nextjs-studio 0.2.0 → 0.4.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -2,13 +2,16 @@
2
2
 
3
3
  // src/bin/nextjs-studio.ts
4
4
  import { existsSync } from "fs";
5
- import path from "path";
5
+ import fs2 from "fs/promises";
6
+ import path2 from "path";
6
7
  import { spawn } from "child_process";
7
8
  import { Command } from "commander";
8
9
 
9
10
  // src/shared/constants.ts
10
11
  var CONTENTS_DIR = "contents";
11
12
  var CLI_PORT = 3030;
13
+ var SUPPORTED_EXTENSIONS = [".mdx", ".json"];
14
+ var COLLECTION_ORDER_FILE = "collection.json";
12
15
  var IMAGE_MIME_TYPES = [
13
16
  "image/png",
14
17
  "image/jpeg",
@@ -28,10 +31,433 @@ var AUDIO_MIME_TYPES = [
28
31
  ];
29
32
  var MEDIA_MIME_TYPES = [...IMAGE_MIME_TYPES, ...VIDEO_MIME_TYPES, ...AUDIO_MIME_TYPES];
30
33
 
34
+ // src/cli/adapters/fs-adapter.ts
35
+ import fs from "fs/promises";
36
+ import path from "path";
37
+ var FsAdapter = class {
38
+ basePath;
39
+ constructor(basePath) {
40
+ this.basePath = path.resolve(basePath);
41
+ }
42
+ resolve(...segments) {
43
+ return path.resolve(this.basePath, ...segments);
44
+ }
45
+ async readFile(filePath) {
46
+ return fs.readFile(this.resolve(filePath), "utf-8");
47
+ }
48
+ async writeFile(filePath, content) {
49
+ const fullPath = this.resolve(filePath);
50
+ await fs.mkdir(path.dirname(fullPath), { recursive: true });
51
+ await fs.writeFile(fullPath, content, "utf-8");
52
+ }
53
+ async deleteFile(filePath) {
54
+ await fs.unlink(this.resolve(filePath));
55
+ }
56
+ async exists(filePath) {
57
+ try {
58
+ await fs.access(this.resolve(filePath));
59
+ return true;
60
+ } catch {
61
+ return false;
62
+ }
63
+ }
64
+ async getStats(filePath) {
65
+ const fullPath = this.resolve(filePath);
66
+ const stats = await fs.stat(fullPath);
67
+ return { path: filePath, size: stats.size, modifiedAt: stats.mtime };
68
+ }
69
+ async listFiles(dirPath, extensions) {
70
+ const fullPath = this.resolve(dirPath);
71
+ const filterExts = extensions ?? SUPPORTED_EXTENSIONS;
72
+ let entries;
73
+ try {
74
+ entries = await fs.readdir(fullPath, { withFileTypes: true });
75
+ } catch {
76
+ return [];
77
+ }
78
+ return entries.filter((entry) => entry.isFile() && filterExts.some((ext) => entry.name.endsWith(ext))).map((entry) => this.join(dirPath, entry.name));
79
+ }
80
+ async listDirectories(dirPath) {
81
+ const fullPath = this.resolve(dirPath);
82
+ let entries;
83
+ try {
84
+ entries = await fs.readdir(fullPath, { withFileTypes: true });
85
+ } catch {
86
+ return [];
87
+ }
88
+ return entries.filter((entry) => entry.isDirectory()).map((entry) => this.join(dirPath, entry.name));
89
+ }
90
+ async readBuffer(filePath) {
91
+ return fs.readFile(this.resolve(filePath));
92
+ }
93
+ async writeBuffer(filePath, data) {
94
+ const fullPath = this.resolve(filePath);
95
+ await fs.mkdir(path.dirname(fullPath), { recursive: true });
96
+ await fs.writeFile(fullPath, data);
97
+ }
98
+ async listAllFiles(dirPath) {
99
+ const fullPath = this.resolve(dirPath);
100
+ let entries;
101
+ try {
102
+ entries = await fs.readdir(fullPath, { withFileTypes: true });
103
+ } catch {
104
+ return [];
105
+ }
106
+ const results = [];
107
+ for (const entry of entries) {
108
+ if (!entry.isFile()) continue;
109
+ const relativePath = this.join(dirPath, entry.name);
110
+ const stats = await fs.stat(this.resolve(relativePath));
111
+ results.push({ name: entry.name, relativePath, size: stats.size, modifiedAt: stats.mtime });
112
+ }
113
+ return results;
114
+ }
115
+ join(...segments) {
116
+ return path.join(...segments);
117
+ }
118
+ basename(filePath) {
119
+ return path.basename(filePath);
120
+ }
121
+ extname(filePath) {
122
+ return path.extname(filePath);
123
+ }
124
+ relative(from, to) {
125
+ return path.relative(from, to);
126
+ }
127
+ normalizeSlug(relativePath, ext) {
128
+ return relativePath.replace(ext, "").split(path.sep).join("/");
129
+ }
130
+ };
131
+
132
+ // src/core/indexer.ts
133
+ import slugify from "@sindresorhus/slugify";
134
+
135
+ // src/core/parsers/parser-mdx.ts
136
+ import matter from "gray-matter";
137
+ function parseMdx(content) {
138
+ const { data, content: body } = matter(content);
139
+ return { data, body: body.trim() };
140
+ }
141
+
142
+ // src/core/parsers/parser-json.ts
143
+ function parseJson(content) {
144
+ const parsed = JSON.parse(content);
145
+ if (Array.isArray(parsed)) {
146
+ return {
147
+ type: "json-array",
148
+ entries: parsed
149
+ };
150
+ }
151
+ if (typeof parsed === "object" && parsed !== null) {
152
+ return {
153
+ type: "json-object",
154
+ data: parsed
155
+ };
156
+ }
157
+ throw new Error("JSON content must be an array or object");
158
+ }
159
+
160
+ // src/core/schema-inferrer.ts
161
+ var RE_ISO_DATE = /^\d{4}-\d{2}-\d{2}$/;
162
+ var RE_ISO_DATETIME = /^\d{4}-\d{2}-\d{2}[T ]\d{2}:\d{2}(:\d{2}(\.\d+)?)?(Z|[+-]\d{2}:?\d{2})?$/;
163
+ var RE_EMAIL = /^[^\s@]+@[^\s@]+\.[^\s@]+$/;
164
+ var RE_URL = /^https?:\/\/.+/;
165
+ var LONG_TEXT_THRESHOLD = 200;
166
+ function isISODate(value) {
167
+ return RE_ISO_DATE.test(value);
168
+ }
169
+ function isISODateTime(value) {
170
+ return RE_ISO_DATETIME.test(value);
171
+ }
172
+ function isEmail(value) {
173
+ return RE_EMAIL.test(value);
174
+ }
175
+ function isUrl(value) {
176
+ return RE_URL.test(value);
177
+ }
178
+ function inferStringField(name, strings) {
179
+ if (strings.every(isEmail)) return { name, type: "email" };
180
+ if (strings.every(isUrl)) return { name, type: "url" };
181
+ if (strings.every(isISODateTime)) return { name, type: "date", includeTime: true };
182
+ if (strings.every(isISODate)) return { name, type: "date" };
183
+ const isLong = strings.some((s) => s.length > LONG_TEXT_THRESHOLD || s.includes("\n"));
184
+ return { name, type: isLong ? "long-text" : "text" };
185
+ }
186
+ function inferArrayField(name, items) {
187
+ if (items.length === 0) return { name, type: "array", itemFields: [] };
188
+ if (items.every((item) => typeof item === "string")) {
189
+ const unique = [...new Set(items)].slice(0, 50);
190
+ const options = unique.map((v) => ({ label: v, value: v }));
191
+ return { name, type: "multi-select", options };
192
+ }
193
+ if (items.every((item) => typeof item === "object" && item !== null && !Array.isArray(item))) {
194
+ return { name, type: "array", itemFields: inferFields(items) };
195
+ }
196
+ return { name, type: "array", itemFields: [] };
197
+ }
198
+ function inferFieldDefinition(name, values) {
199
+ const present = values.filter((v) => v !== null && v !== void 0);
200
+ if (present.length === 0) return { name, type: "text" };
201
+ if (present.every((v) => typeof v === "boolean")) return { name, type: "boolean" };
202
+ if (present.every((v) => typeof v === "number")) {
203
+ const format = present.every((v) => Number.isInteger(v)) ? "integer" : "decimal";
204
+ return { name, type: "number", format };
205
+ }
206
+ if (present.every((v) => typeof v === "string")) {
207
+ return inferStringField(name, present);
208
+ }
209
+ if (present.every((v) => Array.isArray(v))) {
210
+ return inferArrayField(name, present.flat());
211
+ }
212
+ if (present.every((v) => typeof v === "object" && v !== null && !Array.isArray(v))) {
213
+ return { name, type: "object", fields: inferFields(present) };
214
+ }
215
+ return { name, type: "text" };
216
+ }
217
+ function inferFields(rows) {
218
+ const keySet = new Set(rows.flatMap((row) => Object.keys(row)));
219
+ return Array.from(keySet).map((key) => inferFieldDefinition(key, rows.map((row) => row[key])));
220
+ }
221
+ function inferSchema(entries, collectionName) {
222
+ const rows = entries.map((entry) => entry.data);
223
+ return { collection: collectionName, fields: inferFields(rows) };
224
+ }
225
+
226
+ // src/core/indexer.ts
227
+ var ContentIndex = class {
228
+ entries = /* @__PURE__ */ new Map();
229
+ collections = /* @__PURE__ */ new Map();
230
+ fs;
231
+ constructor(fsAdapter) {
232
+ this.fs = fsAdapter;
233
+ }
234
+ async build(config) {
235
+ this.clear();
236
+ const dirs = await this.fs.listDirectories(".");
237
+ for (const dir of dirs) {
238
+ const dirName = this.fs.basename(dir);
239
+ const collectionName = slugify(dirName);
240
+ const collectionConfig = config?.collections?.[collectionName];
241
+ await this.indexCollection(dirName, collectionName, collectionConfig?.schema);
242
+ }
243
+ }
244
+ getCollection(name) {
245
+ return this.entries.get(name) ?? [];
246
+ }
247
+ getCollections() {
248
+ return Array.from(this.collections.values());
249
+ }
250
+ clear() {
251
+ this.entries.clear();
252
+ this.collections.clear();
253
+ }
254
+ async indexCollection(dirName, collectionName, manualSchema) {
255
+ const entries = [];
256
+ await this.scanDir(dirName, collectionName, dirName, entries);
257
+ const orderPath = this.fs.join(dirName, COLLECTION_ORDER_FILE);
258
+ const ordering = await this.readOrdering(orderPath);
259
+ if (ordering) {
260
+ this.applyOrdering(entries, ordering);
261
+ }
262
+ const schema = manualSchema ?? inferSchema(entries, collectionName);
263
+ this.entries.set(collectionName, entries);
264
+ this.collections.set(collectionName, {
265
+ name: collectionName,
266
+ type: this.detectCollectionType(entries),
267
+ count: entries.length,
268
+ basePath: dirName,
269
+ schema
270
+ });
271
+ }
272
+ async scanDir(dirName, collectionName, dirPath, entries) {
273
+ const subDirs = await this.fs.listDirectories(dirPath);
274
+ for (const subDir of subDirs) {
275
+ await this.scanDir(dirName, collectionName, subDir, entries);
276
+ }
277
+ const files = await this.fs.listFiles(dirPath);
278
+ for (const filePath of files) {
279
+ const fileName = this.fs.basename(filePath);
280
+ if (fileName === COLLECTION_ORDER_FILE) continue;
281
+ const ext = this.fs.extname(fileName);
282
+ const content = await this.fs.readFile(filePath);
283
+ const relativePath = this.fs.relative(dirName, filePath);
284
+ const slug = this.fs.normalizeSlug(relativePath, ext).split("/").map((segment) => slugify(segment)).join("/");
285
+ if (ext === ".mdx") {
286
+ entries.push(this.buildMdxEntry(collectionName, slug, content));
287
+ } else if (ext === ".json") {
288
+ entries.push(...this.buildJsonEntries(collectionName, slug, content));
289
+ }
290
+ }
291
+ }
292
+ buildMdxEntry(collectionName, slug, content) {
293
+ const parsed = parseMdx(content);
294
+ return {
295
+ collection: collectionName,
296
+ slug,
297
+ path: `/${collectionName}/${slug}`,
298
+ body: parsed.body,
299
+ data: parsed.data
300
+ };
301
+ }
302
+ buildJsonEntries(collectionName, slug, content) {
303
+ const parsed = parseJson(content);
304
+ if (parsed.type === "json-array") {
305
+ return parsed.entries.map((data, index) => {
306
+ const entrySlug = typeof data["slug"] === "string" ? slugify(data["slug"]) : `${slug}/${index}`;
307
+ return {
308
+ collection: collectionName,
309
+ slug: entrySlug,
310
+ path: `/${collectionName}/${entrySlug}`,
311
+ data
312
+ };
313
+ });
314
+ }
315
+ return [{ collection: collectionName, slug, path: `/${collectionName}/${slug}`, data: parsed.data }];
316
+ }
317
+ async readOrdering(orderPath) {
318
+ if (!await this.fs.exists(orderPath)) return null;
319
+ try {
320
+ const content = await this.fs.readFile(orderPath);
321
+ const parsed = JSON.parse(content);
322
+ if (Array.isArray(parsed)) return parsed;
323
+ } catch (error) {
324
+ console.warn(`[Nextjs Studio] Failed to parse ordering file: ${orderPath}`, error);
325
+ }
326
+ return null;
327
+ }
328
+ applyOrdering(entries, ordering) {
329
+ const orderMap = new Map(ordering.map((slug, index) => [slug, index]));
330
+ entries.sort((a, b) => {
331
+ const aIndex = orderMap.get(a.slug) ?? Infinity;
332
+ const bIndex = orderMap.get(b.slug) ?? Infinity;
333
+ return aIndex - bIndex;
334
+ });
335
+ }
336
+ detectCollectionType(entries) {
337
+ if (entries.length === 0) return "mdx";
338
+ const first = entries[0];
339
+ if (first.body !== void 0) return "mdx";
340
+ if (entries.length === 1 && !first.slug.includes("/")) return "json-object";
341
+ return "json-array";
342
+ }
343
+ };
344
+
345
+ // src/core/content-store.ts
346
+ var store = null;
347
+ async function loadContent(fsAdapter, config) {
348
+ const index = new ContentIndex(fsAdapter);
349
+ await index.build(config);
350
+ store = index;
351
+ return index;
352
+ }
353
+
354
+ // src/core/type-generator.ts
355
+ function indent(code, spaces = 2) {
356
+ return code.split("\n").map((line) => line.trim() === "" ? "" : " ".repeat(spaces) + line).join("\n");
357
+ }
358
+ function toPascalCase(str) {
359
+ return str.replace(/[-_\s]+(.)/g, (_, c) => c.toUpperCase()).replace(/^(.)/, (_, c) => c.toUpperCase());
360
+ }
361
+ function fieldToTsType(field) {
362
+ switch (field.type) {
363
+ case "text":
364
+ case "long-text":
365
+ return "string";
366
+ case "email":
367
+ return "Email";
368
+ case "url":
369
+ return "HttpUrl";
370
+ case "media":
371
+ return "MediaPath";
372
+ case "id":
373
+ return "ID";
374
+ case "slug":
375
+ return "Slug";
376
+ case "date":
377
+ return field.includeTime ? "Date" : "ISODate";
378
+ case "created-time":
379
+ case "updated-time":
380
+ return "Date";
381
+ case "number":
382
+ return "number";
383
+ case "boolean":
384
+ return "boolean";
385
+ case "select":
386
+ case "status": {
387
+ const values = field.options.map((o) => JSON.stringify(o.value));
388
+ return values.length > 0 ? values.join(" | ") : "string";
389
+ }
390
+ case "multi-select": {
391
+ const values = field.options.map((o) => JSON.stringify(o.value));
392
+ const union = values.length > 0 ? values.join(" | ") : "string";
393
+ return `Array<${union}>`;
394
+ }
395
+ case "object":
396
+ return generateObjectType(field.fields);
397
+ case "array":
398
+ return `Array<${generateObjectType(field.itemFields)}>`;
399
+ case "relation":
400
+ return field.multiple === true ? "ID[]" : "ID";
401
+ case "formula":
402
+ if (field.resultType === "number") return "number";
403
+ if (field.resultType === "boolean") return "boolean";
404
+ return "string";
405
+ default:
406
+ return "unknown";
407
+ }
408
+ }
409
+ function generateObjectType(fields) {
410
+ if (fields.length === 0) return "Record<string, unknown>";
411
+ const lines = fields.map((field) => {
412
+ const optional = field.required === false ? "?" : "";
413
+ const tsType = fieldToTsType(field);
414
+ const comment = field.description ? `/** ${field.description} */
415
+ ` : "";
416
+ return `${comment}${field.name}${optional}: ${tsType};`;
417
+ });
418
+ return `{
419
+ ${indent(lines.join("\n"))}
420
+ }`;
421
+ }
422
+ function generateInterfaceForSchema(schema) {
423
+ const name = toPascalCase(schema.collection) + "Entry";
424
+ const label = schema.label ?? schema.collection;
425
+ const body = generateObjectType(schema.fields);
426
+ return `/** Data shape for the "${label}" collection. */
427
+ export interface ${name} ${body}`;
428
+ }
429
+ function generateCollectionTypes(schemas) {
430
+ const banner = [
431
+ "// This file is auto-generated by nextjs-studio.",
432
+ "// Do not edit manually \u2014 re-run `npx nextjs-studio --generate-types` to update.",
433
+ "",
434
+ "// Branded scalar types \u2014 structurally strings/numbers but semantically distinct.",
435
+ "declare const __brand: unique symbol;",
436
+ "type Brand<T, B extends string> = T & { readonly [__brand]: B };",
437
+ "",
438
+ "export type Email = Brand<string, 'Email'>;",
439
+ "export type HttpUrl = Brand<string, 'HttpUrl'>;",
440
+ "export type ISODate = Brand<string, 'ISODate'>;",
441
+ "export type MediaPath = Brand<string, 'MediaPath'>;",
442
+ "export type ID = Brand<string, 'ID'>;",
443
+ "export type Slug = Brand<string, 'Slug'>;"
444
+ ].join("\n");
445
+ const interfaces = schemas.map(generateInterfaceForSchema).join("\n\n");
446
+ const collectionRegistry = [
447
+ "// Augment the nextjs-studio module so queryCollection() is fully typed.",
448
+ "declare module 'nextjs-studio' {",
449
+ " interface CollectionTypeMap {",
450
+ schemas.map((schema) => ` ${JSON.stringify(schema.collection)}: ${toPascalCase(schema.collection)}Entry;`).join("\n"),
451
+ " }",
452
+ "}"
453
+ ].join("\n");
454
+ return [banner, interfaces, collectionRegistry].join("\n\n") + "\n";
455
+ }
456
+
31
457
  // package.json
32
458
  var package_default = {
33
459
  name: "nextjs-studio",
34
- version: "0.2.0",
460
+ version: "0.4.0",
35
461
  description: "A Git-based, local-first CMS for Next.js projects",
36
462
  keywords: [
37
463
  "nextjs",
@@ -45,7 +471,7 @@ var package_default = {
45
471
  homepage: "https://github.com/TiagoDanin/Nextjs-Studio",
46
472
  repository: {
47
473
  type: "git",
48
- url: "https://github.com/TiagoDanin/Nextjs-Studio.git"
474
+ url: "git+https://github.com/TiagoDanin/Nextjs-Studio.git"
49
475
  },
50
476
  license: "MIT",
51
477
  author: "Tiago Danin",
@@ -58,7 +484,9 @@ var package_default = {
58
484
  },
59
485
  main: "./dist/core/index.js",
60
486
  types: "./dist/core/index.d.ts",
61
- bin: "./dist/bin/nextjs-studio.js",
487
+ bin: {
488
+ "nextjs-studio": "dist/bin/nextjs-studio.js"
489
+ },
62
490
  files: [
63
491
  "dist",
64
492
  "README.md",
@@ -130,42 +558,62 @@ var package_default = {
130
558
 
131
559
  // src/bin/nextjs-studio.ts
132
560
  var { version } = package_default;
133
- var program = new Command().name("Nextjs Studio").description("Local-first CMS for Next.js projects").version(version).option("-d, --dir <path>", "Path to contents directory", CONTENTS_DIR).option("-p, --port <number>", "Port to run the studio on", String(CLI_PORT)).parse();
561
+ var program = new Command().name("Nextjs Studio").description("Local-first CMS for Next.js projects").version(version).option("-d, --dir <path>", "Path to contents directory", CONTENTS_DIR).option("-p, --port <number>", "Port to run the studio on", String(CLI_PORT)).option("--generate-types", "Generate TypeScript types for content collections").parse();
134
562
  var opts = program.opts();
135
- var contentsDir = path.resolve(opts.dir);
563
+ var contentsDir = path2.resolve(opts.dir);
136
564
  var port = Number(opts.port);
137
- var uiDir = path.resolve(import.meta.dirname, "../cli/ui");
138
- var standaloneServer = path.resolve(uiDir, ".next/standalone/src/cli/ui/server.js");
139
- var serverEnv = {
140
- ...process.env,
141
- STUDIO_CONTENTS_DIR: contentsDir,
142
- PORT: String(port),
143
- HOSTNAME: "0.0.0.0"
144
- };
145
- console.log(`Nextjs Studio v${version}`);
146
- console.log(`Contents: ${contentsDir}`);
147
- console.log(`Starting on http://localhost:${port}`);
148
- function createServerProcess() {
565
+ async function runGenerateTypes(sourceDir) {
566
+ const outDir = path2.resolve(".studio");
567
+ const outFile = path2.join(outDir, "studio.d.ts");
568
+ console.log(`Generating types from ${sourceDir}...`);
569
+ const fsAdapter = new FsAdapter(sourceDir);
570
+ const index = await loadContent(fsAdapter);
571
+ const schemas = index.getCollections().flatMap((c) => c.schema ? [c.schema] : []);
572
+ const code = generateCollectionTypes(schemas);
573
+ await fs2.mkdir(outDir, { recursive: true });
574
+ await fs2.writeFile(outFile, code, "utf-8");
575
+ console.log(`Types written to ${outFile}`);
576
+ }
577
+ function resolveServerProcess(uiDir2, serverPort, env) {
578
+ const standaloneServer = path2.resolve(uiDir2, ".next/standalone/src/cli/ui/server.js");
149
579
  if (existsSync(standaloneServer)) {
150
- return spawn("node", [standaloneServer], { stdio: "inherit", env: serverEnv });
580
+ return spawn("node", [standaloneServer], { stdio: "inherit", env });
151
581
  }
152
- return spawn("npx", ["next", "dev", "--port", String(port), "--webpack"], {
153
- cwd: uiDir,
154
- stdio: "inherit",
155
- shell: true,
156
- env: serverEnv
157
- });
582
+ const uiPackageJson = path2.resolve(uiDir2, "package.json");
583
+ if (existsSync(uiPackageJson)) {
584
+ const nextBin = path2.resolve(uiDir2, "../../../node_modules/next/dist/bin/next");
585
+ return spawn("node", [nextBin, "dev", "--port", String(serverPort), "--webpack"], {
586
+ cwd: uiDir2,
587
+ stdio: "inherit",
588
+ env
589
+ });
590
+ }
591
+ return null;
158
592
  }
159
- function registerSignalForwarding(child) {
593
+ function forwardSignals(child) {
160
594
  for (const signal of ["SIGINT", "SIGTERM"]) {
161
595
  process.on(signal, () => child.kill(signal));
162
596
  }
163
597
  }
164
- var serverProcess = createServerProcess();
598
+ if (opts.generateTypes) {
599
+ await runGenerateTypes(contentsDir);
600
+ process.exit(0);
601
+ }
602
+ var uiDir = path2.resolve(import.meta.dirname, "../cli/ui");
603
+ var serverEnv = { ...process.env, STUDIO_CONTENTS_DIR: contentsDir, PORT: String(port), HOSTNAME: "0.0.0.0" };
604
+ var serverProcess = resolveServerProcess(uiDir, port, serverEnv);
605
+ if (!serverProcess) {
606
+ console.error("Error: Studio UI server not found.");
607
+ console.error("The pre-built UI is not included in this installation.");
608
+ process.exit(1);
609
+ }
610
+ console.log(`Nextjs Studio v${version}`);
611
+ console.log(`Contents: ${contentsDir}`);
612
+ console.log(`Starting on http://localhost:${port}`);
165
613
  serverProcess.on("error", (error) => {
166
614
  console.error("Failed to start server:", error.message);
167
615
  process.exit(1);
168
616
  });
169
617
  serverProcess.on("close", (code) => process.exit(code ?? 0));
170
- registerSignalForwarding(serverProcess);
618
+ forwardSignals(serverProcess);
171
619
  //# sourceMappingURL=nextjs-studio.js.map
@@ -1 +1 @@
1
- {"version":3,"sources":["../../src/bin/nextjs-studio.ts","../../src/shared/constants.ts","../../package.json"],"sourcesContent":["#!/usr/bin/env node\n\n/**\n * @context bin layer — CLI entry point at src/bin/nextjs-studio.ts\n * @does Parses CLI args, resolves paths, and spawns the UI server process\n * @depends src/shared/constants.ts\n * @do Add new CLI flags here; keep only process bootstrap logic\n * @dont Import UI components, access the filesystem beyond existsSync, or contain business logic\n */\n\nimport { existsSync } from \"node:fs\";\nimport path from \"node:path\";\nimport { spawn, type ChildProcess } from \"node:child_process\";\nimport { Command } from \"commander\";\nimport { CLI_PORT, CONTENTS_DIR } from \"../shared/constants.js\";\nimport pkg from \"../../package.json\" with { type: \"json\" };\n\nconst { version } = pkg;\n\nconst program = new Command()\n .name(\"Nextjs Studio\")\n .description(\"Local-first CMS for Next.js projects\")\n .version(version)\n .option(\"-d, --dir <path>\", \"Path to contents directory\", CONTENTS_DIR)\n .option(\"-p, --port <number>\", \"Port to run the studio on\", String(CLI_PORT))\n .parse();\n\nconst opts = program.opts<{ dir: string; port: string }>();\nconst contentsDir = path.resolve(opts.dir);\nconst port = Number(opts.port);\n\nconst uiDir = path.resolve(import.meta.dirname, \"../cli/ui\");\nconst standaloneServer = path.resolve(uiDir, \".next/standalone/src/cli/ui/server.js\");\n\nconst serverEnv = {\n ...process.env,\n STUDIO_CONTENTS_DIR: contentsDir,\n PORT: String(port),\n HOSTNAME: \"0.0.0.0\",\n};\n\nconsole.log(`Nextjs Studio v${version}`);\nconsole.log(`Contents: ${contentsDir}`);\nconsole.log(`Starting on http://localhost:${port}`);\n\nfunction createServerProcess(): ChildProcess {\n if (existsSync(standaloneServer)) {\n return spawn(\"node\", [standaloneServer], { stdio: \"inherit\", env: serverEnv });\n }\n\n return spawn(\"npx\", [\"next\", \"dev\", \"--port\", String(port), \"--webpack\"], {\n cwd: uiDir,\n stdio: \"inherit\",\n shell: true,\n env: serverEnv,\n });\n}\n\nfunction registerSignalForwarding(child: ChildProcess): void {\n for (const signal of [\"SIGINT\", \"SIGTERM\"] as const) {\n process.on(signal, () => child.kill(signal));\n }\n}\n\nconst serverProcess = createServerProcess();\n\nserverProcess.on(\"error\", (error) => {\n console.error(\"Failed to start server:\", error.message);\n process.exit(1);\n});\n\nserverProcess.on(\"close\", (code) => process.exit(code ?? 0));\n\nregisterSignalForwarding(serverProcess);\n","/**\n * @context Shared layer — constants at src/shared/constants.ts\n * @does Defines project-wide constants shared across core, CLI, and UI layers\n * @depends none\n * @do Add new shared constants here\n * @dont Import from CLI or UI; constants must be framework-agnostic\n */\n\nexport const CONTENTS_DIR = \"contents\";\nexport const CLI_PORT = 3030;\nexport const CONFIG_FILE = \"studio.config.ts\";\nexport const SUPPORTED_EXTENSIONS = [\".mdx\", \".json\"] as const;\nexport const COLLECTION_ORDER_FILE = \"collection.json\";\nexport const WATCHER_DEBOUNCE_MS = 5_000;\nexport const MEDIA_DIR = \"media\";\n\nexport const IMAGE_MIME_TYPES = [\n \"image/png\",\n \"image/jpeg\",\n \"image/gif\",\n \"image/webp\",\n \"image/svg+xml\",\n \"image/avif\",\n] as const;\n\nexport const VIDEO_MIME_TYPES = [\"video/mp4\", \"video/webm\", \"video/ogg\"] as const;\n\nexport const AUDIO_MIME_TYPES = [\n \"audio/mpeg\",\n \"audio/ogg\",\n \"audio/wav\",\n \"audio/webm\",\n \"audio/aac\",\n \"audio/flac\",\n] as const;\n\nexport const MEDIA_MIME_TYPES = [...IMAGE_MIME_TYPES, ...VIDEO_MIME_TYPES, ...AUDIO_MIME_TYPES] as const;\n\nexport const IMAGE_EXTENSIONS = [\".png\", \".jpg\", \".jpeg\", \".gif\", \".webp\", \".svg\", \".avif\"] as const;\nexport const VIDEO_EXTENSIONS = [\".mp4\", \".webm\", \".ogv\"] as const;\nexport const AUDIO_EXTENSIONS = [\".mp3\", \".ogg\", \".wav\", \".m4a\", \".aac\", \".flac\"] as const;\n","{\n \"name\": \"nextjs-studio\",\n \"version\": \"0.2.0\",\n \"description\": \"A Git-based, local-first CMS for Next.js projects\",\n \"keywords\": [\n \"nextjs\",\n \"cms\",\n \"mdx\",\n \"content\",\n \"studio\",\n \"static-site\",\n \"local-first\"\n ],\n \"homepage\": \"https://github.com/TiagoDanin/Nextjs-Studio\",\n \"repository\": {\n \"type\": \"git\",\n \"url\": \"https://github.com/TiagoDanin/Nextjs-Studio.git\"\n },\n \"license\": \"MIT\",\n \"author\": \"Tiago Danin\",\n \"type\": \"module\",\n \"exports\": {\n \".\": {\n \"types\": \"./dist/core/index.d.ts\",\n \"import\": \"./dist/core/index.js\"\n }\n },\n \"main\": \"./dist/core/index.js\",\n \"types\": \"./dist/core/index.d.ts\",\n \"bin\": \"./dist/bin/nextjs-studio.js\",\n \"files\": [\n \"dist\",\n \"README.md\",\n \"LICENSE\"\n ],\n \"scripts\": {\n \"dev\": \"tsx src/bin/nextjs-studio.ts --dir example/contents\",\n \"studio:dev\": \"cross-env STUDIO_CONTENTS_DIR=example/contents next dev --port 3030 --webpack src/cli/ui\",\n \"studio:build\": \"next build --webpack src/cli/ui\",\n \"build\": \"tsup && yarn studio:build\",\n \"lint\": \"eslint src/\",\n \"type-check\": \"tsc --noEmit\",\n \"test\": \"vitest run\",\n \"test:watch\": \"vitest\"\n },\n \"engines\": {\n \"node\": \">=22.10.0\"\n },\n \"packageManager\": \"yarn@4.6.0\",\n \"devDependencies\": {\n \"@radix-ui/react-collapsible\": \"^1.1.12\",\n \"@radix-ui/react-label\": \"^2.1.8\",\n \"@radix-ui/react-switch\": \"^1.2.6\",\n \"@tailwindcss/postcss\": \"^4.1.18\",\n \"@tanstack/react-table\": \"^8.21.3\",\n \"@tiptap/extension-bubble-menu\": \"^3.20.0\",\n \"@tiptap/extension-code-block-lowlight\": \"^3.20.0\",\n \"@tiptap/extension-file-handler\": \"^3.20.0\",\n \"@tiptap/extension-image\": \"^3.20.0\",\n \"@tiptap/extension-link\": \"^3.20.0\",\n \"@tiptap/extension-placeholder\": \"^3.20.0\",\n \"@tiptap/react\": \"^3.20.0\",\n \"@tiptap/starter-kit\": \"^3.20.0\",\n \"@tiptap/suggestion\": \"^3.20.0\",\n \"@types/lodash-es\": \"^4.17.12\",\n \"@types/node\": \"^25.2.3\",\n \"@types/react\": \"^19\",\n \"@types/react-dom\": \"^19\",\n \"class-variance-authority\": \"^0.7.1\",\n \"clsx\": \"^2.1.1\",\n \"cross-env\": \"^10.1.0\",\n \"eslint\": \"^10.0.0\",\n \"lowlight\": \"^3.3.0\",\n \"lucide-react\": \"^0.574.0\",\n \"mermaid\": \"^11.6.0\",\n \"next\": \"^16.1.6\",\n \"next-themes\": \"^0.4.6\",\n \"react\": \"^19.2.4\",\n \"react-dom\": \"^19.2.4\",\n \"tailwind-merge\": \"^3.4.1\",\n \"tailwindcss\": \"^4.1.18\",\n \"tippy.js\": \"^6.3.7\",\n \"tiptap-extension-global-drag-handle\": \"^0.1.18\",\n \"tiptap-markdown\": \"^0.9.0\",\n \"tsup\": \"^8.5.1\",\n \"tsx\": \"^4.21.0\",\n \"typescript\": \"^5.9.3\",\n \"vitest\": \"^4.0.18\",\n \"zustand\": \"^5.0.11\"\n },\n \"dependencies\": {\n \"@sindresorhus/slugify\": \"^3.0.0\",\n \"chokidar\": \"^5.0.0\",\n \"commander\": \"^14.0.3\",\n \"gray-matter\": \"^4.0.3\",\n \"lodash-es\": \"^4.17.23\"\n }\n}"],"mappings":";;;AAUA,SAAS,kBAAkB;AAC3B,OAAO,UAAU;AACjB,SAAS,aAAgC;AACzC,SAAS,eAAe;;;ACLjB,IAAM,eAAe;AACrB,IAAM,WAAW;AAOjB,IAAM,mBAAmB;AAAA,EAC9B;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AACF;AAEO,IAAM,mBAAmB,CAAC,aAAa,cAAc,WAAW;AAEhE,IAAM,mBAAmB;AAAA,EAC9B;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AACF;AAEO,IAAM,mBAAmB,CAAC,GAAG,kBAAkB,GAAG,kBAAkB,GAAG,gBAAgB;;;ACpC9F;AAAA,EACI,MAAQ;AAAA,EACR,SAAW;AAAA,EACX,aAAe;AAAA,EACf,UAAY;AAAA,IACR;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,EACJ;AAAA,EACA,UAAY;AAAA,EACZ,YAAc;AAAA,IACV,MAAQ;AAAA,IACR,KAAO;AAAA,EACX;AAAA,EACA,SAAW;AAAA,EACX,QAAU;AAAA,EACV,MAAQ;AAAA,EACR,SAAW;AAAA,IACP,KAAK;AAAA,MACD,OAAS;AAAA,MACT,QAAU;AAAA,IACd;AAAA,EACJ;AAAA,EACA,MAAQ;AAAA,EACR,OAAS;AAAA,EACT,KAAO;AAAA,EACP,OAAS;AAAA,IACL;AAAA,IACA;AAAA,IACA;AAAA,EACJ;AAAA,EACA,SAAW;AAAA,IACP,KAAO;AAAA,IACP,cAAc;AAAA,IACd,gBAAgB;AAAA,IAChB,OAAS;AAAA,IACT,MAAQ;AAAA,IACR,cAAc;AAAA,IACd,MAAQ;AAAA,IACR,cAAc;AAAA,EAClB;AAAA,EACA,SAAW;AAAA,IACP,MAAQ;AAAA,EACZ;AAAA,EACA,gBAAkB;AAAA,EAClB,iBAAmB;AAAA,IACf,+BAA+B;AAAA,IAC/B,yBAAyB;AAAA,IACzB,0BAA0B;AAAA,IAC1B,wBAAwB;AAAA,IACxB,yBAAyB;AAAA,IACzB,iCAAiC;AAAA,IACjC,yCAAyC;AAAA,IACzC,kCAAkC;AAAA,IAClC,2BAA2B;AAAA,IAC3B,0BAA0B;AAAA,IAC1B,iCAAiC;AAAA,IACjC,iBAAiB;AAAA,IACjB,uBAAuB;AAAA,IACvB,sBAAsB;AAAA,IACtB,oBAAoB;AAAA,IACpB,eAAe;AAAA,IACf,gBAAgB;AAAA,IAChB,oBAAoB;AAAA,IACpB,4BAA4B;AAAA,IAC5B,MAAQ;AAAA,IACR,aAAa;AAAA,IACb,QAAU;AAAA,IACV,UAAY;AAAA,IACZ,gBAAgB;AAAA,IAChB,SAAW;AAAA,IACX,MAAQ;AAAA,IACR,eAAe;AAAA,IACf,OAAS;AAAA,IACT,aAAa;AAAA,IACb,kBAAkB;AAAA,IAClB,aAAe;AAAA,IACf,YAAY;AAAA,IACZ,uCAAuC;AAAA,IACvC,mBAAmB;AAAA,IACnB,MAAQ;AAAA,IACR,KAAO;AAAA,IACP,YAAc;AAAA,IACd,QAAU;AAAA,IACV,SAAW;AAAA,EACf;AAAA,EACA,cAAgB;AAAA,IACZ,yBAAyB;AAAA,IACzB,UAAY;AAAA,IACZ,WAAa;AAAA,IACb,eAAe;AAAA,IACf,aAAa;AAAA,EACjB;AACJ;;;AFhFA,IAAM,EAAE,QAAQ,IAAI;AAEpB,IAAM,UAAU,IAAI,QAAQ,EACzB,KAAK,eAAe,EACpB,YAAY,sCAAsC,EAClD,QAAQ,OAAO,EACf,OAAO,oBAAoB,8BAA8B,YAAY,EACrE,OAAO,uBAAuB,6BAA6B,OAAO,QAAQ,CAAC,EAC3E,MAAM;AAET,IAAM,OAAO,QAAQ,KAAoC;AACzD,IAAM,cAAc,KAAK,QAAQ,KAAK,GAAG;AACzC,IAAM,OAAO,OAAO,KAAK,IAAI;AAE7B,IAAM,QAAQ,KAAK,QAAQ,YAAY,SAAS,WAAW;AAC3D,IAAM,mBAAmB,KAAK,QAAQ,OAAO,uCAAuC;AAEpF,IAAM,YAAY;AAAA,EAChB,GAAG,QAAQ;AAAA,EACX,qBAAqB;AAAA,EACrB,MAAM,OAAO,IAAI;AAAA,EACjB,UAAU;AACZ;AAEA,QAAQ,IAAI,kBAAkB,OAAO,EAAE;AACvC,QAAQ,IAAI,aAAa,WAAW,EAAE;AACtC,QAAQ,IAAI,gCAAgC,IAAI,EAAE;AAElD,SAAS,sBAAoC;AAC3C,MAAI,WAAW,gBAAgB,GAAG;AAChC,WAAO,MAAM,QAAQ,CAAC,gBAAgB,GAAG,EAAE,OAAO,WAAW,KAAK,UAAU,CAAC;AAAA,EAC/E;AAEA,SAAO,MAAM,OAAO,CAAC,QAAQ,OAAO,UAAU,OAAO,IAAI,GAAG,WAAW,GAAG;AAAA,IACxE,KAAK;AAAA,IACL,OAAO;AAAA,IACP,OAAO;AAAA,IACP,KAAK;AAAA,EACP,CAAC;AACH;AAEA,SAAS,yBAAyB,OAA2B;AAC3D,aAAW,UAAU,CAAC,UAAU,SAAS,GAAY;AACnD,YAAQ,GAAG,QAAQ,MAAM,MAAM,KAAK,MAAM,CAAC;AAAA,EAC7C;AACF;AAEA,IAAM,gBAAgB,oBAAoB;AAE1C,cAAc,GAAG,SAAS,CAAC,UAAU;AACnC,UAAQ,MAAM,2BAA2B,MAAM,OAAO;AACtD,UAAQ,KAAK,CAAC;AAChB,CAAC;AAED,cAAc,GAAG,SAAS,CAAC,SAAS,QAAQ,KAAK,QAAQ,CAAC,CAAC;AAE3D,yBAAyB,aAAa;","names":[]}
1
+ {"version":3,"sources":["../../src/bin/nextjs-studio.ts","../../src/shared/constants.ts","../../src/cli/adapters/fs-adapter.ts","../../src/core/indexer.ts","../../src/core/parsers/parser-mdx.ts","../../src/core/parsers/parser-json.ts","../../src/core/schema-inferrer.ts","../../src/core/content-store.ts","../../src/core/type-generator.ts","../../package.json"],"sourcesContent":["#!/usr/bin/env node\n\n/**\n * @context bin layer — CLI entry point at src/bin/nextjs-studio.ts\n * @does Parses CLI args, then either generates types or spawns the UI server process\n * @depends src/shared/constants.ts, src/cli/adapters/fs-adapter.ts, src/core/content-store.ts, src/core/type-generator.ts\n * @do Add new CLI flags here; keep only process bootstrap logic\n * @dont Import UI components or contain parsing/indexing business logic\n */\n\nimport { existsSync } from \"node:fs\";\nimport fs from \"node:fs/promises\";\nimport path from \"node:path\";\nimport { spawn, type ChildProcess } from \"node:child_process\";\nimport { Command } from \"commander\";\nimport { CLI_PORT, CONTENTS_DIR } from \"../shared/constants.js\";\nimport { FsAdapter } from \"../cli/adapters/fs-adapter.js\";\nimport { loadContent } from \"../core/content-store.js\";\nimport { generateCollectionTypes } from \"../core/type-generator.js\";\nimport pkg from \"../../package.json\" with { type: \"json\" };\n\nconst { version } = pkg;\n\nconst program = new Command()\n .name(\"Nextjs Studio\")\n .description(\"Local-first CMS for Next.js projects\")\n .version(version)\n .option(\"-d, --dir <path>\", \"Path to contents directory\", CONTENTS_DIR)\n .option(\"-p, --port <number>\", \"Port to run the studio on\", String(CLI_PORT))\n .option(\"--generate-types\", \"Generate TypeScript types for content collections\")\n .parse();\n\nconst opts = program.opts<{ dir: string; port: string; generateTypes?: boolean }>();\nconst contentsDir = path.resolve(opts.dir);\nconst port = Number(opts.port);\n\nasync function runGenerateTypes(sourceDir: string): Promise<void> {\n const outDir = path.resolve(\".studio\");\n const outFile = path.join(outDir, \"studio.d.ts\");\n\n console.log(`Generating types from ${sourceDir}...`);\n\n const fsAdapter = new FsAdapter(sourceDir);\n const index = await loadContent(fsAdapter);\n const schemas = index.getCollections().flatMap((c) => (c.schema ? [c.schema] : []));\n const code = generateCollectionTypes(schemas);\n\n await fs.mkdir(outDir, { recursive: true });\n await fs.writeFile(outFile, code, \"utf-8\");\n\n console.log(`Types written to ${outFile}`);\n}\n\nfunction resolveServerProcess(\n uiDir: string,\n serverPort: number,\n env: NodeJS.ProcessEnv,\n): ChildProcess | null {\n const standaloneServer = path.resolve(uiDir, \".next/standalone/src/cli/ui/server.js\");\n if (existsSync(standaloneServer)) {\n return spawn(\"node\", [standaloneServer], { stdio: \"inherit\", env });\n }\n\n // Dev mode: UI source present (running from repo with `yarn dev`)\n const uiPackageJson = path.resolve(uiDir, \"package.json\");\n if (existsSync(uiPackageJson)) {\n // Resolve `next` bin from workspace root (3 levels up from src/cli/ui)\n const nextBin = path.resolve(uiDir, \"../../../node_modules/next/dist/bin/next\");\n return spawn(\"node\", [nextBin, \"dev\", \"--port\", String(serverPort), \"--webpack\"], {\n cwd: uiDir,\n stdio: \"inherit\",\n env,\n });\n }\n\n return null;\n}\n\nfunction forwardSignals(child: ChildProcess): void {\n for (const signal of [\"SIGINT\", \"SIGTERM\"] as const) {\n process.on(signal, () => child.kill(signal));\n }\n}\n\nif (opts.generateTypes) {\n await runGenerateTypes(contentsDir);\n process.exit(0);\n}\n\nconst uiDir = path.resolve(import.meta.dirname, \"../cli/ui\");\nconst serverEnv = { ...process.env, STUDIO_CONTENTS_DIR: contentsDir, PORT: String(port), HOSTNAME: \"0.0.0.0\" };\nconst serverProcess = resolveServerProcess(uiDir, port, serverEnv);\n\nif (!serverProcess) {\n console.error(\"Error: Studio UI server not found.\");\n console.error(\"The pre-built UI is not included in this installation.\");\n process.exit(1);\n}\n\nconsole.log(`Nextjs Studio v${version}`);\nconsole.log(`Contents: ${contentsDir}`);\nconsole.log(`Starting on http://localhost:${port}`);\n\nserverProcess.on(\"error\", (error) => {\n console.error(\"Failed to start server:\", error.message);\n process.exit(1);\n});\n\nserverProcess.on(\"close\", (code) => process.exit(code ?? 0));\n\nforwardSignals(serverProcess);\n","/**\n * @context Shared layer — constants at src/shared/constants.ts\n * @does Defines project-wide constants shared across core, CLI, and UI layers\n * @depends none\n * @do Add new shared constants here\n * @dont Import from CLI or UI; constants must be framework-agnostic\n */\n\nexport const CONTENTS_DIR = \"contents\";\nexport const CLI_PORT = 3030;\nexport const CONFIG_FILE = \"studio.config.ts\";\nexport const SUPPORTED_EXTENSIONS = [\".mdx\", \".json\"] as const;\nexport const COLLECTION_ORDER_FILE = \"collection.json\";\nexport const WATCHER_DEBOUNCE_MS = 5_000;\nexport const MEDIA_DIR = \"media\";\n\nexport const IMAGE_MIME_TYPES = [\n \"image/png\",\n \"image/jpeg\",\n \"image/gif\",\n \"image/webp\",\n \"image/svg+xml\",\n \"image/avif\",\n] as const;\n\nexport const VIDEO_MIME_TYPES = [\"video/mp4\", \"video/webm\", \"video/ogg\"] as const;\n\nexport const AUDIO_MIME_TYPES = [\n \"audio/mpeg\",\n \"audio/ogg\",\n \"audio/wav\",\n \"audio/webm\",\n \"audio/aac\",\n \"audio/flac\",\n] as const;\n\nexport const MEDIA_MIME_TYPES = [...IMAGE_MIME_TYPES, ...VIDEO_MIME_TYPES, ...AUDIO_MIME_TYPES] as const;\n\nexport const IMAGE_EXTENSIONS = [\".png\", \".jpg\", \".jpeg\", \".gif\", \".webp\", \".svg\", \".avif\"] as const;\nexport const VIDEO_EXTENSIONS = [\".mp4\", \".webm\", \".ogv\"] as const;\nexport const AUDIO_EXTENSIONS = [\".mp3\", \".ogg\", \".wav\", \".m4a\", \".aac\", \".flac\"] as const;\n","/**\n * @context CLI layer — filesystem adapter at src/cli/adapters/fs-adapter.ts\n * @does Implements IFsAdapter; abstracts all file read/write/list operations behind a single interface\n * @depends src/shared/types.ts, src/shared/constants.ts, src/shared/fs-adapter.interface.ts\n * @do Add new I/O operations here; all file access must go through this adapter\n * @dont Import UI components, run HTTP requests, or contain business logic\n */\n\nimport fs from \"node:fs/promises\";\nimport path from \"node:path\";\nimport type { Dirent } from \"node:fs\";\nimport type { FileInfo, DirectoryFileEntry } from \"../../shared/types.js\";\nimport type { IFsAdapter } from \"../../shared/fs-adapter.interface.js\";\nimport { SUPPORTED_EXTENSIONS } from \"../../shared/constants.js\";\n\nexport class FsAdapter implements IFsAdapter {\n private readonly basePath: string;\n\n constructor(basePath: string) {\n this.basePath = path.resolve(basePath);\n }\n\n private resolve(...segments: string[]): string {\n return path.resolve(this.basePath, ...segments);\n }\n\n async readFile(filePath: string): Promise<string> {\n return fs.readFile(this.resolve(filePath), \"utf-8\");\n }\n\n async writeFile(filePath: string, content: string): Promise<void> {\n const fullPath = this.resolve(filePath);\n await fs.mkdir(path.dirname(fullPath), { recursive: true });\n await fs.writeFile(fullPath, content, \"utf-8\");\n }\n\n async deleteFile(filePath: string): Promise<void> {\n await fs.unlink(this.resolve(filePath));\n }\n\n async exists(filePath: string): Promise<boolean> {\n try {\n await fs.access(this.resolve(filePath));\n return true;\n } catch {\n return false;\n }\n }\n\n async getStats(filePath: string): Promise<FileInfo> {\n const fullPath = this.resolve(filePath);\n const stats = await fs.stat(fullPath);\n return { path: filePath, size: stats.size, modifiedAt: stats.mtime };\n }\n\n async listFiles(dirPath: string, extensions?: readonly string[]): Promise<string[]> {\n const fullPath = this.resolve(dirPath);\n const filterExts = extensions ?? SUPPORTED_EXTENSIONS;\n\n let entries: Dirent[];\n try {\n entries = await fs.readdir(fullPath, { withFileTypes: true });\n } catch {\n return [];\n }\n\n return entries\n .filter((entry) => entry.isFile() && filterExts.some((ext) => entry.name.endsWith(ext)))\n .map((entry) => this.join(dirPath, entry.name));\n }\n\n async listDirectories(dirPath: string): Promise<string[]> {\n const fullPath = this.resolve(dirPath);\n\n let entries: Dirent[];\n try {\n entries = await fs.readdir(fullPath, { withFileTypes: true });\n } catch {\n return [];\n }\n\n return entries\n .filter((entry) => entry.isDirectory())\n .map((entry) => this.join(dirPath, entry.name));\n }\n\n async readBuffer(filePath: string): Promise<Buffer> {\n return fs.readFile(this.resolve(filePath));\n }\n\n async writeBuffer(filePath: string, data: Buffer): Promise<void> {\n const fullPath = this.resolve(filePath);\n await fs.mkdir(path.dirname(fullPath), { recursive: true });\n await fs.writeFile(fullPath, data);\n }\n\n async listAllFiles(dirPath: string): Promise<DirectoryFileEntry[]> {\n const fullPath = this.resolve(dirPath);\n\n let entries: Dirent[];\n try {\n entries = await fs.readdir(fullPath, { withFileTypes: true });\n } catch {\n return [];\n }\n\n const results: DirectoryFileEntry[] = [];\n for (const entry of entries) {\n if (!entry.isFile()) continue;\n const relativePath = this.join(dirPath, entry.name);\n const stats = await fs.stat(this.resolve(relativePath));\n results.push({ name: entry.name, relativePath, size: stats.size, modifiedAt: stats.mtime });\n }\n return results;\n }\n\n join(...segments: string[]): string {\n return path.join(...segments);\n }\n\n basename(filePath: string): string {\n return path.basename(filePath);\n }\n\n extname(filePath: string): string {\n return path.extname(filePath);\n }\n\n relative(from: string, to: string): string {\n return path.relative(from, to);\n }\n\n normalizeSlug(relativePath: string, ext: string): string {\n return relativePath.replace(ext, \"\").split(path.sep).join(\"/\");\n }\n}\n","/**\n * @context Core layer — content indexer at src/core/indexer.ts\n * @does Scans the contents directory, parses MDX/JSON files, and builds an in-memory index\n * @depends src/shared/types.ts, src/shared/constants.ts, src/shared/fs-adapter.interface.ts, src/core/parsers/, src/core/schema-inferrer.ts\n * @do Add new file type handling here; extend indexCollection for new collection behaviors\n * @dont Import from CLI or UI; instantiate FsAdapter; access the filesystem directly\n */\n\nimport slugify from \"@sindresorhus/slugify\";\nimport type { CollectionSchema } from \"../shared/fields.js\";\nimport type { ContentEntry, Collection, StudioConfig } from \"../shared/types.js\";\nimport type { IFsAdapter } from \"../shared/fs-adapter.interface.js\";\nimport { COLLECTION_ORDER_FILE } from \"../shared/constants.js\";\nimport { parseMdx } from \"./parsers/parser-mdx.js\";\nimport { parseJson } from \"./parsers/parser-json.js\";\nimport { inferSchema } from \"./schema-inferrer.js\";\n\nexport class ContentIndex {\n private readonly entries = new Map<string, ContentEntry[]>();\n private readonly collections = new Map<string, Collection>();\n private readonly fs: IFsAdapter;\n\n constructor(fsAdapter: IFsAdapter) {\n this.fs = fsAdapter;\n }\n\n async build(config?: StudioConfig): Promise<void> {\n this.clear();\n const dirs = await this.fs.listDirectories(\".\");\n\n for (const dir of dirs) {\n const dirName = this.fs.basename(dir);\n const collectionName = slugify(dirName);\n const collectionConfig = config?.collections?.[collectionName];\n await this.indexCollection(dirName, collectionName, collectionConfig?.schema);\n }\n }\n\n getCollection(name: string): ContentEntry[] {\n return this.entries.get(name) ?? [];\n }\n\n getCollections(): Collection[] {\n return Array.from(this.collections.values());\n }\n\n clear(): void {\n this.entries.clear();\n this.collections.clear();\n }\n\n private async indexCollection(\n dirName: string,\n collectionName: string,\n manualSchema?: CollectionSchema,\n ): Promise<void> {\n const entries: ContentEntry[] = [];\n await this.scanDir(dirName, collectionName, dirName, entries);\n\n const orderPath = this.fs.join(dirName, COLLECTION_ORDER_FILE);\n const ordering = await this.readOrdering(orderPath);\n if (ordering) {\n this.applyOrdering(entries, ordering);\n }\n\n const schema = manualSchema ?? inferSchema(entries, collectionName);\n\n this.entries.set(collectionName, entries);\n this.collections.set(collectionName, {\n name: collectionName,\n type: this.detectCollectionType(entries),\n count: entries.length,\n basePath: dirName,\n schema,\n });\n }\n\n private async scanDir(\n dirName: string,\n collectionName: string,\n dirPath: string,\n entries: ContentEntry[],\n ): Promise<void> {\n const subDirs = await this.fs.listDirectories(dirPath);\n for (const subDir of subDirs) {\n await this.scanDir(dirName, collectionName, subDir, entries);\n }\n\n const files = await this.fs.listFiles(dirPath);\n for (const filePath of files) {\n const fileName = this.fs.basename(filePath);\n if (fileName === COLLECTION_ORDER_FILE) continue;\n\n const ext = this.fs.extname(fileName);\n const content = await this.fs.readFile(filePath);\n const relativePath = this.fs.relative(dirName, filePath);\n const slug = this.fs\n .normalizeSlug(relativePath, ext)\n .split(\"/\")\n .map((segment) => slugify(segment))\n .join(\"/\");\n\n if (ext === \".mdx\") {\n entries.push(this.buildMdxEntry(collectionName, slug, content));\n } else if (ext === \".json\") {\n entries.push(...this.buildJsonEntries(collectionName, slug, content));\n }\n }\n }\n\n private buildMdxEntry(collectionName: string, slug: string, content: string): ContentEntry {\n const parsed = parseMdx(content);\n return {\n collection: collectionName,\n slug,\n path: `/${collectionName}/${slug}`,\n body: parsed.body,\n data: parsed.data,\n };\n }\n\n private buildJsonEntries(collectionName: string, slug: string, content: string): ContentEntry[] {\n const parsed = parseJson(content);\n\n if (parsed.type === \"json-array\") {\n return parsed.entries.map((data, index) => {\n const entrySlug =\n typeof data[\"slug\"] === \"string\" ? slugify(data[\"slug\"]) : `${slug}/${index}`;\n return {\n collection: collectionName,\n slug: entrySlug,\n path: `/${collectionName}/${entrySlug}`,\n data,\n };\n });\n }\n\n return [{ collection: collectionName, slug, path: `/${collectionName}/${slug}`, data: parsed.data }];\n }\n\n private async readOrdering(orderPath: string): Promise<string[] | null> {\n if (!(await this.fs.exists(orderPath))) return null;\n\n try {\n const content = await this.fs.readFile(orderPath);\n const parsed: unknown = JSON.parse(content);\n if (Array.isArray(parsed)) return parsed as string[];\n } catch (error) {\n console.warn(`[Nextjs Studio] Failed to parse ordering file: ${orderPath}`, error);\n }\n return null;\n }\n\n private applyOrdering(entries: ContentEntry[], ordering: string[]): void {\n const orderMap = new Map(ordering.map((slug, index) => [slug, index]));\n entries.sort((a, b) => {\n const aIndex = orderMap.get(a.slug) ?? Infinity;\n const bIndex = orderMap.get(b.slug) ?? Infinity;\n return aIndex - bIndex;\n });\n }\n\n private detectCollectionType(entries: ContentEntry[]): Collection[\"type\"] {\n if (entries.length === 0) return \"mdx\";\n const first = entries[0];\n if (first.body !== undefined) return \"mdx\";\n if (entries.length === 1 && !first.slug.includes(\"/\")) return \"json-object\";\n return \"json-array\";\n }\n}\n","/**\n * @context Core layer — MDX parser/serializer at src/core/parsers/parser-mdx.ts\n * @does Parses .mdx content into frontmatter + body, and serializes them back to MDX strings\n * @depends none (gray-matter is an external dep)\n * @do Add MDX transform steps here; both parse and serialize live here intentionally\n * @dont Access the filesystem; import from CLI or UI; handle JSON content\n */\n\nimport matter from \"gray-matter\";\n\nexport interface ParsedMdx {\n data: Record<string, unknown>;\n body: string;\n}\n\nexport function parseMdx(content: string): ParsedMdx {\n const { data, content: body } = matter(content);\n return { data, body: body.trim() };\n}\n\nexport function serializeMdx(data: Record<string, unknown>, body: string): string {\n return matter.stringify(body, data);\n}\n","/**\n * @context Core layer — JSON parser at src/core/parsers/parser-json.ts\n * @does Parses JSON content strings into typed ParsedJson results (array or object)\n * @depends none\n * @do Extend ParsedJson variants here if new JSON structures are supported\n * @dont Access the filesystem; import from CLI or UI; contain serialization logic\n */\n\nexport interface ParsedJsonArray {\n type: \"json-array\";\n entries: Record<string, unknown>[];\n}\n\nexport interface ParsedJsonObject {\n type: \"json-object\";\n data: Record<string, unknown>;\n}\n\nexport type ParsedJson = ParsedJsonArray | ParsedJsonObject;\n\nexport function parseJson(content: string): ParsedJson {\n const parsed: unknown = JSON.parse(content);\n\n if (Array.isArray(parsed)) {\n return {\n type: \"json-array\",\n entries: parsed as Record<string, unknown>[],\n };\n }\n\n if (typeof parsed === \"object\" && parsed !== null) {\n return {\n type: \"json-object\",\n data: parsed as Record<string, unknown>,\n };\n }\n\n throw new Error(\"JSON content must be an array or object\");\n}\n","/**\n * @context Core layer — schema inferrer at src/core/schema-inferrer.ts\n * @does Infers a CollectionSchema from actual content entries when no manual schema is defined\n * @depends src/shared/types.ts, src/shared/fields.ts\n * @do Add new type detection heuristics here (e.g. color, phone)\n * @dont Import from CLI or UI; access the filesystem; perform I/O\n */\n\nimport type { ContentEntry } from \"../shared/types.js\";\nimport type { CollectionSchema, FieldDefinition, SelectOption } from \"../shared/fields.js\";\n\n// Value detector patterns\nconst RE_ISO_DATE = /^\\d{4}-\\d{2}-\\d{2}$/;\nconst RE_ISO_DATETIME =\n /^\\d{4}-\\d{2}-\\d{2}[T ]\\d{2}:\\d{2}(:\\d{2}(\\.\\d+)?)?(Z|[+-]\\d{2}:?\\d{2})?$/;\nconst RE_EMAIL = /^[^\\s@]+@[^\\s@]+\\.[^\\s@]+$/;\nconst RE_URL = /^https?:\\/\\/.+/;\nconst LONG_TEXT_THRESHOLD = 200;\n\nfunction isISODate(value: string): boolean {\n return RE_ISO_DATE.test(value);\n}\n\nfunction isISODateTime(value: string): boolean {\n return RE_ISO_DATETIME.test(value);\n}\n\nfunction isEmail(value: string): boolean {\n return RE_EMAIL.test(value);\n}\n\nfunction isUrl(value: string): boolean {\n return RE_URL.test(value);\n}\n\nfunction inferStringField(name: string, strings: string[]): FieldDefinition {\n if (strings.every(isEmail)) return { name, type: \"email\" };\n if (strings.every(isUrl)) return { name, type: \"url\" };\n if (strings.every(isISODateTime)) return { name, type: \"date\", includeTime: true };\n if (strings.every(isISODate)) return { name, type: \"date\" };\n\n const isLong = strings.some((s) => s.length > LONG_TEXT_THRESHOLD || s.includes(\"\\n\"));\n return { name, type: isLong ? \"long-text\" : \"text\" };\n}\n\nfunction inferArrayField(name: string, items: unknown[]): FieldDefinition {\n if (items.length === 0) return { name, type: \"array\", itemFields: [] };\n\n if (items.every((item) => typeof item === \"string\")) {\n const unique = [...new Set(items as string[])].slice(0, 50);\n const options: SelectOption[] = unique.map((v) => ({ label: v, value: v }));\n return { name, type: \"multi-select\", options };\n }\n\n if (items.every((item) => typeof item === \"object\" && item !== null && !Array.isArray(item))) {\n return { name, type: \"array\", itemFields: inferFields(items as Record<string, unknown>[]) };\n }\n\n return { name, type: \"array\", itemFields: [] };\n}\n\nfunction inferFieldDefinition(name: string, values: unknown[]): FieldDefinition {\n const present = values.filter((v) => v !== null && v !== undefined);\n\n if (present.length === 0) return { name, type: \"text\" };\n if (present.every((v) => typeof v === \"boolean\")) return { name, type: \"boolean\" };\n\n if (present.every((v) => typeof v === \"number\")) {\n const format = present.every((v) => Number.isInteger(v)) ? \"integer\" : \"decimal\";\n return { name, type: \"number\", format };\n }\n\n if (present.every((v) => typeof v === \"string\")) {\n return inferStringField(name, present as string[]);\n }\n\n if (present.every((v) => Array.isArray(v))) {\n return inferArrayField(name, (present as unknown[][]).flat());\n }\n\n if (present.every((v) => typeof v === \"object\" && v !== null && !Array.isArray(v))) {\n return { name, type: \"object\", fields: inferFields(present as Record<string, unknown>[]) };\n }\n\n return { name, type: \"text\" };\n}\n\nfunction inferFields(rows: Record<string, unknown>[]): FieldDefinition[] {\n const keySet = new Set<string>(rows.flatMap((row) => Object.keys(row)));\n return Array.from(keySet).map((key) => inferFieldDefinition(key, rows.map((row) => row[key])));\n}\n\n/**\n * Infer a `CollectionSchema` from the data of a set of content entries.\n *\n * The result is a best-effort approximation — string fields that look like\n * emails, URLs, or ISO dates get the correct semantic type. Everything else\n * falls back to `text`.\n */\nexport function inferSchema(entries: ContentEntry[], collectionName: string): CollectionSchema {\n const rows = entries.map((entry) => entry.data as Record<string, unknown>);\n return { collection: collectionName, fields: inferFields(rows) };\n}\n","/**\n * @context Core layer — content store at src/core/content-store.ts\n * @does Manages a singleton ContentIndex; exposes loadContent() and getStore() for consumers\n * @depends src/core/indexer.ts, src/shared/fs-adapter.interface.ts, src/shared/types.ts\n * @do Use this as the single access point for in-memory indexed content\n * @dont Import from CLI or UI; instantiate FsAdapter here; contain parsing or I/O logic\n */\n\nimport type { IFsAdapter } from \"../shared/fs-adapter.interface.js\";\nimport type { StudioConfig } from \"../shared/types.js\";\nimport { ContentIndex } from \"./indexer.js\";\n\nlet store: ContentIndex | null = null;\n\nexport function getStore(): ContentIndex {\n if (!store) {\n throw new Error(\"Content not loaded. Call loadContent() before querying.\");\n }\n return store;\n}\n\nexport async function loadContent(\n fsAdapter: IFsAdapter,\n config?: StudioConfig,\n): Promise<ContentIndex> {\n const index = new ContentIndex(fsAdapter);\n await index.build(config);\n store = index;\n return index;\n}\n","/**\n * @context Core layer — type generator at src/core/type-generator.ts\n * @does Converts CollectionSchema definitions into TypeScript declaration strings for .d.ts output\n * @depends src/shared/fields.ts\n * @do Add new field-to-type mappings here as new field types are introduced\n * @dont Import from CLI or UI; access the filesystem; perform I/O\n */\n\nimport type { FieldDefinition, CollectionSchema } from \"../shared/fields.js\";\n\nfunction indent(code: string, spaces = 2): string {\n return code\n .split(\"\\n\")\n .map((line) => (line.trim() === \"\" ? \"\" : \" \".repeat(spaces) + line))\n .join(\"\\n\");\n}\n\nfunction toPascalCase(str: string): string {\n return str\n .replace(/[-_\\s]+(.)/g, (_, c: string) => c.toUpperCase())\n .replace(/^(.)/, (_, c: string) => c.toUpperCase());\n}\n\nfunction fieldToTsType(field: FieldDefinition): string {\n switch (field.type) {\n case \"text\":\n case \"long-text\":\n return \"string\";\n\n case \"email\":\n return \"Email\";\n\n case \"url\":\n return \"HttpUrl\";\n\n case \"media\":\n return \"MediaPath\";\n\n case \"id\":\n return \"ID\";\n\n case \"slug\":\n return \"Slug\";\n\n case \"date\":\n return field.includeTime ? \"Date\" : \"ISODate\";\n\n case \"created-time\":\n case \"updated-time\":\n return \"Date\";\n\n case \"number\":\n return \"number\";\n\n case \"boolean\":\n return \"boolean\";\n\n case \"select\":\n case \"status\": {\n const values = field.options.map((o) => JSON.stringify(o.value));\n return values.length > 0 ? values.join(\" | \") : \"string\";\n }\n\n case \"multi-select\": {\n const values = field.options.map((o) => JSON.stringify(o.value));\n const union = values.length > 0 ? values.join(\" | \") : \"string\";\n return `Array<${union}>`;\n }\n\n case \"object\":\n return generateObjectType(field.fields);\n\n case \"array\":\n return `Array<${generateObjectType(field.itemFields)}>`;\n\n case \"relation\":\n return field.multiple === true ? \"ID[]\" : \"ID\";\n\n case \"formula\":\n if (field.resultType === \"number\") return \"number\";\n if (field.resultType === \"boolean\") return \"boolean\";\n return \"string\";\n\n default:\n return \"unknown\";\n }\n}\n\nfunction generateObjectType(fields: FieldDefinition[]): string {\n if (fields.length === 0) return \"Record<string, unknown>\";\n\n const lines = fields.map((field) => {\n const optional = field.required === false ? \"?\" : \"\";\n const tsType = fieldToTsType(field);\n const comment = field.description ? `/** ${field.description} */\\n` : \"\";\n return `${comment}${field.name}${optional}: ${tsType};`;\n });\n\n return `{\\n${indent(lines.join(\"\\n\"))}\\n}`;\n}\n\nexport function generateInterfaceForSchema(schema: CollectionSchema): string {\n const name = toPascalCase(schema.collection) + \"Entry\";\n const label = schema.label ?? schema.collection;\n const body = generateObjectType(schema.fields);\n return `/** Data shape for the \"${label}\" collection. */\\nexport interface ${name} ${body}`;\n}\n\n/**\n * Generate a complete TypeScript declaration file for all provided schemas.\n *\n * @example\n * ```ts\n * const code = generateCollectionTypes([blogSchema, authorSchema]);\n * await fs.writeFile(\".studio/types.d.ts\", code, \"utf-8\");\n * ```\n */\nexport function generateCollectionTypes(schemas: CollectionSchema[]): string {\n const banner = [\n \"// This file is auto-generated by nextjs-studio.\",\n \"// Do not edit manually — re-run `npx nextjs-studio --generate-types` to update.\",\n \"\",\n \"// Branded scalar types — structurally strings/numbers but semantically distinct.\",\n \"declare const __brand: unique symbol;\",\n \"type Brand<T, B extends string> = T & { readonly [__brand]: B };\",\n \"\",\n \"export type Email = Brand<string, 'Email'>;\",\n \"export type HttpUrl = Brand<string, 'HttpUrl'>;\",\n \"export type ISODate = Brand<string, 'ISODate'>;\",\n \"export type MediaPath = Brand<string, 'MediaPath'>;\",\n \"export type ID = Brand<string, 'ID'>;\",\n \"export type Slug = Brand<string, 'Slug'>;\",\n ].join(\"\\n\");\n\n const interfaces = schemas.map(generateInterfaceForSchema).join(\"\\n\\n\");\n\n const collectionRegistry = [\n \"// Augment the nextjs-studio module so queryCollection() is fully typed.\",\n \"declare module 'nextjs-studio' {\",\n \" interface CollectionTypeMap {\",\n schemas.map((schema) => ` ${JSON.stringify(schema.collection)}: ${toPascalCase(schema.collection)}Entry;`).join(\"\\n\"),\n \" }\",\n \"}\",\n ].join(\"\\n\");\n\n return [banner, interfaces, collectionRegistry].join(\"\\n\\n\") + \"\\n\";\n}\n","{\n \"name\": \"nextjs-studio\",\n \"version\": \"0.4.0\",\n \"description\": \"A Git-based, local-first CMS for Next.js projects\",\n \"keywords\": [\n \"nextjs\",\n \"cms\",\n \"mdx\",\n \"content\",\n \"studio\",\n \"static-site\",\n \"local-first\"\n ],\n \"homepage\": \"https://github.com/TiagoDanin/Nextjs-Studio\",\n \"repository\": {\n \"type\": \"git\",\n \"url\": \"git+https://github.com/TiagoDanin/Nextjs-Studio.git\"\n },\n \"license\": \"MIT\",\n \"author\": \"Tiago Danin\",\n \"type\": \"module\",\n \"exports\": {\n \".\": {\n \"types\": \"./dist/core/index.d.ts\",\n \"import\": \"./dist/core/index.js\"\n }\n },\n \"main\": \"./dist/core/index.js\",\n \"types\": \"./dist/core/index.d.ts\",\n \"bin\": {\n \"nextjs-studio\": \"dist/bin/nextjs-studio.js\"\n },\n \"files\": [\n \"dist\",\n \"README.md\",\n \"LICENSE\"\n ],\n \"scripts\": {\n \"dev\": \"tsx src/bin/nextjs-studio.ts --dir example/contents\",\n \"studio:dev\": \"cross-env STUDIO_CONTENTS_DIR=example/contents next dev --port 3030 --webpack src/cli/ui\",\n \"studio:build\": \"next build --webpack src/cli/ui\",\n \"build\": \"tsup && yarn studio:build\",\n \"lint\": \"eslint src/\",\n \"type-check\": \"tsc --noEmit\",\n \"test\": \"vitest run\",\n \"test:watch\": \"vitest\"\n },\n \"engines\": {\n \"node\": \">=22.10.0\"\n },\n \"packageManager\": \"yarn@4.6.0\",\n \"devDependencies\": {\n \"@radix-ui/react-collapsible\": \"^1.1.12\",\n \"@radix-ui/react-label\": \"^2.1.8\",\n \"@radix-ui/react-switch\": \"^1.2.6\",\n \"@tailwindcss/postcss\": \"^4.1.18\",\n \"@tanstack/react-table\": \"^8.21.3\",\n \"@tiptap/extension-bubble-menu\": \"^3.20.0\",\n \"@tiptap/extension-code-block-lowlight\": \"^3.20.0\",\n \"@tiptap/extension-file-handler\": \"^3.20.0\",\n \"@tiptap/extension-image\": \"^3.20.0\",\n \"@tiptap/extension-link\": \"^3.20.0\",\n \"@tiptap/extension-placeholder\": \"^3.20.0\",\n \"@tiptap/react\": \"^3.20.0\",\n \"@tiptap/starter-kit\": \"^3.20.0\",\n \"@tiptap/suggestion\": \"^3.20.0\",\n \"@types/lodash-es\": \"^4.17.12\",\n \"@types/node\": \"^25.2.3\",\n \"@types/react\": \"^19\",\n \"@types/react-dom\": \"^19\",\n \"class-variance-authority\": \"^0.7.1\",\n \"clsx\": \"^2.1.1\",\n \"cross-env\": \"^10.1.0\",\n \"eslint\": \"^10.0.0\",\n \"lowlight\": \"^3.3.0\",\n \"lucide-react\": \"^0.574.0\",\n \"mermaid\": \"^11.6.0\",\n \"next\": \"^16.1.6\",\n \"next-themes\": \"^0.4.6\",\n \"react\": \"^19.2.4\",\n \"react-dom\": \"^19.2.4\",\n \"tailwind-merge\": \"^3.4.1\",\n \"tailwindcss\": \"^4.1.18\",\n \"tippy.js\": \"^6.3.7\",\n \"tiptap-extension-global-drag-handle\": \"^0.1.18\",\n \"tiptap-markdown\": \"^0.9.0\",\n \"tsup\": \"^8.5.1\",\n \"tsx\": \"^4.21.0\",\n \"typescript\": \"^5.9.3\",\n \"vitest\": \"^4.0.18\",\n \"zustand\": \"^5.0.11\"\n },\n \"dependencies\": {\n \"@sindresorhus/slugify\": \"^3.0.0\",\n \"chokidar\": \"^5.0.0\",\n \"commander\": \"^14.0.3\",\n \"gray-matter\": \"^4.0.3\",\n \"lodash-es\": \"^4.17.23\"\n }\n}"],"mappings":";;;AAUA,SAAS,kBAAkB;AAC3B,OAAOA,SAAQ;AACf,OAAOC,WAAU;AACjB,SAAS,aAAgC;AACzC,SAAS,eAAe;;;ACNjB,IAAM,eAAe;AACrB,IAAM,WAAW;AAEjB,IAAM,uBAAuB,CAAC,QAAQ,OAAO;AAC7C,IAAM,wBAAwB;AAI9B,IAAM,mBAAmB;AAAA,EAC9B;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AACF;AAEO,IAAM,mBAAmB,CAAC,aAAa,cAAc,WAAW;AAEhE,IAAM,mBAAmB;AAAA,EAC9B;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AACF;AAEO,IAAM,mBAAmB,CAAC,GAAG,kBAAkB,GAAG,kBAAkB,GAAG,gBAAgB;;;AC5B9F,OAAO,QAAQ;AACf,OAAO,UAAU;AAMV,IAAM,YAAN,MAAsC;AAAA,EAC1B;AAAA,EAEjB,YAAY,UAAkB;AAC5B,SAAK,WAAW,KAAK,QAAQ,QAAQ;AAAA,EACvC;AAAA,EAEQ,WAAW,UAA4B;AAC7C,WAAO,KAAK,QAAQ,KAAK,UAAU,GAAG,QAAQ;AAAA,EAChD;AAAA,EAEA,MAAM,SAAS,UAAmC;AAChD,WAAO,GAAG,SAAS,KAAK,QAAQ,QAAQ,GAAG,OAAO;AAAA,EACpD;AAAA,EAEA,MAAM,UAAU,UAAkB,SAAgC;AAChE,UAAM,WAAW,KAAK,QAAQ,QAAQ;AACtC,UAAM,GAAG,MAAM,KAAK,QAAQ,QAAQ,GAAG,EAAE,WAAW,KAAK,CAAC;AAC1D,UAAM,GAAG,UAAU,UAAU,SAAS,OAAO;AAAA,EAC/C;AAAA,EAEA,MAAM,WAAW,UAAiC;AAChD,UAAM,GAAG,OAAO,KAAK,QAAQ,QAAQ,CAAC;AAAA,EACxC;AAAA,EAEA,MAAM,OAAO,UAAoC;AAC/C,QAAI;AACF,YAAM,GAAG,OAAO,KAAK,QAAQ,QAAQ,CAAC;AACtC,aAAO;AAAA,IACT,QAAQ;AACN,aAAO;AAAA,IACT;AAAA,EACF;AAAA,EAEA,MAAM,SAAS,UAAqC;AAClD,UAAM,WAAW,KAAK,QAAQ,QAAQ;AACtC,UAAM,QAAQ,MAAM,GAAG,KAAK,QAAQ;AACpC,WAAO,EAAE,MAAM,UAAU,MAAM,MAAM,MAAM,YAAY,MAAM,MAAM;AAAA,EACrE;AAAA,EAEA,MAAM,UAAU,SAAiB,YAAmD;AAClF,UAAM,WAAW,KAAK,QAAQ,OAAO;AACrC,UAAM,aAAa,cAAc;AAEjC,QAAI;AACJ,QAAI;AACF,gBAAU,MAAM,GAAG,QAAQ,UAAU,EAAE,eAAe,KAAK,CAAC;AAAA,IAC9D,QAAQ;AACN,aAAO,CAAC;AAAA,IACV;AAEA,WAAO,QACJ,OAAO,CAAC,UAAU,MAAM,OAAO,KAAK,WAAW,KAAK,CAAC,QAAQ,MAAM,KAAK,SAAS,GAAG,CAAC,CAAC,EACtF,IAAI,CAAC,UAAU,KAAK,KAAK,SAAS,MAAM,IAAI,CAAC;AAAA,EAClD;AAAA,EAEA,MAAM,gBAAgB,SAAoC;AACxD,UAAM,WAAW,KAAK,QAAQ,OAAO;AAErC,QAAI;AACJ,QAAI;AACF,gBAAU,MAAM,GAAG,QAAQ,UAAU,EAAE,eAAe,KAAK,CAAC;AAAA,IAC9D,QAAQ;AACN,aAAO,CAAC;AAAA,IACV;AAEA,WAAO,QACJ,OAAO,CAAC,UAAU,MAAM,YAAY,CAAC,EACrC,IAAI,CAAC,UAAU,KAAK,KAAK,SAAS,MAAM,IAAI,CAAC;AAAA,EAClD;AAAA,EAEA,MAAM,WAAW,UAAmC;AAClD,WAAO,GAAG,SAAS,KAAK,QAAQ,QAAQ,CAAC;AAAA,EAC3C;AAAA,EAEA,MAAM,YAAY,UAAkB,MAA6B;AAC/D,UAAM,WAAW,KAAK,QAAQ,QAAQ;AACtC,UAAM,GAAG,MAAM,KAAK,QAAQ,QAAQ,GAAG,EAAE,WAAW,KAAK,CAAC;AAC1D,UAAM,GAAG,UAAU,UAAU,IAAI;AAAA,EACnC;AAAA,EAEA,MAAM,aAAa,SAAgD;AACjE,UAAM,WAAW,KAAK,QAAQ,OAAO;AAErC,QAAI;AACJ,QAAI;AACF,gBAAU,MAAM,GAAG,QAAQ,UAAU,EAAE,eAAe,KAAK,CAAC;AAAA,IAC9D,QAAQ;AACN,aAAO,CAAC;AAAA,IACV;AAEA,UAAM,UAAgC,CAAC;AACvC,eAAW,SAAS,SAAS;AAC3B,UAAI,CAAC,MAAM,OAAO,EAAG;AACrB,YAAM,eAAe,KAAK,KAAK,SAAS,MAAM,IAAI;AAClD,YAAM,QAAQ,MAAM,GAAG,KAAK,KAAK,QAAQ,YAAY,CAAC;AACtD,cAAQ,KAAK,EAAE,MAAM,MAAM,MAAM,cAAc,MAAM,MAAM,MAAM,YAAY,MAAM,MAAM,CAAC;AAAA,IAC5F;AACA,WAAO;AAAA,EACT;AAAA,EAEA,QAAQ,UAA4B;AAClC,WAAO,KAAK,KAAK,GAAG,QAAQ;AAAA,EAC9B;AAAA,EAEA,SAAS,UAA0B;AACjC,WAAO,KAAK,SAAS,QAAQ;AAAA,EAC/B;AAAA,EAEA,QAAQ,UAA0B;AAChC,WAAO,KAAK,QAAQ,QAAQ;AAAA,EAC9B;AAAA,EAEA,SAAS,MAAc,IAAoB;AACzC,WAAO,KAAK,SAAS,MAAM,EAAE;AAAA,EAC/B;AAAA,EAEA,cAAc,cAAsB,KAAqB;AACvD,WAAO,aAAa,QAAQ,KAAK,EAAE,EAAE,MAAM,KAAK,GAAG,EAAE,KAAK,GAAG;AAAA,EAC/D;AACF;;;AC/HA,OAAO,aAAa;;;ACApB,OAAO,YAAY;AAOZ,SAAS,SAAS,SAA4B;AACnD,QAAM,EAAE,MAAM,SAAS,KAAK,IAAI,OAAO,OAAO;AAC9C,SAAO,EAAE,MAAM,MAAM,KAAK,KAAK,EAAE;AACnC;;;ACEO,SAAS,UAAU,SAA6B;AACrD,QAAM,SAAkB,KAAK,MAAM,OAAO;AAE1C,MAAI,MAAM,QAAQ,MAAM,GAAG;AACzB,WAAO;AAAA,MACL,MAAM;AAAA,MACN,SAAS;AAAA,IACX;AAAA,EACF;AAEA,MAAI,OAAO,WAAW,YAAY,WAAW,MAAM;AACjD,WAAO;AAAA,MACL,MAAM;AAAA,MACN,MAAM;AAAA,IACR;AAAA,EACF;AAEA,QAAM,IAAI,MAAM,yCAAyC;AAC3D;;;AC1BA,IAAM,cAAc;AACpB,IAAM,kBACJ;AACF,IAAM,WAAW;AACjB,IAAM,SAAS;AACf,IAAM,sBAAsB;AAE5B,SAAS,UAAU,OAAwB;AACzC,SAAO,YAAY,KAAK,KAAK;AAC/B;AAEA,SAAS,cAAc,OAAwB;AAC7C,SAAO,gBAAgB,KAAK,KAAK;AACnC;AAEA,SAAS,QAAQ,OAAwB;AACvC,SAAO,SAAS,KAAK,KAAK;AAC5B;AAEA,SAAS,MAAM,OAAwB;AACrC,SAAO,OAAO,KAAK,KAAK;AAC1B;AAEA,SAAS,iBAAiB,MAAc,SAAoC;AAC1E,MAAI,QAAQ,MAAM,OAAO,EAAG,QAAO,EAAE,MAAM,MAAM,QAAQ;AACzD,MAAI,QAAQ,MAAM,KAAK,EAAG,QAAO,EAAE,MAAM,MAAM,MAAM;AACrD,MAAI,QAAQ,MAAM,aAAa,EAAG,QAAO,EAAE,MAAM,MAAM,QAAQ,aAAa,KAAK;AACjF,MAAI,QAAQ,MAAM,SAAS,EAAG,QAAO,EAAE,MAAM,MAAM,OAAO;AAE1D,QAAM,SAAS,QAAQ,KAAK,CAAC,MAAM,EAAE,SAAS,uBAAuB,EAAE,SAAS,IAAI,CAAC;AACrF,SAAO,EAAE,MAAM,MAAM,SAAS,cAAc,OAAO;AACrD;AAEA,SAAS,gBAAgB,MAAc,OAAmC;AACxE,MAAI,MAAM,WAAW,EAAG,QAAO,EAAE,MAAM,MAAM,SAAS,YAAY,CAAC,EAAE;AAErE,MAAI,MAAM,MAAM,CAAC,SAAS,OAAO,SAAS,QAAQ,GAAG;AACnD,UAAM,SAAS,CAAC,GAAG,IAAI,IAAI,KAAiB,CAAC,EAAE,MAAM,GAAG,EAAE;AAC1D,UAAM,UAA0B,OAAO,IAAI,CAAC,OAAO,EAAE,OAAO,GAAG,OAAO,EAAE,EAAE;AAC1E,WAAO,EAAE,MAAM,MAAM,gBAAgB,QAAQ;AAAA,EAC/C;AAEA,MAAI,MAAM,MAAM,CAAC,SAAS,OAAO,SAAS,YAAY,SAAS,QAAQ,CAAC,MAAM,QAAQ,IAAI,CAAC,GAAG;AAC5F,WAAO,EAAE,MAAM,MAAM,SAAS,YAAY,YAAY,KAAkC,EAAE;AAAA,EAC5F;AAEA,SAAO,EAAE,MAAM,MAAM,SAAS,YAAY,CAAC,EAAE;AAC/C;AAEA,SAAS,qBAAqB,MAAc,QAAoC;AAC9E,QAAM,UAAU,OAAO,OAAO,CAAC,MAAM,MAAM,QAAQ,MAAM,MAAS;AAElE,MAAI,QAAQ,WAAW,EAAG,QAAO,EAAE,MAAM,MAAM,OAAO;AACtD,MAAI,QAAQ,MAAM,CAAC,MAAM,OAAO,MAAM,SAAS,EAAG,QAAO,EAAE,MAAM,MAAM,UAAU;AAEjF,MAAI,QAAQ,MAAM,CAAC,MAAM,OAAO,MAAM,QAAQ,GAAG;AAC/C,UAAM,SAAS,QAAQ,MAAM,CAAC,MAAM,OAAO,UAAU,CAAC,CAAC,IAAI,YAAY;AACvE,WAAO,EAAE,MAAM,MAAM,UAAU,OAAO;AAAA,EACxC;AAEA,MAAI,QAAQ,MAAM,CAAC,MAAM,OAAO,MAAM,QAAQ,GAAG;AAC/C,WAAO,iBAAiB,MAAM,OAAmB;AAAA,EACnD;AAEA,MAAI,QAAQ,MAAM,CAAC,MAAM,MAAM,QAAQ,CAAC,CAAC,GAAG;AAC1C,WAAO,gBAAgB,MAAO,QAAwB,KAAK,CAAC;AAAA,EAC9D;AAEA,MAAI,QAAQ,MAAM,CAAC,MAAM,OAAO,MAAM,YAAY,MAAM,QAAQ,CAAC,MAAM,QAAQ,CAAC,CAAC,GAAG;AAClF,WAAO,EAAE,MAAM,MAAM,UAAU,QAAQ,YAAY,OAAoC,EAAE;AAAA,EAC3F;AAEA,SAAO,EAAE,MAAM,MAAM,OAAO;AAC9B;AAEA,SAAS,YAAY,MAAoD;AACvE,QAAM,SAAS,IAAI,IAAY,KAAK,QAAQ,CAAC,QAAQ,OAAO,KAAK,GAAG,CAAC,CAAC;AACtE,SAAO,MAAM,KAAK,MAAM,EAAE,IAAI,CAAC,QAAQ,qBAAqB,KAAK,KAAK,IAAI,CAAC,QAAQ,IAAI,GAAG,CAAC,CAAC,CAAC;AAC/F;AASO,SAAS,YAAY,SAAyB,gBAA0C;AAC7F,QAAM,OAAO,QAAQ,IAAI,CAAC,UAAU,MAAM,IAA+B;AACzE,SAAO,EAAE,YAAY,gBAAgB,QAAQ,YAAY,IAAI,EAAE;AACjE;;;AHrFO,IAAM,eAAN,MAAmB;AAAA,EACP,UAAU,oBAAI,IAA4B;AAAA,EAC1C,cAAc,oBAAI,IAAwB;AAAA,EAC1C;AAAA,EAEjB,YAAY,WAAuB;AACjC,SAAK,KAAK;AAAA,EACZ;AAAA,EAEA,MAAM,MAAM,QAAsC;AAChD,SAAK,MAAM;AACX,UAAM,OAAO,MAAM,KAAK,GAAG,gBAAgB,GAAG;AAE9C,eAAW,OAAO,MAAM;AACtB,YAAM,UAAU,KAAK,GAAG,SAAS,GAAG;AACpC,YAAM,iBAAiB,QAAQ,OAAO;AACtC,YAAM,mBAAmB,QAAQ,cAAc,cAAc;AAC7D,YAAM,KAAK,gBAAgB,SAAS,gBAAgB,kBAAkB,MAAM;AAAA,IAC9E;AAAA,EACF;AAAA,EAEA,cAAc,MAA8B;AAC1C,WAAO,KAAK,QAAQ,IAAI,IAAI,KAAK,CAAC;AAAA,EACpC;AAAA,EAEA,iBAA+B;AAC7B,WAAO,MAAM,KAAK,KAAK,YAAY,OAAO,CAAC;AAAA,EAC7C;AAAA,EAEA,QAAc;AACZ,SAAK,QAAQ,MAAM;AACnB,SAAK,YAAY,MAAM;AAAA,EACzB;AAAA,EAEA,MAAc,gBACZ,SACA,gBACA,cACe;AACf,UAAM,UAA0B,CAAC;AACjC,UAAM,KAAK,QAAQ,SAAS,gBAAgB,SAAS,OAAO;AAE5D,UAAM,YAAY,KAAK,GAAG,KAAK,SAAS,qBAAqB;AAC7D,UAAM,WAAW,MAAM,KAAK,aAAa,SAAS;AAClD,QAAI,UAAU;AACZ,WAAK,cAAc,SAAS,QAAQ;AAAA,IACtC;AAEA,UAAM,SAAS,gBAAgB,YAAY,SAAS,cAAc;AAElE,SAAK,QAAQ,IAAI,gBAAgB,OAAO;AACxC,SAAK,YAAY,IAAI,gBAAgB;AAAA,MACnC,MAAM;AAAA,MACN,MAAM,KAAK,qBAAqB,OAAO;AAAA,MACvC,OAAO,QAAQ;AAAA,MACf,UAAU;AAAA,MACV;AAAA,IACF,CAAC;AAAA,EACH;AAAA,EAEA,MAAc,QACZ,SACA,gBACA,SACA,SACe;AACf,UAAM,UAAU,MAAM,KAAK,GAAG,gBAAgB,OAAO;AACrD,eAAW,UAAU,SAAS;AAC5B,YAAM,KAAK,QAAQ,SAAS,gBAAgB,QAAQ,OAAO;AAAA,IAC7D;AAEA,UAAM,QAAQ,MAAM,KAAK,GAAG,UAAU,OAAO;AAC7C,eAAW,YAAY,OAAO;AAC5B,YAAM,WAAW,KAAK,GAAG,SAAS,QAAQ;AAC1C,UAAI,aAAa,sBAAuB;AAExC,YAAM,MAAM,KAAK,GAAG,QAAQ,QAAQ;AACpC,YAAM,UAAU,MAAM,KAAK,GAAG,SAAS,QAAQ;AAC/C,YAAM,eAAe,KAAK,GAAG,SAAS,SAAS,QAAQ;AACvD,YAAM,OAAO,KAAK,GACf,cAAc,cAAc,GAAG,EAC/B,MAAM,GAAG,EACT,IAAI,CAAC,YAAY,QAAQ,OAAO,CAAC,EACjC,KAAK,GAAG;AAEX,UAAI,QAAQ,QAAQ;AAClB,gBAAQ,KAAK,KAAK,cAAc,gBAAgB,MAAM,OAAO,CAAC;AAAA,MAChE,WAAW,QAAQ,SAAS;AAC1B,gBAAQ,KAAK,GAAG,KAAK,iBAAiB,gBAAgB,MAAM,OAAO,CAAC;AAAA,MACtE;AAAA,IACF;AAAA,EACF;AAAA,EAEQ,cAAc,gBAAwB,MAAc,SAA+B;AACzF,UAAM,SAAS,SAAS,OAAO;AAC/B,WAAO;AAAA,MACL,YAAY;AAAA,MACZ;AAAA,MACA,MAAM,IAAI,cAAc,IAAI,IAAI;AAAA,MAChC,MAAM,OAAO;AAAA,MACb,MAAM,OAAO;AAAA,IACf;AAAA,EACF;AAAA,EAEQ,iBAAiB,gBAAwB,MAAc,SAAiC;AAC9F,UAAM,SAAS,UAAU,OAAO;AAEhC,QAAI,OAAO,SAAS,cAAc;AAChC,aAAO,OAAO,QAAQ,IAAI,CAAC,MAAM,UAAU;AACzC,cAAM,YACJ,OAAO,KAAK,MAAM,MAAM,WAAW,QAAQ,KAAK,MAAM,CAAC,IAAI,GAAG,IAAI,IAAI,KAAK;AAC7E,eAAO;AAAA,UACL,YAAY;AAAA,UACZ,MAAM;AAAA,UACN,MAAM,IAAI,cAAc,IAAI,SAAS;AAAA,UACrC;AAAA,QACF;AAAA,MACF,CAAC;AAAA,IACH;AAEA,WAAO,CAAC,EAAE,YAAY,gBAAgB,MAAM,MAAM,IAAI,cAAc,IAAI,IAAI,IAAI,MAAM,OAAO,KAAK,CAAC;AAAA,EACrG;AAAA,EAEA,MAAc,aAAa,WAA6C;AACtE,QAAI,CAAE,MAAM,KAAK,GAAG,OAAO,SAAS,EAAI,QAAO;AAE/C,QAAI;AACF,YAAM,UAAU,MAAM,KAAK,GAAG,SAAS,SAAS;AAChD,YAAM,SAAkB,KAAK,MAAM,OAAO;AAC1C,UAAI,MAAM,QAAQ,MAAM,EAAG,QAAO;AAAA,IACpC,SAAS,OAAO;AACd,cAAQ,KAAK,kDAAkD,SAAS,IAAI,KAAK;AAAA,IACnF;AACA,WAAO;AAAA,EACT;AAAA,EAEQ,cAAc,SAAyB,UAA0B;AACvE,UAAM,WAAW,IAAI,IAAI,SAAS,IAAI,CAAC,MAAM,UAAU,CAAC,MAAM,KAAK,CAAC,CAAC;AACrE,YAAQ,KAAK,CAAC,GAAG,MAAM;AACrB,YAAM,SAAS,SAAS,IAAI,EAAE,IAAI,KAAK;AACvC,YAAM,SAAS,SAAS,IAAI,EAAE,IAAI,KAAK;AACvC,aAAO,SAAS;AAAA,IAClB,CAAC;AAAA,EACH;AAAA,EAEQ,qBAAqB,SAA6C;AACxE,QAAI,QAAQ,WAAW,EAAG,QAAO;AACjC,UAAM,QAAQ,QAAQ,CAAC;AACvB,QAAI,MAAM,SAAS,OAAW,QAAO;AACrC,QAAI,QAAQ,WAAW,KAAK,CAAC,MAAM,KAAK,SAAS,GAAG,EAAG,QAAO;AAC9D,WAAO;AAAA,EACT;AACF;;;AI7JA,IAAI,QAA6B;AASjC,eAAsB,YACpB,WACA,QACuB;AACvB,QAAM,QAAQ,IAAI,aAAa,SAAS;AACxC,QAAM,MAAM,MAAM,MAAM;AACxB,UAAQ;AACR,SAAO;AACT;;;ACnBA,SAAS,OAAO,MAAc,SAAS,GAAW;AAChD,SAAO,KACJ,MAAM,IAAI,EACV,IAAI,CAAC,SAAU,KAAK,KAAK,MAAM,KAAK,KAAK,IAAI,OAAO,MAAM,IAAI,IAAK,EACnE,KAAK,IAAI;AACd;AAEA,SAAS,aAAa,KAAqB;AACzC,SAAO,IACJ,QAAQ,eAAe,CAAC,GAAG,MAAc,EAAE,YAAY,CAAC,EACxD,QAAQ,QAAQ,CAAC,GAAG,MAAc,EAAE,YAAY,CAAC;AACtD;AAEA,SAAS,cAAc,OAAgC;AACrD,UAAQ,MAAM,MAAM;AAAA,IAClB,KAAK;AAAA,IACL,KAAK;AACH,aAAO;AAAA,IAET,KAAK;AACH,aAAO;AAAA,IAET,KAAK;AACH,aAAO;AAAA,IAET,KAAK;AACH,aAAO;AAAA,IAET,KAAK;AACH,aAAO;AAAA,IAET,KAAK;AACH,aAAO;AAAA,IAET,KAAK;AACH,aAAO,MAAM,cAAc,SAAS;AAAA,IAEtC,KAAK;AAAA,IACL,KAAK;AACH,aAAO;AAAA,IAET,KAAK;AACH,aAAO;AAAA,IAET,KAAK;AACH,aAAO;AAAA,IAET,KAAK;AAAA,IACL,KAAK,UAAU;AACb,YAAM,SAAS,MAAM,QAAQ,IAAI,CAAC,MAAM,KAAK,UAAU,EAAE,KAAK,CAAC;AAC/D,aAAO,OAAO,SAAS,IAAI,OAAO,KAAK,KAAK,IAAI;AAAA,IAClD;AAAA,IAEA,KAAK,gBAAgB;AACnB,YAAM,SAAS,MAAM,QAAQ,IAAI,CAAC,MAAM,KAAK,UAAU,EAAE,KAAK,CAAC;AAC/D,YAAM,QAAQ,OAAO,SAAS,IAAI,OAAO,KAAK,KAAK,IAAI;AACvD,aAAO,SAAS,KAAK;AAAA,IACvB;AAAA,IAEA,KAAK;AACH,aAAO,mBAAmB,MAAM,MAAM;AAAA,IAExC,KAAK;AACH,aAAO,SAAS,mBAAmB,MAAM,UAAU,CAAC;AAAA,IAEtD,KAAK;AACH,aAAO,MAAM,aAAa,OAAO,SAAS;AAAA,IAE5C,KAAK;AACH,UAAI,MAAM,eAAe,SAAU,QAAO;AAC1C,UAAI,MAAM,eAAe,UAAW,QAAO;AAC3C,aAAO;AAAA,IAET;AACE,aAAO;AAAA,EACX;AACF;AAEA,SAAS,mBAAmB,QAAmC;AAC7D,MAAI,OAAO,WAAW,EAAG,QAAO;AAEhC,QAAM,QAAQ,OAAO,IAAI,CAAC,UAAU;AAClC,UAAM,WAAW,MAAM,aAAa,QAAQ,MAAM;AAClD,UAAM,SAAS,cAAc,KAAK;AAClC,UAAM,UAAU,MAAM,cAAc,OAAO,MAAM,WAAW;AAAA,IAAU;AACtE,WAAO,GAAG,OAAO,GAAG,MAAM,IAAI,GAAG,QAAQ,KAAK,MAAM;AAAA,EACtD,CAAC;AAED,SAAO;AAAA,EAAM,OAAO,MAAM,KAAK,IAAI,CAAC,CAAC;AAAA;AACvC;AAEO,SAAS,2BAA2B,QAAkC;AAC3E,QAAM,OAAO,aAAa,OAAO,UAAU,IAAI;AAC/C,QAAM,QAAQ,OAAO,SAAS,OAAO;AACrC,QAAM,OAAO,mBAAmB,OAAO,MAAM;AAC7C,SAAO,2BAA2B,KAAK;AAAA,mBAAsC,IAAI,IAAI,IAAI;AAC3F;AAWO,SAAS,wBAAwB,SAAqC;AAC3E,QAAM,SAAS;AAAA,IACb;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,EACF,EAAE,KAAK,IAAI;AAEX,QAAM,aAAa,QAAQ,IAAI,0BAA0B,EAAE,KAAK,MAAM;AAEtE,QAAM,qBAAqB;AAAA,IACzB;AAAA,IACA;AAAA,IACA;AAAA,IACA,QAAQ,IAAI,CAAC,WAAW,OAAO,KAAK,UAAU,OAAO,UAAU,CAAC,KAAK,aAAa,OAAO,UAAU,CAAC,QAAQ,EAAE,KAAK,IAAI;AAAA,IACvH;AAAA,IACA;AAAA,EACF,EAAE,KAAK,IAAI;AAEX,SAAO,CAAC,QAAQ,YAAY,kBAAkB,EAAE,KAAK,MAAM,IAAI;AACjE;;;AClJA;AAAA,EACI,MAAQ;AAAA,EACR,SAAW;AAAA,EACX,aAAe;AAAA,EACf,UAAY;AAAA,IACR;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,EACJ;AAAA,EACA,UAAY;AAAA,EACZ,YAAc;AAAA,IACV,MAAQ;AAAA,IACR,KAAO;AAAA,EACX;AAAA,EACA,SAAW;AAAA,EACX,QAAU;AAAA,EACV,MAAQ;AAAA,EACR,SAAW;AAAA,IACP,KAAK;AAAA,MACD,OAAS;AAAA,MACT,QAAU;AAAA,IACd;AAAA,EACJ;AAAA,EACA,MAAQ;AAAA,EACR,OAAS;AAAA,EACT,KAAO;AAAA,IACH,iBAAiB;AAAA,EACrB;AAAA,EACA,OAAS;AAAA,IACL;AAAA,IACA;AAAA,IACA;AAAA,EACJ;AAAA,EACA,SAAW;AAAA,IACP,KAAO;AAAA,IACP,cAAc;AAAA,IACd,gBAAgB;AAAA,IAChB,OAAS;AAAA,IACT,MAAQ;AAAA,IACR,cAAc;AAAA,IACd,MAAQ;AAAA,IACR,cAAc;AAAA,EAClB;AAAA,EACA,SAAW;AAAA,IACP,MAAQ;AAAA,EACZ;AAAA,EACA,gBAAkB;AAAA,EAClB,iBAAmB;AAAA,IACf,+BAA+B;AAAA,IAC/B,yBAAyB;AAAA,IACzB,0BAA0B;AAAA,IAC1B,wBAAwB;AAAA,IACxB,yBAAyB;AAAA,IACzB,iCAAiC;AAAA,IACjC,yCAAyC;AAAA,IACzC,kCAAkC;AAAA,IAClC,2BAA2B;AAAA,IAC3B,0BAA0B;AAAA,IAC1B,iCAAiC;AAAA,IACjC,iBAAiB;AAAA,IACjB,uBAAuB;AAAA,IACvB,sBAAsB;AAAA,IACtB,oBAAoB;AAAA,IACpB,eAAe;AAAA,IACf,gBAAgB;AAAA,IAChB,oBAAoB;AAAA,IACpB,4BAA4B;AAAA,IAC5B,MAAQ;AAAA,IACR,aAAa;AAAA,IACb,QAAU;AAAA,IACV,UAAY;AAAA,IACZ,gBAAgB;AAAA,IAChB,SAAW;AAAA,IACX,MAAQ;AAAA,IACR,eAAe;AAAA,IACf,OAAS;AAAA,IACT,aAAa;AAAA,IACb,kBAAkB;AAAA,IAClB,aAAe;AAAA,IACf,YAAY;AAAA,IACZ,uCAAuC;AAAA,IACvC,mBAAmB;AAAA,IACnB,MAAQ;AAAA,IACR,KAAO;AAAA,IACP,YAAc;AAAA,IACd,QAAU;AAAA,IACV,SAAW;AAAA,EACf;AAAA,EACA,cAAgB;AAAA,IACZ,yBAAyB;AAAA,IACzB,UAAY;AAAA,IACZ,WAAa;AAAA,IACb,eAAe;AAAA,IACf,aAAa;AAAA,EACjB;AACJ;;;AT9EA,IAAM,EAAE,QAAQ,IAAI;AAEpB,IAAM,UAAU,IAAI,QAAQ,EACzB,KAAK,eAAe,EACpB,YAAY,sCAAsC,EAClD,QAAQ,OAAO,EACf,OAAO,oBAAoB,8BAA8B,YAAY,EACrE,OAAO,uBAAuB,6BAA6B,OAAO,QAAQ,CAAC,EAC3E,OAAO,oBAAoB,mDAAmD,EAC9E,MAAM;AAET,IAAM,OAAO,QAAQ,KAA6D;AAClF,IAAM,cAAcC,MAAK,QAAQ,KAAK,GAAG;AACzC,IAAM,OAAO,OAAO,KAAK,IAAI;AAE7B,eAAe,iBAAiB,WAAkC;AAChE,QAAM,SAASA,MAAK,QAAQ,SAAS;AACrC,QAAM,UAAUA,MAAK,KAAK,QAAQ,aAAa;AAE/C,UAAQ,IAAI,yBAAyB,SAAS,KAAK;AAEnD,QAAM,YAAY,IAAI,UAAU,SAAS;AACzC,QAAM,QAAQ,MAAM,YAAY,SAAS;AACzC,QAAM,UAAU,MAAM,eAAe,EAAE,QAAQ,CAAC,MAAO,EAAE,SAAS,CAAC,EAAE,MAAM,IAAI,CAAC,CAAE;AAClF,QAAM,OAAO,wBAAwB,OAAO;AAE5C,QAAMC,IAAG,MAAM,QAAQ,EAAE,WAAW,KAAK,CAAC;AAC1C,QAAMA,IAAG,UAAU,SAAS,MAAM,OAAO;AAEzC,UAAQ,IAAI,oBAAoB,OAAO,EAAE;AAC3C;AAEA,SAAS,qBACPC,QACA,YACA,KACqB;AACrB,QAAM,mBAAmBF,MAAK,QAAQE,QAAO,uCAAuC;AACpF,MAAI,WAAW,gBAAgB,GAAG;AAChC,WAAO,MAAM,QAAQ,CAAC,gBAAgB,GAAG,EAAE,OAAO,WAAW,IAAI,CAAC;AAAA,EACpE;AAGA,QAAM,gBAAgBF,MAAK,QAAQE,QAAO,cAAc;AACxD,MAAI,WAAW,aAAa,GAAG;AAE7B,UAAM,UAAUF,MAAK,QAAQE,QAAO,0CAA0C;AAC9E,WAAO,MAAM,QAAQ,CAAC,SAAS,OAAO,UAAU,OAAO,UAAU,GAAG,WAAW,GAAG;AAAA,MAChF,KAAKA;AAAA,MACL,OAAO;AAAA,MACP;AAAA,IACF,CAAC;AAAA,EACH;AAEA,SAAO;AACT;AAEA,SAAS,eAAe,OAA2B;AACjD,aAAW,UAAU,CAAC,UAAU,SAAS,GAAY;AACnD,YAAQ,GAAG,QAAQ,MAAM,MAAM,KAAK,MAAM,CAAC;AAAA,EAC7C;AACF;AAEA,IAAI,KAAK,eAAe;AACtB,QAAM,iBAAiB,WAAW;AAClC,UAAQ,KAAK,CAAC;AAChB;AAEA,IAAM,QAAQF,MAAK,QAAQ,YAAY,SAAS,WAAW;AAC3D,IAAM,YAAY,EAAE,GAAG,QAAQ,KAAK,qBAAqB,aAAa,MAAM,OAAO,IAAI,GAAG,UAAU,UAAU;AAC9G,IAAM,gBAAgB,qBAAqB,OAAO,MAAM,SAAS;AAEjE,IAAI,CAAC,eAAe;AAClB,UAAQ,MAAM,oCAAoC;AAClD,UAAQ,MAAM,wDAAwD;AACtE,UAAQ,KAAK,CAAC;AAChB;AAEA,QAAQ,IAAI,kBAAkB,OAAO,EAAE;AACvC,QAAQ,IAAI,aAAa,WAAW,EAAE;AACtC,QAAQ,IAAI,gCAAgC,IAAI,EAAE;AAElD,cAAc,GAAG,SAAS,CAAC,UAAU;AACnC,UAAQ,MAAM,2BAA2B,MAAM,OAAO;AACtD,UAAQ,KAAK,CAAC;AAChB,CAAC;AAED,cAAc,GAAG,SAAS,CAAC,SAAS,QAAQ,KAAK,QAAQ,CAAC,CAAC;AAE3D,eAAe,aAAa;","names":["fs","path","path","fs","uiDir"]}
@@ -210,6 +210,20 @@ interface CollectionConfig {
210
210
  sync?: string;
211
211
  };
212
212
  }
213
+ /**
214
+ * Augmentable map of collection names to their typed entry shapes.
215
+ * Extend this via generated types or manual declaration:
216
+ *
217
+ * @example
218
+ * // .studio/studio.d.ts (auto-generated)
219
+ * declare module 'nextjs-studio' {
220
+ * interface CollectionTypeMap {
221
+ * posts: { title: string; date: string; slug: string };
222
+ * }
223
+ * }
224
+ */
225
+ interface CollectionTypeMap {
226
+ }
213
227
  /**
214
228
  * Query options for the content query builder.
215
229
  */
@@ -264,7 +278,7 @@ interface DirectoryFileEntry {
264
278
  * queryCollection("pages").where({ "hero.title": "Welcome" }).all();
265
279
  * ```
266
280
  */
267
- declare class QueryBuilder {
281
+ declare class QueryBuilder<T = Record<string, unknown>> {
268
282
  private readonly collectionName;
269
283
  private options;
270
284
  constructor(collection: string);
@@ -272,14 +286,15 @@ declare class QueryBuilder {
272
286
  sort(field: string, order?: "asc" | "desc"): this;
273
287
  limit(count: number): this;
274
288
  offset(count: number): this;
275
- all(): ContentEntry[];
276
- first(): ContentEntry | undefined;
289
+ all(): T[];
290
+ first(): T | undefined;
277
291
  count(): number;
278
292
  }
279
293
  /**
280
294
  * Entry point for querying a content collection.
281
295
  */
282
- declare function queryCollection(collection: string): QueryBuilder;
296
+ declare function queryCollection<K extends keyof CollectionTypeMap>(name: K): QueryBuilder<CollectionTypeMap[K]>;
297
+ declare function queryCollection(name: string): QueryBuilder<Record<string, unknown>>;
283
298
 
284
299
  /**
285
300
  * @context Core layer — studio initializer at src/core/init.ts
@@ -309,6 +324,26 @@ declare function initStudio(contentsDir?: string, config?: StudioConfig): Promis
309
324
  * Returns true if the content store has been initialized.
310
325
  */
311
326
  declare function isStudioInitialized(): boolean;
327
+ /**
328
+ * Ensures the content store is initialized. Safe to call multiple times — only
329
+ * initializes once. Ideal for use at the top of Next.js server components and
330
+ * `generateStaticParams` / `generateMetadata` functions.
331
+ *
332
+ * @param contentsDir - Path to the contents directory. Defaults to `./contents` relative to cwd.
333
+ * @param config - Optional studio config for schemas and scripts.
334
+ *
335
+ * @example
336
+ * ```ts
337
+ * import { ensureContentLoaded, queryCollection } from "nextjs-studio";
338
+ *
339
+ * export default async function Page() {
340
+ * await ensureContentLoaded();
341
+ * const posts = queryCollection("posts");
342
+ * // ...
343
+ * }
344
+ * ```
345
+ */
346
+ declare function ensureContentLoaded(contentsDir?: string, config?: StudioConfig): Promise<void>;
312
347
 
313
348
  /**
314
349
  * @context Shared layer — FS adapter interface at src/shared/fs-adapter.interface.ts
@@ -436,4 +471,4 @@ declare function fieldLabel(field: Pick<BaseField, "name" | "label">): string;
436
471
  */
437
472
  declare function keyLabel(name: string): string;
438
473
 
439
- export { type Collection, type CollectionConfig, type CollectionSchema, type ContentEntry, ContentIndex, type FieldDefinition, type FieldType, type InferFieldValue, type InferSchemaData, type QueryOptions, type StudioConfig, fieldLabel, initStudio, isStudioInitialized, keyLabel, loadContent, queryCollection };
474
+ export { type Collection, type CollectionConfig, type CollectionSchema, type ContentEntry, ContentIndex, type FieldDefinition, type FieldType, type InferFieldValue, type InferSchemaData, type QueryOptions, type StudioConfig, ensureContentLoaded, fieldLabel, initStudio, isStudioInitialized, keyLabel, loadContent, queryCollection };
@@ -289,7 +289,7 @@ var QueryBuilder = class {
289
289
  }
290
290
  const start = this.options.offset ?? 0;
291
291
  const end = this.options.limit ? start + this.options.limit : void 0;
292
- return slice(entries, start, end);
292
+ return slice(entries, start, end).map((e) => e.data);
293
293
  }
294
294
  first() {
295
295
  return this.limit(1).all()[0];
@@ -298,8 +298,8 @@ var QueryBuilder = class {
298
298
  return this.all().length;
299
299
  }
300
300
  };
301
- function queryCollection(collection) {
302
- return new QueryBuilder(collection);
301
+ function queryCollection(name) {
302
+ return new QueryBuilder(name);
303
303
  }
304
304
 
305
305
  // src/core/init.ts
@@ -416,6 +416,11 @@ function isStudioInitialized() {
416
416
  return false;
417
417
  }
418
418
  }
419
+ async function ensureContentLoaded(contentsDir, config) {
420
+ if (!isStudioInitialized()) {
421
+ await initStudio(contentsDir, config);
422
+ }
423
+ }
419
424
 
420
425
  // src/shared/field-utils.ts
421
426
  function fieldLabel(field) {
@@ -427,6 +432,7 @@ function keyLabel(name) {
427
432
  }
428
433
  export {
429
434
  ContentIndex,
435
+ ensureContentLoaded,
430
436
  fieldLabel,
431
437
  initStudio,
432
438
  isStudioInitialized,
@@ -1 +1 @@
1
- {"version":3,"sources":["../../src/core/query-builder.ts","../../src/core/indexer.ts","../../src/shared/constants.ts","../../src/core/parsers/parser-mdx.ts","../../src/core/parsers/parser-json.ts","../../src/core/schema-inferrer.ts","../../src/core/content-store.ts","../../src/core/init.ts","../../src/cli/adapters/fs-adapter.ts","../../src/shared/field-utils.ts"],"sourcesContent":["/**\n * @context Core layer — query builder at src/core/query-builder.ts\n * @does Provides a fluent API to filter, sort, and paginate content entries from a collection\n * @depends src/shared/types.ts, src/core/content-store.ts\n * @do Add new query capabilities here (e.g. search, groupBy)\n * @dont Import from CLI or UI; access the filesystem; perform I/O\n */\n\nimport { filter, orderBy, get, slice } from \"lodash-es\";\nimport type { ContentEntry, QueryOptions } from \"../shared/types.js\";\nimport { getStore } from \"./content-store.js\";\n\n/**\n * Fluent query builder for content collections.\n *\n * ```ts\n * const posts = queryCollection(\"blog\")\n * .where({ published: true })\n * .sort(\"date\", \"desc\")\n * .limit(10)\n * .all();\n * ```\n *\n * Supports dot notation for nested properties:\n * ```ts\n * queryCollection(\"pages\").where({ \"hero.title\": \"Welcome\" }).all();\n * ```\n */\nexport class QueryBuilder {\n private readonly collectionName: string;\n private options: QueryOptions = {};\n\n constructor(collection: string) {\n this.collectionName = collection;\n }\n\n where(conditions: Record<string, unknown>): this {\n this.options.where = { ...this.options.where, ...conditions };\n return this;\n }\n\n sort(field: string, order: \"asc\" | \"desc\" = \"asc\"): this {\n this.options.sort = { field, order };\n return this;\n }\n\n limit(count: number): this {\n this.options.limit = count;\n return this;\n }\n\n offset(count: number): this {\n this.options.offset = count;\n return this;\n }\n\n all(): ContentEntry[] {\n let entries = [...getStore().getCollection(this.collectionName)];\n\n if (this.options.where) {\n const conditions = this.options.where;\n entries = filter(entries, (entry) =>\n Object.entries(conditions).every(([key, value]) => get(entry.data, key) === value),\n );\n }\n\n if (this.options.sort) {\n const { field, order } = this.options.sort;\n entries = orderBy(entries, [(entry) => get(entry.data, field)], [order]);\n }\n\n const start = this.options.offset ?? 0;\n const end = this.options.limit ? start + this.options.limit : undefined;\n return slice(entries, start, end);\n }\n\n first(): ContentEntry | undefined {\n return this.limit(1).all()[0];\n }\n\n count(): number {\n return this.all().length;\n }\n}\n\n/**\n * Entry point for querying a content collection.\n */\nexport function queryCollection(collection: string): QueryBuilder {\n return new QueryBuilder(collection);\n}\n","/**\n * @context Core layer — content indexer at src/core/indexer.ts\n * @does Scans the contents directory, parses MDX/JSON files, and builds an in-memory index\n * @depends src/shared/types.ts, src/shared/constants.ts, src/shared/fs-adapter.interface.ts, src/core/parsers/, src/core/schema-inferrer.ts\n * @do Add new file type handling here; extend indexCollection for new collection behaviors\n * @dont Import from CLI or UI; instantiate FsAdapter; access the filesystem directly\n */\n\nimport slugify from \"@sindresorhus/slugify\";\nimport type { CollectionSchema } from \"../shared/fields.js\";\nimport type { ContentEntry, Collection, StudioConfig } from \"../shared/types.js\";\nimport type { IFsAdapter } from \"../shared/fs-adapter.interface.js\";\nimport { COLLECTION_ORDER_FILE } from \"../shared/constants.js\";\nimport { parseMdx } from \"./parsers/parser-mdx.js\";\nimport { parseJson } from \"./parsers/parser-json.js\";\nimport { inferSchema } from \"./schema-inferrer.js\";\n\nexport class ContentIndex {\n private readonly entries = new Map<string, ContentEntry[]>();\n private readonly collections = new Map<string, Collection>();\n private readonly fs: IFsAdapter;\n\n constructor(fsAdapter: IFsAdapter) {\n this.fs = fsAdapter;\n }\n\n async build(config?: StudioConfig): Promise<void> {\n this.clear();\n const dirs = await this.fs.listDirectories(\".\");\n\n for (const dir of dirs) {\n const dirName = this.fs.basename(dir);\n const collectionName = slugify(dirName);\n const collectionConfig = config?.collections?.[collectionName];\n await this.indexCollection(dirName, collectionName, collectionConfig?.schema);\n }\n }\n\n getCollection(name: string): ContentEntry[] {\n return this.entries.get(name) ?? [];\n }\n\n getCollections(): Collection[] {\n return Array.from(this.collections.values());\n }\n\n clear(): void {\n this.entries.clear();\n this.collections.clear();\n }\n\n private async indexCollection(\n dirName: string,\n collectionName: string,\n manualSchema?: CollectionSchema,\n ): Promise<void> {\n const entries: ContentEntry[] = [];\n await this.scanDir(dirName, collectionName, dirName, entries);\n\n const orderPath = this.fs.join(dirName, COLLECTION_ORDER_FILE);\n const ordering = await this.readOrdering(orderPath);\n if (ordering) {\n this.applyOrdering(entries, ordering);\n }\n\n const schema = manualSchema ?? inferSchema(entries, collectionName);\n\n this.entries.set(collectionName, entries);\n this.collections.set(collectionName, {\n name: collectionName,\n type: this.detectCollectionType(entries),\n count: entries.length,\n basePath: dirName,\n schema,\n });\n }\n\n private async scanDir(\n dirName: string,\n collectionName: string,\n dirPath: string,\n entries: ContentEntry[],\n ): Promise<void> {\n const subDirs = await this.fs.listDirectories(dirPath);\n for (const subDir of subDirs) {\n await this.scanDir(dirName, collectionName, subDir, entries);\n }\n\n const files = await this.fs.listFiles(dirPath);\n for (const filePath of files) {\n const fileName = this.fs.basename(filePath);\n if (fileName === COLLECTION_ORDER_FILE) continue;\n\n const ext = this.fs.extname(fileName);\n const content = await this.fs.readFile(filePath);\n const relativePath = this.fs.relative(dirName, filePath);\n const slug = this.fs\n .normalizeSlug(relativePath, ext)\n .split(\"/\")\n .map((segment) => slugify(segment))\n .join(\"/\");\n\n if (ext === \".mdx\") {\n entries.push(this.buildMdxEntry(collectionName, slug, content));\n } else if (ext === \".json\") {\n entries.push(...this.buildJsonEntries(collectionName, slug, content));\n }\n }\n }\n\n private buildMdxEntry(collectionName: string, slug: string, content: string): ContentEntry {\n const parsed = parseMdx(content);\n return {\n collection: collectionName,\n slug,\n path: `/${collectionName}/${slug}`,\n body: parsed.body,\n data: parsed.data,\n };\n }\n\n private buildJsonEntries(collectionName: string, slug: string, content: string): ContentEntry[] {\n const parsed = parseJson(content);\n\n if (parsed.type === \"json-array\") {\n return parsed.entries.map((data, index) => {\n const entrySlug =\n typeof data[\"slug\"] === \"string\" ? slugify(data[\"slug\"]) : `${slug}/${index}`;\n return {\n collection: collectionName,\n slug: entrySlug,\n path: `/${collectionName}/${entrySlug}`,\n data,\n };\n });\n }\n\n return [{ collection: collectionName, slug, path: `/${collectionName}/${slug}`, data: parsed.data }];\n }\n\n private async readOrdering(orderPath: string): Promise<string[] | null> {\n if (!(await this.fs.exists(orderPath))) return null;\n\n try {\n const content = await this.fs.readFile(orderPath);\n const parsed: unknown = JSON.parse(content);\n if (Array.isArray(parsed)) return parsed as string[];\n } catch (error) {\n console.warn(`[Nextjs Studio] Failed to parse ordering file: ${orderPath}`, error);\n }\n return null;\n }\n\n private applyOrdering(entries: ContentEntry[], ordering: string[]): void {\n const orderMap = new Map(ordering.map((slug, index) => [slug, index]));\n entries.sort((a, b) => {\n const aIndex = orderMap.get(a.slug) ?? Infinity;\n const bIndex = orderMap.get(b.slug) ?? Infinity;\n return aIndex - bIndex;\n });\n }\n\n private detectCollectionType(entries: ContentEntry[]): Collection[\"type\"] {\n if (entries.length === 0) return \"mdx\";\n const first = entries[0];\n if (first.body !== undefined) return \"mdx\";\n if (entries.length === 1 && !first.slug.includes(\"/\")) return \"json-object\";\n return \"json-array\";\n }\n}\n","/**\n * @context Shared layer — constants at src/shared/constants.ts\n * @does Defines project-wide constants shared across core, CLI, and UI layers\n * @depends none\n * @do Add new shared constants here\n * @dont Import from CLI or UI; constants must be framework-agnostic\n */\n\nexport const CONTENTS_DIR = \"contents\";\nexport const CLI_PORT = 3030;\nexport const CONFIG_FILE = \"studio.config.ts\";\nexport const SUPPORTED_EXTENSIONS = [\".mdx\", \".json\"] as const;\nexport const COLLECTION_ORDER_FILE = \"collection.json\";\nexport const WATCHER_DEBOUNCE_MS = 5_000;\nexport const MEDIA_DIR = \"media\";\n\nexport const IMAGE_MIME_TYPES = [\n \"image/png\",\n \"image/jpeg\",\n \"image/gif\",\n \"image/webp\",\n \"image/svg+xml\",\n \"image/avif\",\n] as const;\n\nexport const VIDEO_MIME_TYPES = [\"video/mp4\", \"video/webm\", \"video/ogg\"] as const;\n\nexport const AUDIO_MIME_TYPES = [\n \"audio/mpeg\",\n \"audio/ogg\",\n \"audio/wav\",\n \"audio/webm\",\n \"audio/aac\",\n \"audio/flac\",\n] as const;\n\nexport const MEDIA_MIME_TYPES = [...IMAGE_MIME_TYPES, ...VIDEO_MIME_TYPES, ...AUDIO_MIME_TYPES] as const;\n\nexport const IMAGE_EXTENSIONS = [\".png\", \".jpg\", \".jpeg\", \".gif\", \".webp\", \".svg\", \".avif\"] as const;\nexport const VIDEO_EXTENSIONS = [\".mp4\", \".webm\", \".ogv\"] as const;\nexport const AUDIO_EXTENSIONS = [\".mp3\", \".ogg\", \".wav\", \".m4a\", \".aac\", \".flac\"] as const;\n","/**\n * @context Core layer — MDX parser/serializer at src/core/parsers/parser-mdx.ts\n * @does Parses .mdx content into frontmatter + body, and serializes them back to MDX strings\n * @depends none (gray-matter is an external dep)\n * @do Add MDX transform steps here; both parse and serialize live here intentionally\n * @dont Access the filesystem; import from CLI or UI; handle JSON content\n */\n\nimport matter from \"gray-matter\";\n\nexport interface ParsedMdx {\n data: Record<string, unknown>;\n body: string;\n}\n\nexport function parseMdx(content: string): ParsedMdx {\n const { data, content: body } = matter(content);\n return { data, body: body.trim() };\n}\n\nexport function serializeMdx(data: Record<string, unknown>, body: string): string {\n return matter.stringify(body, data);\n}\n","/**\n * @context Core layer — JSON parser at src/core/parsers/parser-json.ts\n * @does Parses JSON content strings into typed ParsedJson results (array or object)\n * @depends none\n * @do Extend ParsedJson variants here if new JSON structures are supported\n * @dont Access the filesystem; import from CLI or UI; contain serialization logic\n */\n\nexport interface ParsedJsonArray {\n type: \"json-array\";\n entries: Record<string, unknown>[];\n}\n\nexport interface ParsedJsonObject {\n type: \"json-object\";\n data: Record<string, unknown>;\n}\n\nexport type ParsedJson = ParsedJsonArray | ParsedJsonObject;\n\nexport function parseJson(content: string): ParsedJson {\n const parsed: unknown = JSON.parse(content);\n\n if (Array.isArray(parsed)) {\n return {\n type: \"json-array\",\n entries: parsed as Record<string, unknown>[],\n };\n }\n\n if (typeof parsed === \"object\" && parsed !== null) {\n return {\n type: \"json-object\",\n data: parsed as Record<string, unknown>,\n };\n }\n\n throw new Error(\"JSON content must be an array or object\");\n}\n","/**\n * @context Core layer — schema inferrer at src/core/schema-inferrer.ts\n * @does Infers a CollectionSchema from actual content entries when no manual schema is defined\n * @depends src/shared/types.ts, src/shared/fields.ts\n * @do Add new type detection heuristics here (e.g. color, phone)\n * @dont Import from CLI or UI; access the filesystem; perform I/O\n */\n\nimport type { ContentEntry } from \"../shared/types.js\";\nimport type { CollectionSchema, FieldDefinition, SelectOption } from \"../shared/fields.js\";\n\n// Value detector patterns\nconst RE_ISO_DATE = /^\\d{4}-\\d{2}-\\d{2}$/;\nconst RE_ISO_DATETIME =\n /^\\d{4}-\\d{2}-\\d{2}[T ]\\d{2}:\\d{2}(:\\d{2}(\\.\\d+)?)?(Z|[+-]\\d{2}:?\\d{2})?$/;\nconst RE_EMAIL = /^[^\\s@]+@[^\\s@]+\\.[^\\s@]+$/;\nconst RE_URL = /^https?:\\/\\/.+/;\nconst LONG_TEXT_THRESHOLD = 200;\n\nfunction isISODate(value: string): boolean {\n return RE_ISO_DATE.test(value);\n}\n\nfunction isISODateTime(value: string): boolean {\n return RE_ISO_DATETIME.test(value);\n}\n\nfunction isEmail(value: string): boolean {\n return RE_EMAIL.test(value);\n}\n\nfunction isUrl(value: string): boolean {\n return RE_URL.test(value);\n}\n\nfunction inferStringField(name: string, strings: string[]): FieldDefinition {\n if (strings.every(isEmail)) return { name, type: \"email\" };\n if (strings.every(isUrl)) return { name, type: \"url\" };\n if (strings.every(isISODateTime)) return { name, type: \"date\", includeTime: true };\n if (strings.every(isISODate)) return { name, type: \"date\" };\n\n const isLong = strings.some((s) => s.length > LONG_TEXT_THRESHOLD || s.includes(\"\\n\"));\n return { name, type: isLong ? \"long-text\" : \"text\" };\n}\n\nfunction inferArrayField(name: string, items: unknown[]): FieldDefinition {\n if (items.length === 0) return { name, type: \"array\", itemFields: [] };\n\n if (items.every((item) => typeof item === \"string\")) {\n const unique = [...new Set(items as string[])].slice(0, 50);\n const options: SelectOption[] = unique.map((v) => ({ label: v, value: v }));\n return { name, type: \"multi-select\", options };\n }\n\n if (items.every((item) => typeof item === \"object\" && item !== null && !Array.isArray(item))) {\n return { name, type: \"array\", itemFields: inferFields(items as Record<string, unknown>[]) };\n }\n\n return { name, type: \"array\", itemFields: [] };\n}\n\nfunction inferFieldDefinition(name: string, values: unknown[]): FieldDefinition {\n const present = values.filter((v) => v !== null && v !== undefined);\n\n if (present.length === 0) return { name, type: \"text\" };\n if (present.every((v) => typeof v === \"boolean\")) return { name, type: \"boolean\" };\n\n if (present.every((v) => typeof v === \"number\")) {\n const format = present.every((v) => Number.isInteger(v)) ? \"integer\" : \"decimal\";\n return { name, type: \"number\", format };\n }\n\n if (present.every((v) => typeof v === \"string\")) {\n return inferStringField(name, present as string[]);\n }\n\n if (present.every((v) => Array.isArray(v))) {\n return inferArrayField(name, (present as unknown[][]).flat());\n }\n\n if (present.every((v) => typeof v === \"object\" && v !== null && !Array.isArray(v))) {\n return { name, type: \"object\", fields: inferFields(present as Record<string, unknown>[]) };\n }\n\n return { name, type: \"text\" };\n}\n\nfunction inferFields(rows: Record<string, unknown>[]): FieldDefinition[] {\n const keySet = new Set<string>(rows.flatMap((row) => Object.keys(row)));\n return Array.from(keySet).map((key) => inferFieldDefinition(key, rows.map((row) => row[key])));\n}\n\n/**\n * Infer a `CollectionSchema` from the data of a set of content entries.\n *\n * The result is a best-effort approximation — string fields that look like\n * emails, URLs, or ISO dates get the correct semantic type. Everything else\n * falls back to `text`.\n */\nexport function inferSchema(entries: ContentEntry[], collectionName: string): CollectionSchema {\n const rows = entries.map((entry) => entry.data as Record<string, unknown>);\n return { collection: collectionName, fields: inferFields(rows) };\n}\n","/**\n * @context Core layer — content store at src/core/content-store.ts\n * @does Manages a singleton ContentIndex; exposes loadContent() and getStore() for consumers\n * @depends src/core/indexer.ts, src/shared/fs-adapter.interface.ts, src/shared/types.ts\n * @do Use this as the single access point for in-memory indexed content\n * @dont Import from CLI or UI; instantiate FsAdapter here; contain parsing or I/O logic\n */\n\nimport type { IFsAdapter } from \"../shared/fs-adapter.interface.js\";\nimport type { StudioConfig } from \"../shared/types.js\";\nimport { ContentIndex } from \"./indexer.js\";\n\nlet store: ContentIndex | null = null;\n\nexport function getStore(): ContentIndex {\n if (!store) {\n throw new Error(\"Content not loaded. Call loadContent() before querying.\");\n }\n return store;\n}\n\nexport async function loadContent(\n fsAdapter: IFsAdapter,\n config?: StudioConfig,\n): Promise<ContentIndex> {\n const index = new ContentIndex(fsAdapter);\n await index.build(config);\n store = index;\n return index;\n}\n","/**\r\n * @context Core layer — studio initializer at src/core/init.ts\r\n * @does Convenience function to initialize the content store for Node.js environments\r\n * @depends src/core/content-store.ts, src/cli/adapters/fs-adapter.ts\r\n * @do Use this as the entry point for Next.js and other Node.js consumers\r\n * @dont Import from UI; use in browser environments\r\n */\r\n\r\nimport path from \"node:path\";\r\nimport { FsAdapter } from \"../cli/adapters/fs-adapter.js\";\r\nimport { loadContent, getStore } from \"./content-store.js\";\r\nimport type { StudioConfig } from \"../shared/types.js\";\r\n\r\n/**\r\n * Initialize the content store from the filesystem.\r\n * Call this once before using `queryCollection()`.\r\n *\r\n * @param contentsDir - Path to the contents directory. Defaults to `./contents` relative to cwd.\r\n * @param config - Optional studio config for schemas and scripts.\r\n *\r\n * @example\r\n * ```ts\r\n * import { initStudio, queryCollection } from \"nextjs-studio\";\r\n *\r\n * await initStudio();\r\n * const posts = queryCollection(\"posts\").all();\r\n * ```\r\n */\r\nexport async function initStudio(\r\n contentsDir?: string,\r\n config?: StudioConfig,\r\n): Promise<void> {\r\n const dir = contentsDir ?? path.join(process.cwd(), \"contents\");\r\n await loadContent(new FsAdapter(dir), config);\r\n}\r\n\r\n/**\r\n * Returns true if the content store has been initialized.\r\n */\r\nexport function isStudioInitialized(): boolean {\r\n try {\r\n getStore();\r\n return true;\r\n } catch {\r\n return false;\r\n }\r\n}\r\n","/**\n * @context CLI layer — filesystem adapter at src/cli/adapters/fs-adapter.ts\n * @does Implements IFsAdapter; abstracts all file read/write/list operations behind a single interface\n * @depends src/shared/types.ts, src/shared/constants.ts, src/shared/fs-adapter.interface.ts\n * @do Add new I/O operations here; all file access must go through this adapter\n * @dont Import UI components, run HTTP requests, or contain business logic\n */\n\nimport fs from \"node:fs/promises\";\nimport path from \"node:path\";\nimport type { Dirent } from \"node:fs\";\nimport type { FileInfo, DirectoryFileEntry } from \"../../shared/types.js\";\nimport type { IFsAdapter } from \"../../shared/fs-adapter.interface.js\";\nimport { SUPPORTED_EXTENSIONS } from \"../../shared/constants.js\";\n\nexport class FsAdapter implements IFsAdapter {\n private readonly basePath: string;\n\n constructor(basePath: string) {\n this.basePath = path.resolve(basePath);\n }\n\n private resolve(...segments: string[]): string {\n return path.resolve(this.basePath, ...segments);\n }\n\n async readFile(filePath: string): Promise<string> {\n return fs.readFile(this.resolve(filePath), \"utf-8\");\n }\n\n async writeFile(filePath: string, content: string): Promise<void> {\n const fullPath = this.resolve(filePath);\n await fs.mkdir(path.dirname(fullPath), { recursive: true });\n await fs.writeFile(fullPath, content, \"utf-8\");\n }\n\n async deleteFile(filePath: string): Promise<void> {\n await fs.unlink(this.resolve(filePath));\n }\n\n async exists(filePath: string): Promise<boolean> {\n try {\n await fs.access(this.resolve(filePath));\n return true;\n } catch {\n return false;\n }\n }\n\n async getStats(filePath: string): Promise<FileInfo> {\n const fullPath = this.resolve(filePath);\n const stats = await fs.stat(fullPath);\n return { path: filePath, size: stats.size, modifiedAt: stats.mtime };\n }\n\n async listFiles(dirPath: string, extensions?: readonly string[]): Promise<string[]> {\n const fullPath = this.resolve(dirPath);\n const filterExts = extensions ?? SUPPORTED_EXTENSIONS;\n\n let entries: Dirent[];\n try {\n entries = await fs.readdir(fullPath, { withFileTypes: true });\n } catch {\n return [];\n }\n\n return entries\n .filter((entry) => entry.isFile() && filterExts.some((ext) => entry.name.endsWith(ext)))\n .map((entry) => this.join(dirPath, entry.name));\n }\n\n async listDirectories(dirPath: string): Promise<string[]> {\n const fullPath = this.resolve(dirPath);\n\n let entries: Dirent[];\n try {\n entries = await fs.readdir(fullPath, { withFileTypes: true });\n } catch {\n return [];\n }\n\n return entries\n .filter((entry) => entry.isDirectory())\n .map((entry) => this.join(dirPath, entry.name));\n }\n\n async readBuffer(filePath: string): Promise<Buffer> {\n return fs.readFile(this.resolve(filePath));\n }\n\n async writeBuffer(filePath: string, data: Buffer): Promise<void> {\n const fullPath = this.resolve(filePath);\n await fs.mkdir(path.dirname(fullPath), { recursive: true });\n await fs.writeFile(fullPath, data);\n }\n\n async listAllFiles(dirPath: string): Promise<DirectoryFileEntry[]> {\n const fullPath = this.resolve(dirPath);\n\n let entries: Dirent[];\n try {\n entries = await fs.readdir(fullPath, { withFileTypes: true });\n } catch {\n return [];\n }\n\n const results: DirectoryFileEntry[] = [];\n for (const entry of entries) {\n if (!entry.isFile()) continue;\n const relativePath = this.join(dirPath, entry.name);\n const stats = await fs.stat(this.resolve(relativePath));\n results.push({ name: entry.name, relativePath, size: stats.size, modifiedAt: stats.mtime });\n }\n return results;\n }\n\n join(...segments: string[]): string {\n return path.join(...segments);\n }\n\n basename(filePath: string): string {\n return path.basename(filePath);\n }\n\n extname(filePath: string): string {\n return path.extname(filePath);\n }\n\n relative(from: string, to: string): string {\n return path.relative(from, to);\n }\n\n normalizeSlug(relativePath: string, ext: string): string {\n return relativePath.replace(ext, \"\").split(path.sep).join(\"/\");\n }\n}\n","/**\n * @context Shared layer — field label utilities at src/shared/field-utils.ts\n * @does Resolves human-readable labels for field definitions and raw key strings\n * @depends src/shared/fields.ts\n * @do Add field-related utility functions here\n * @dont Import from CLI or UI; contain field type definitions or schema logic\n */\n\nimport type { BaseField } from \"./fields.js\";\n\n/**\n * Resolve the human-readable label for a field.\n *\n * When the field definition has an explicit `label`, that is returned as-is.\n * Otherwise the `name` (camelCase / kebab-case / snake_case) is converted to Title Case:\n *\n * @example\n * fieldLabel({ name: \"siteName\", type: \"text\" }) // \"Site Name\"\n * fieldLabel({ name: \"created_at\", type: \"date\" }) // \"Created At\"\n * fieldLabel({ name: \"bio\", type: \"long-text\", label: \"About\" }) // \"About\"\n */\nexport function fieldLabel(field: Pick<BaseField, \"name\" | \"label\">): string {\n if (field.label) return field.label;\n return field.name\n .replace(/[-_](.)/g, (_, c: string) => ` ${c.toUpperCase()}`)\n .replace(/([A-Z])/g, \" $1\")\n .trim()\n .replace(/\\b\\w/g, (c) => c.toUpperCase());\n}\n\n/**\n * Resolve the label for a raw key string (no field definition available).\n * Useful for dynamic keys that have no schema entry.\n */\nexport function keyLabel(name: string): string {\n return fieldLabel({ name });\n}\n"],"mappings":";AAQA,SAAS,QAAQ,SAAS,KAAK,aAAa;;;ACA5C,OAAO,aAAa;;;ACGb,IAAM,uBAAuB,CAAC,QAAQ,OAAO;AAC7C,IAAM,wBAAwB;AAI9B,IAAM,mBAAmB;AAAA,EAC9B;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AACF;AAEO,IAAM,mBAAmB,CAAC,aAAa,cAAc,WAAW;AAEhE,IAAM,mBAAmB;AAAA,EAC9B;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AACF;AAEO,IAAM,mBAAmB,CAAC,GAAG,kBAAkB,GAAG,kBAAkB,GAAG,gBAAgB;;;AC5B9F,OAAO,YAAY;AAOZ,SAAS,SAAS,SAA4B;AACnD,QAAM,EAAE,MAAM,SAAS,KAAK,IAAI,OAAO,OAAO;AAC9C,SAAO,EAAE,MAAM,MAAM,KAAK,KAAK,EAAE;AACnC;;;ACEO,SAAS,UAAU,SAA6B;AACrD,QAAM,SAAkB,KAAK,MAAM,OAAO;AAE1C,MAAI,MAAM,QAAQ,MAAM,GAAG;AACzB,WAAO;AAAA,MACL,MAAM;AAAA,MACN,SAAS;AAAA,IACX;AAAA,EACF;AAEA,MAAI,OAAO,WAAW,YAAY,WAAW,MAAM;AACjD,WAAO;AAAA,MACL,MAAM;AAAA,MACN,MAAM;AAAA,IACR;AAAA,EACF;AAEA,QAAM,IAAI,MAAM,yCAAyC;AAC3D;;;AC1BA,IAAM,cAAc;AACpB,IAAM,kBACJ;AACF,IAAM,WAAW;AACjB,IAAM,SAAS;AACf,IAAM,sBAAsB;AAE5B,SAAS,UAAU,OAAwB;AACzC,SAAO,YAAY,KAAK,KAAK;AAC/B;AAEA,SAAS,cAAc,OAAwB;AAC7C,SAAO,gBAAgB,KAAK,KAAK;AACnC;AAEA,SAAS,QAAQ,OAAwB;AACvC,SAAO,SAAS,KAAK,KAAK;AAC5B;AAEA,SAAS,MAAM,OAAwB;AACrC,SAAO,OAAO,KAAK,KAAK;AAC1B;AAEA,SAAS,iBAAiB,MAAc,SAAoC;AAC1E,MAAI,QAAQ,MAAM,OAAO,EAAG,QAAO,EAAE,MAAM,MAAM,QAAQ;AACzD,MAAI,QAAQ,MAAM,KAAK,EAAG,QAAO,EAAE,MAAM,MAAM,MAAM;AACrD,MAAI,QAAQ,MAAM,aAAa,EAAG,QAAO,EAAE,MAAM,MAAM,QAAQ,aAAa,KAAK;AACjF,MAAI,QAAQ,MAAM,SAAS,EAAG,QAAO,EAAE,MAAM,MAAM,OAAO;AAE1D,QAAM,SAAS,QAAQ,KAAK,CAAC,MAAM,EAAE,SAAS,uBAAuB,EAAE,SAAS,IAAI,CAAC;AACrF,SAAO,EAAE,MAAM,MAAM,SAAS,cAAc,OAAO;AACrD;AAEA,SAAS,gBAAgB,MAAc,OAAmC;AACxE,MAAI,MAAM,WAAW,EAAG,QAAO,EAAE,MAAM,MAAM,SAAS,YAAY,CAAC,EAAE;AAErE,MAAI,MAAM,MAAM,CAAC,SAAS,OAAO,SAAS,QAAQ,GAAG;AACnD,UAAM,SAAS,CAAC,GAAG,IAAI,IAAI,KAAiB,CAAC,EAAE,MAAM,GAAG,EAAE;AAC1D,UAAM,UAA0B,OAAO,IAAI,CAAC,OAAO,EAAE,OAAO,GAAG,OAAO,EAAE,EAAE;AAC1E,WAAO,EAAE,MAAM,MAAM,gBAAgB,QAAQ;AAAA,EAC/C;AAEA,MAAI,MAAM,MAAM,CAAC,SAAS,OAAO,SAAS,YAAY,SAAS,QAAQ,CAAC,MAAM,QAAQ,IAAI,CAAC,GAAG;AAC5F,WAAO,EAAE,MAAM,MAAM,SAAS,YAAY,YAAY,KAAkC,EAAE;AAAA,EAC5F;AAEA,SAAO,EAAE,MAAM,MAAM,SAAS,YAAY,CAAC,EAAE;AAC/C;AAEA,SAAS,qBAAqB,MAAc,QAAoC;AAC9E,QAAM,UAAU,OAAO,OAAO,CAAC,MAAM,MAAM,QAAQ,MAAM,MAAS;AAElE,MAAI,QAAQ,WAAW,EAAG,QAAO,EAAE,MAAM,MAAM,OAAO;AACtD,MAAI,QAAQ,MAAM,CAAC,MAAM,OAAO,MAAM,SAAS,EAAG,QAAO,EAAE,MAAM,MAAM,UAAU;AAEjF,MAAI,QAAQ,MAAM,CAAC,MAAM,OAAO,MAAM,QAAQ,GAAG;AAC/C,UAAM,SAAS,QAAQ,MAAM,CAAC,MAAM,OAAO,UAAU,CAAC,CAAC,IAAI,YAAY;AACvE,WAAO,EAAE,MAAM,MAAM,UAAU,OAAO;AAAA,EACxC;AAEA,MAAI,QAAQ,MAAM,CAAC,MAAM,OAAO,MAAM,QAAQ,GAAG;AAC/C,WAAO,iBAAiB,MAAM,OAAmB;AAAA,EACnD;AAEA,MAAI,QAAQ,MAAM,CAAC,MAAM,MAAM,QAAQ,CAAC,CAAC,GAAG;AAC1C,WAAO,gBAAgB,MAAO,QAAwB,KAAK,CAAC;AAAA,EAC9D;AAEA,MAAI,QAAQ,MAAM,CAAC,MAAM,OAAO,MAAM,YAAY,MAAM,QAAQ,CAAC,MAAM,QAAQ,CAAC,CAAC,GAAG;AAClF,WAAO,EAAE,MAAM,MAAM,UAAU,QAAQ,YAAY,OAAoC,EAAE;AAAA,EAC3F;AAEA,SAAO,EAAE,MAAM,MAAM,OAAO;AAC9B;AAEA,SAAS,YAAY,MAAoD;AACvE,QAAM,SAAS,IAAI,IAAY,KAAK,QAAQ,CAAC,QAAQ,OAAO,KAAK,GAAG,CAAC,CAAC;AACtE,SAAO,MAAM,KAAK,MAAM,EAAE,IAAI,CAAC,QAAQ,qBAAqB,KAAK,KAAK,IAAI,CAAC,QAAQ,IAAI,GAAG,CAAC,CAAC,CAAC;AAC/F;AASO,SAAS,YAAY,SAAyB,gBAA0C;AAC7F,QAAM,OAAO,QAAQ,IAAI,CAAC,UAAU,MAAM,IAA+B;AACzE,SAAO,EAAE,YAAY,gBAAgB,QAAQ,YAAY,IAAI,EAAE;AACjE;;;AJrFO,IAAM,eAAN,MAAmB;AAAA,EACP,UAAU,oBAAI,IAA4B;AAAA,EAC1C,cAAc,oBAAI,IAAwB;AAAA,EAC1C;AAAA,EAEjB,YAAY,WAAuB;AACjC,SAAK,KAAK;AAAA,EACZ;AAAA,EAEA,MAAM,MAAM,QAAsC;AAChD,SAAK,MAAM;AACX,UAAM,OAAO,MAAM,KAAK,GAAG,gBAAgB,GAAG;AAE9C,eAAW,OAAO,MAAM;AACtB,YAAM,UAAU,KAAK,GAAG,SAAS,GAAG;AACpC,YAAM,iBAAiB,QAAQ,OAAO;AACtC,YAAM,mBAAmB,QAAQ,cAAc,cAAc;AAC7D,YAAM,KAAK,gBAAgB,SAAS,gBAAgB,kBAAkB,MAAM;AAAA,IAC9E;AAAA,EACF;AAAA,EAEA,cAAc,MAA8B;AAC1C,WAAO,KAAK,QAAQ,IAAI,IAAI,KAAK,CAAC;AAAA,EACpC;AAAA,EAEA,iBAA+B;AAC7B,WAAO,MAAM,KAAK,KAAK,YAAY,OAAO,CAAC;AAAA,EAC7C;AAAA,EAEA,QAAc;AACZ,SAAK,QAAQ,MAAM;AACnB,SAAK,YAAY,MAAM;AAAA,EACzB;AAAA,EAEA,MAAc,gBACZ,SACA,gBACA,cACe;AACf,UAAM,UAA0B,CAAC;AACjC,UAAM,KAAK,QAAQ,SAAS,gBAAgB,SAAS,OAAO;AAE5D,UAAM,YAAY,KAAK,GAAG,KAAK,SAAS,qBAAqB;AAC7D,UAAM,WAAW,MAAM,KAAK,aAAa,SAAS;AAClD,QAAI,UAAU;AACZ,WAAK,cAAc,SAAS,QAAQ;AAAA,IACtC;AAEA,UAAM,SAAS,gBAAgB,YAAY,SAAS,cAAc;AAElE,SAAK,QAAQ,IAAI,gBAAgB,OAAO;AACxC,SAAK,YAAY,IAAI,gBAAgB;AAAA,MACnC,MAAM;AAAA,MACN,MAAM,KAAK,qBAAqB,OAAO;AAAA,MACvC,OAAO,QAAQ;AAAA,MACf,UAAU;AAAA,MACV;AAAA,IACF,CAAC;AAAA,EACH;AAAA,EAEA,MAAc,QACZ,SACA,gBACA,SACA,SACe;AACf,UAAM,UAAU,MAAM,KAAK,GAAG,gBAAgB,OAAO;AACrD,eAAW,UAAU,SAAS;AAC5B,YAAM,KAAK,QAAQ,SAAS,gBAAgB,QAAQ,OAAO;AAAA,IAC7D;AAEA,UAAM,QAAQ,MAAM,KAAK,GAAG,UAAU,OAAO;AAC7C,eAAW,YAAY,OAAO;AAC5B,YAAM,WAAW,KAAK,GAAG,SAAS,QAAQ;AAC1C,UAAI,aAAa,sBAAuB;AAExC,YAAM,MAAM,KAAK,GAAG,QAAQ,QAAQ;AACpC,YAAM,UAAU,MAAM,KAAK,GAAG,SAAS,QAAQ;AAC/C,YAAM,eAAe,KAAK,GAAG,SAAS,SAAS,QAAQ;AACvD,YAAM,OAAO,KAAK,GACf,cAAc,cAAc,GAAG,EAC/B,MAAM,GAAG,EACT,IAAI,CAAC,YAAY,QAAQ,OAAO,CAAC,EACjC,KAAK,GAAG;AAEX,UAAI,QAAQ,QAAQ;AAClB,gBAAQ,KAAK,KAAK,cAAc,gBAAgB,MAAM,OAAO,CAAC;AAAA,MAChE,WAAW,QAAQ,SAAS;AAC1B,gBAAQ,KAAK,GAAG,KAAK,iBAAiB,gBAAgB,MAAM,OAAO,CAAC;AAAA,MACtE;AAAA,IACF;AAAA,EACF;AAAA,EAEQ,cAAc,gBAAwB,MAAc,SAA+B;AACzF,UAAM,SAAS,SAAS,OAAO;AAC/B,WAAO;AAAA,MACL,YAAY;AAAA,MACZ;AAAA,MACA,MAAM,IAAI,cAAc,IAAI,IAAI;AAAA,MAChC,MAAM,OAAO;AAAA,MACb,MAAM,OAAO;AAAA,IACf;AAAA,EACF;AAAA,EAEQ,iBAAiB,gBAAwB,MAAc,SAAiC;AAC9F,UAAM,SAAS,UAAU,OAAO;AAEhC,QAAI,OAAO,SAAS,cAAc;AAChC,aAAO,OAAO,QAAQ,IAAI,CAAC,MAAM,UAAU;AACzC,cAAM,YACJ,OAAO,KAAK,MAAM,MAAM,WAAW,QAAQ,KAAK,MAAM,CAAC,IAAI,GAAG,IAAI,IAAI,KAAK;AAC7E,eAAO;AAAA,UACL,YAAY;AAAA,UACZ,MAAM;AAAA,UACN,MAAM,IAAI,cAAc,IAAI,SAAS;AAAA,UACrC;AAAA,QACF;AAAA,MACF,CAAC;AAAA,IACH;AAEA,WAAO,CAAC,EAAE,YAAY,gBAAgB,MAAM,MAAM,IAAI,cAAc,IAAI,IAAI,IAAI,MAAM,OAAO,KAAK,CAAC;AAAA,EACrG;AAAA,EAEA,MAAc,aAAa,WAA6C;AACtE,QAAI,CAAE,MAAM,KAAK,GAAG,OAAO,SAAS,EAAI,QAAO;AAE/C,QAAI;AACF,YAAM,UAAU,MAAM,KAAK,GAAG,SAAS,SAAS;AAChD,YAAM,SAAkB,KAAK,MAAM,OAAO;AAC1C,UAAI,MAAM,QAAQ,MAAM,EAAG,QAAO;AAAA,IACpC,SAAS,OAAO;AACd,cAAQ,KAAK,kDAAkD,SAAS,IAAI,KAAK;AAAA,IACnF;AACA,WAAO;AAAA,EACT;AAAA,EAEQ,cAAc,SAAyB,UAA0B;AACvE,UAAM,WAAW,IAAI,IAAI,SAAS,IAAI,CAAC,MAAM,UAAU,CAAC,MAAM,KAAK,CAAC,CAAC;AACrE,YAAQ,KAAK,CAAC,GAAG,MAAM;AACrB,YAAM,SAAS,SAAS,IAAI,EAAE,IAAI,KAAK;AACvC,YAAM,SAAS,SAAS,IAAI,EAAE,IAAI,KAAK;AACvC,aAAO,SAAS;AAAA,IAClB,CAAC;AAAA,EACH;AAAA,EAEQ,qBAAqB,SAA6C;AACxE,QAAI,QAAQ,WAAW,EAAG,QAAO;AACjC,UAAM,QAAQ,QAAQ,CAAC;AACvB,QAAI,MAAM,SAAS,OAAW,QAAO;AACrC,QAAI,QAAQ,WAAW,KAAK,CAAC,MAAM,KAAK,SAAS,GAAG,EAAG,QAAO;AAC9D,WAAO;AAAA,EACT;AACF;;;AK7JA,IAAI,QAA6B;AAE1B,SAAS,WAAyB;AACvC,MAAI,CAAC,OAAO;AACV,UAAM,IAAI,MAAM,yDAAyD;AAAA,EAC3E;AACA,SAAO;AACT;AAEA,eAAsB,YACpB,WACA,QACuB;AACvB,QAAM,QAAQ,IAAI,aAAa,SAAS;AACxC,QAAM,MAAM,MAAM,MAAM;AACxB,UAAQ;AACR,SAAO;AACT;;;ANDO,IAAM,eAAN,MAAmB;AAAA,EACP;AAAA,EACT,UAAwB,CAAC;AAAA,EAEjC,YAAY,YAAoB;AAC9B,SAAK,iBAAiB;AAAA,EACxB;AAAA,EAEA,MAAM,YAA2C;AAC/C,SAAK,QAAQ,QAAQ,EAAE,GAAG,KAAK,QAAQ,OAAO,GAAG,WAAW;AAC5D,WAAO;AAAA,EACT;AAAA,EAEA,KAAK,OAAe,QAAwB,OAAa;AACvD,SAAK,QAAQ,OAAO,EAAE,OAAO,MAAM;AACnC,WAAO;AAAA,EACT;AAAA,EAEA,MAAM,OAAqB;AACzB,SAAK,QAAQ,QAAQ;AACrB,WAAO;AAAA,EACT;AAAA,EAEA,OAAO,OAAqB;AAC1B,SAAK,QAAQ,SAAS;AACtB,WAAO;AAAA,EACT;AAAA,EAEA,MAAsB;AACpB,QAAI,UAAU,CAAC,GAAG,SAAS,EAAE,cAAc,KAAK,cAAc,CAAC;AAE/D,QAAI,KAAK,QAAQ,OAAO;AACtB,YAAM,aAAa,KAAK,QAAQ;AAChC,gBAAU;AAAA,QAAO;AAAA,QAAS,CAAC,UACzB,OAAO,QAAQ,UAAU,EAAE,MAAM,CAAC,CAAC,KAAK,KAAK,MAAM,IAAI,MAAM,MAAM,GAAG,MAAM,KAAK;AAAA,MACnF;AAAA,IACF;AAEA,QAAI,KAAK,QAAQ,MAAM;AACrB,YAAM,EAAE,OAAO,MAAM,IAAI,KAAK,QAAQ;AACtC,gBAAU,QAAQ,SAAS,CAAC,CAAC,UAAU,IAAI,MAAM,MAAM,KAAK,CAAC,GAAG,CAAC,KAAK,CAAC;AAAA,IACzE;AAEA,UAAM,QAAQ,KAAK,QAAQ,UAAU;AACrC,UAAM,MAAM,KAAK,QAAQ,QAAQ,QAAQ,KAAK,QAAQ,QAAQ;AAC9D,WAAO,MAAM,SAAS,OAAO,GAAG;AAAA,EAClC;AAAA,EAEA,QAAkC;AAChC,WAAO,KAAK,MAAM,CAAC,EAAE,IAAI,EAAE,CAAC;AAAA,EAC9B;AAAA,EAEA,QAAgB;AACd,WAAO,KAAK,IAAI,EAAE;AAAA,EACpB;AACF;AAKO,SAAS,gBAAgB,YAAkC;AAChE,SAAO,IAAI,aAAa,UAAU;AACpC;;;AOlFA,OAAOA,WAAU;;;ACAjB,OAAO,QAAQ;AACf,OAAO,UAAU;AAMV,IAAM,YAAN,MAAsC;AAAA,EAC1B;AAAA,EAEjB,YAAY,UAAkB;AAC5B,SAAK,WAAW,KAAK,QAAQ,QAAQ;AAAA,EACvC;AAAA,EAEQ,WAAW,UAA4B;AAC7C,WAAO,KAAK,QAAQ,KAAK,UAAU,GAAG,QAAQ;AAAA,EAChD;AAAA,EAEA,MAAM,SAAS,UAAmC;AAChD,WAAO,GAAG,SAAS,KAAK,QAAQ,QAAQ,GAAG,OAAO;AAAA,EACpD;AAAA,EAEA,MAAM,UAAU,UAAkB,SAAgC;AAChE,UAAM,WAAW,KAAK,QAAQ,QAAQ;AACtC,UAAM,GAAG,MAAM,KAAK,QAAQ,QAAQ,GAAG,EAAE,WAAW,KAAK,CAAC;AAC1D,UAAM,GAAG,UAAU,UAAU,SAAS,OAAO;AAAA,EAC/C;AAAA,EAEA,MAAM,WAAW,UAAiC;AAChD,UAAM,GAAG,OAAO,KAAK,QAAQ,QAAQ,CAAC;AAAA,EACxC;AAAA,EAEA,MAAM,OAAO,UAAoC;AAC/C,QAAI;AACF,YAAM,GAAG,OAAO,KAAK,QAAQ,QAAQ,CAAC;AACtC,aAAO;AAAA,IACT,QAAQ;AACN,aAAO;AAAA,IACT;AAAA,EACF;AAAA,EAEA,MAAM,SAAS,UAAqC;AAClD,UAAM,WAAW,KAAK,QAAQ,QAAQ;AACtC,UAAM,QAAQ,MAAM,GAAG,KAAK,QAAQ;AACpC,WAAO,EAAE,MAAM,UAAU,MAAM,MAAM,MAAM,YAAY,MAAM,MAAM;AAAA,EACrE;AAAA,EAEA,MAAM,UAAU,SAAiB,YAAmD;AAClF,UAAM,WAAW,KAAK,QAAQ,OAAO;AACrC,UAAM,aAAa,cAAc;AAEjC,QAAI;AACJ,QAAI;AACF,gBAAU,MAAM,GAAG,QAAQ,UAAU,EAAE,eAAe,KAAK,CAAC;AAAA,IAC9D,QAAQ;AACN,aAAO,CAAC;AAAA,IACV;AAEA,WAAO,QACJ,OAAO,CAAC,UAAU,MAAM,OAAO,KAAK,WAAW,KAAK,CAAC,QAAQ,MAAM,KAAK,SAAS,GAAG,CAAC,CAAC,EACtF,IAAI,CAAC,UAAU,KAAK,KAAK,SAAS,MAAM,IAAI,CAAC;AAAA,EAClD;AAAA,EAEA,MAAM,gBAAgB,SAAoC;AACxD,UAAM,WAAW,KAAK,QAAQ,OAAO;AAErC,QAAI;AACJ,QAAI;AACF,gBAAU,MAAM,GAAG,QAAQ,UAAU,EAAE,eAAe,KAAK,CAAC;AAAA,IAC9D,QAAQ;AACN,aAAO,CAAC;AAAA,IACV;AAEA,WAAO,QACJ,OAAO,CAAC,UAAU,MAAM,YAAY,CAAC,EACrC,IAAI,CAAC,UAAU,KAAK,KAAK,SAAS,MAAM,IAAI,CAAC;AAAA,EAClD;AAAA,EAEA,MAAM,WAAW,UAAmC;AAClD,WAAO,GAAG,SAAS,KAAK,QAAQ,QAAQ,CAAC;AAAA,EAC3C;AAAA,EAEA,MAAM,YAAY,UAAkB,MAA6B;AAC/D,UAAM,WAAW,KAAK,QAAQ,QAAQ;AACtC,UAAM,GAAG,MAAM,KAAK,QAAQ,QAAQ,GAAG,EAAE,WAAW,KAAK,CAAC;AAC1D,UAAM,GAAG,UAAU,UAAU,IAAI;AAAA,EACnC;AAAA,EAEA,MAAM,aAAa,SAAgD;AACjE,UAAM,WAAW,KAAK,QAAQ,OAAO;AAErC,QAAI;AACJ,QAAI;AACF,gBAAU,MAAM,GAAG,QAAQ,UAAU,EAAE,eAAe,KAAK,CAAC;AAAA,IAC9D,QAAQ;AACN,aAAO,CAAC;AAAA,IACV;AAEA,UAAM,UAAgC,CAAC;AACvC,eAAW,SAAS,SAAS;AAC3B,UAAI,CAAC,MAAM,OAAO,EAAG;AACrB,YAAM,eAAe,KAAK,KAAK,SAAS,MAAM,IAAI;AAClD,YAAM,QAAQ,MAAM,GAAG,KAAK,KAAK,QAAQ,YAAY,CAAC;AACtD,cAAQ,KAAK,EAAE,MAAM,MAAM,MAAM,cAAc,MAAM,MAAM,MAAM,YAAY,MAAM,MAAM,CAAC;AAAA,IAC5F;AACA,WAAO;AAAA,EACT;AAAA,EAEA,QAAQ,UAA4B;AAClC,WAAO,KAAK,KAAK,GAAG,QAAQ;AAAA,EAC9B;AAAA,EAEA,SAAS,UAA0B;AACjC,WAAO,KAAK,SAAS,QAAQ;AAAA,EAC/B;AAAA,EAEA,QAAQ,UAA0B;AAChC,WAAO,KAAK,QAAQ,QAAQ;AAAA,EAC9B;AAAA,EAEA,SAAS,MAAc,IAAoB;AACzC,WAAO,KAAK,SAAS,MAAM,EAAE;AAAA,EAC/B;AAAA,EAEA,cAAc,cAAsB,KAAqB;AACvD,WAAO,aAAa,QAAQ,KAAK,EAAE,EAAE,MAAM,KAAK,GAAG,EAAE,KAAK,GAAG;AAAA,EAC/D;AACF;;;AD3GA,eAAsB,WACpB,aACA,QACe;AACf,QAAM,MAAM,eAAeC,MAAK,KAAK,QAAQ,IAAI,GAAG,UAAU;AAC9D,QAAM,YAAY,IAAI,UAAU,GAAG,GAAG,MAAM;AAC9C;AAKO,SAAS,sBAA+B;AAC7C,MAAI;AACF,aAAS;AACT,WAAO;AAAA,EACT,QAAQ;AACN,WAAO;AAAA,EACT;AACF;;;AEzBO,SAAS,WAAW,OAAkD;AAC3E,MAAI,MAAM,MAAO,QAAO,MAAM;AAC9B,SAAO,MAAM,KACV,QAAQ,YAAY,CAAC,GAAG,MAAc,IAAI,EAAE,YAAY,CAAC,EAAE,EAC3D,QAAQ,YAAY,KAAK,EACzB,KAAK,EACL,QAAQ,SAAS,CAAC,MAAM,EAAE,YAAY,CAAC;AAC5C;AAMO,SAAS,SAAS,MAAsB;AAC7C,SAAO,WAAW,EAAE,KAAK,CAAC;AAC5B;","names":["path","path"]}
1
+ {"version":3,"sources":["../../src/core/query-builder.ts","../../src/core/indexer.ts","../../src/shared/constants.ts","../../src/core/parsers/parser-mdx.ts","../../src/core/parsers/parser-json.ts","../../src/core/schema-inferrer.ts","../../src/core/content-store.ts","../../src/core/init.ts","../../src/cli/adapters/fs-adapter.ts","../../src/shared/field-utils.ts"],"sourcesContent":["/**\n * @context Core layer — query builder at src/core/query-builder.ts\n * @does Provides a fluent API to filter, sort, and paginate content entries from a collection\n * @depends src/shared/types.ts, src/core/content-store.ts\n * @do Add new query capabilities here (e.g. search, groupBy)\n * @dont Import from CLI or UI; access the filesystem; perform I/O\n */\n\nimport { filter, orderBy, get, slice } from \"lodash-es\";\nimport type { QueryOptions } from \"../shared/types.js\";\nimport type { CollectionTypeMap } from \"../shared/types.js\";\nimport { getStore } from \"./content-store.js\";\n\n/**\n * Fluent query builder for content collections.\n *\n * ```ts\n * const posts = queryCollection(\"blog\")\n * .where({ published: true })\n * .sort(\"date\", \"desc\")\n * .limit(10)\n * .all();\n * ```\n *\n * Supports dot notation for nested properties:\n * ```ts\n * queryCollection(\"pages\").where({ \"hero.title\": \"Welcome\" }).all();\n * ```\n */\nexport class QueryBuilder<T = Record<string, unknown>> {\n private readonly collectionName: string;\n private options: QueryOptions = {};\n\n constructor(collection: string) {\n this.collectionName = collection;\n }\n\n where(conditions: Record<string, unknown>): this {\n this.options.where = { ...this.options.where, ...conditions };\n return this;\n }\n\n sort(field: string, order: \"asc\" | \"desc\" = \"asc\"): this {\n this.options.sort = { field, order };\n return this;\n }\n\n limit(count: number): this {\n this.options.limit = count;\n return this;\n }\n\n offset(count: number): this {\n this.options.offset = count;\n return this;\n }\n\n all(): T[] {\n let entries = [...getStore().getCollection(this.collectionName)];\n\n if (this.options.where) {\n const conditions = this.options.where;\n entries = filter(entries, (entry) =>\n Object.entries(conditions).every(([key, value]) => get(entry.data, key) === value),\n );\n }\n\n if (this.options.sort) {\n const { field, order } = this.options.sort;\n entries = orderBy(entries, [(entry) => get(entry.data, field)], [order]);\n }\n\n const start = this.options.offset ?? 0;\n const end = this.options.limit ? start + this.options.limit : undefined;\n return slice(entries, start, end).map((e) => e.data as unknown as T);\n }\n\n first(): T | undefined {\n return this.limit(1).all()[0];\n }\n\n count(): number {\n return this.all().length;\n }\n}\n\n/**\n * Entry point for querying a content collection.\n */\nexport function queryCollection<K extends keyof CollectionTypeMap>(\n name: K,\n): QueryBuilder<CollectionTypeMap[K]>;\nexport function queryCollection(name: string): QueryBuilder<Record<string, unknown>>;\nexport function queryCollection(name: string): QueryBuilder {\n return new QueryBuilder(name);\n}\n","/**\n * @context Core layer — content indexer at src/core/indexer.ts\n * @does Scans the contents directory, parses MDX/JSON files, and builds an in-memory index\n * @depends src/shared/types.ts, src/shared/constants.ts, src/shared/fs-adapter.interface.ts, src/core/parsers/, src/core/schema-inferrer.ts\n * @do Add new file type handling here; extend indexCollection for new collection behaviors\n * @dont Import from CLI or UI; instantiate FsAdapter; access the filesystem directly\n */\n\nimport slugify from \"@sindresorhus/slugify\";\nimport type { CollectionSchema } from \"../shared/fields.js\";\nimport type { ContentEntry, Collection, StudioConfig } from \"../shared/types.js\";\nimport type { IFsAdapter } from \"../shared/fs-adapter.interface.js\";\nimport { COLLECTION_ORDER_FILE } from \"../shared/constants.js\";\nimport { parseMdx } from \"./parsers/parser-mdx.js\";\nimport { parseJson } from \"./parsers/parser-json.js\";\nimport { inferSchema } from \"./schema-inferrer.js\";\n\nexport class ContentIndex {\n private readonly entries = new Map<string, ContentEntry[]>();\n private readonly collections = new Map<string, Collection>();\n private readonly fs: IFsAdapter;\n\n constructor(fsAdapter: IFsAdapter) {\n this.fs = fsAdapter;\n }\n\n async build(config?: StudioConfig): Promise<void> {\n this.clear();\n const dirs = await this.fs.listDirectories(\".\");\n\n for (const dir of dirs) {\n const dirName = this.fs.basename(dir);\n const collectionName = slugify(dirName);\n const collectionConfig = config?.collections?.[collectionName];\n await this.indexCollection(dirName, collectionName, collectionConfig?.schema);\n }\n }\n\n getCollection(name: string): ContentEntry[] {\n return this.entries.get(name) ?? [];\n }\n\n getCollections(): Collection[] {\n return Array.from(this.collections.values());\n }\n\n clear(): void {\n this.entries.clear();\n this.collections.clear();\n }\n\n private async indexCollection(\n dirName: string,\n collectionName: string,\n manualSchema?: CollectionSchema,\n ): Promise<void> {\n const entries: ContentEntry[] = [];\n await this.scanDir(dirName, collectionName, dirName, entries);\n\n const orderPath = this.fs.join(dirName, COLLECTION_ORDER_FILE);\n const ordering = await this.readOrdering(orderPath);\n if (ordering) {\n this.applyOrdering(entries, ordering);\n }\n\n const schema = manualSchema ?? inferSchema(entries, collectionName);\n\n this.entries.set(collectionName, entries);\n this.collections.set(collectionName, {\n name: collectionName,\n type: this.detectCollectionType(entries),\n count: entries.length,\n basePath: dirName,\n schema,\n });\n }\n\n private async scanDir(\n dirName: string,\n collectionName: string,\n dirPath: string,\n entries: ContentEntry[],\n ): Promise<void> {\n const subDirs = await this.fs.listDirectories(dirPath);\n for (const subDir of subDirs) {\n await this.scanDir(dirName, collectionName, subDir, entries);\n }\n\n const files = await this.fs.listFiles(dirPath);\n for (const filePath of files) {\n const fileName = this.fs.basename(filePath);\n if (fileName === COLLECTION_ORDER_FILE) continue;\n\n const ext = this.fs.extname(fileName);\n const content = await this.fs.readFile(filePath);\n const relativePath = this.fs.relative(dirName, filePath);\n const slug = this.fs\n .normalizeSlug(relativePath, ext)\n .split(\"/\")\n .map((segment) => slugify(segment))\n .join(\"/\");\n\n if (ext === \".mdx\") {\n entries.push(this.buildMdxEntry(collectionName, slug, content));\n } else if (ext === \".json\") {\n entries.push(...this.buildJsonEntries(collectionName, slug, content));\n }\n }\n }\n\n private buildMdxEntry(collectionName: string, slug: string, content: string): ContentEntry {\n const parsed = parseMdx(content);\n return {\n collection: collectionName,\n slug,\n path: `/${collectionName}/${slug}`,\n body: parsed.body,\n data: parsed.data,\n };\n }\n\n private buildJsonEntries(collectionName: string, slug: string, content: string): ContentEntry[] {\n const parsed = parseJson(content);\n\n if (parsed.type === \"json-array\") {\n return parsed.entries.map((data, index) => {\n const entrySlug =\n typeof data[\"slug\"] === \"string\" ? slugify(data[\"slug\"]) : `${slug}/${index}`;\n return {\n collection: collectionName,\n slug: entrySlug,\n path: `/${collectionName}/${entrySlug}`,\n data,\n };\n });\n }\n\n return [{ collection: collectionName, slug, path: `/${collectionName}/${slug}`, data: parsed.data }];\n }\n\n private async readOrdering(orderPath: string): Promise<string[] | null> {\n if (!(await this.fs.exists(orderPath))) return null;\n\n try {\n const content = await this.fs.readFile(orderPath);\n const parsed: unknown = JSON.parse(content);\n if (Array.isArray(parsed)) return parsed as string[];\n } catch (error) {\n console.warn(`[Nextjs Studio] Failed to parse ordering file: ${orderPath}`, error);\n }\n return null;\n }\n\n private applyOrdering(entries: ContentEntry[], ordering: string[]): void {\n const orderMap = new Map(ordering.map((slug, index) => [slug, index]));\n entries.sort((a, b) => {\n const aIndex = orderMap.get(a.slug) ?? Infinity;\n const bIndex = orderMap.get(b.slug) ?? Infinity;\n return aIndex - bIndex;\n });\n }\n\n private detectCollectionType(entries: ContentEntry[]): Collection[\"type\"] {\n if (entries.length === 0) return \"mdx\";\n const first = entries[0];\n if (first.body !== undefined) return \"mdx\";\n if (entries.length === 1 && !first.slug.includes(\"/\")) return \"json-object\";\n return \"json-array\";\n }\n}\n","/**\n * @context Shared layer — constants at src/shared/constants.ts\n * @does Defines project-wide constants shared across core, CLI, and UI layers\n * @depends none\n * @do Add new shared constants here\n * @dont Import from CLI or UI; constants must be framework-agnostic\n */\n\nexport const CONTENTS_DIR = \"contents\";\nexport const CLI_PORT = 3030;\nexport const CONFIG_FILE = \"studio.config.ts\";\nexport const SUPPORTED_EXTENSIONS = [\".mdx\", \".json\"] as const;\nexport const COLLECTION_ORDER_FILE = \"collection.json\";\nexport const WATCHER_DEBOUNCE_MS = 5_000;\nexport const MEDIA_DIR = \"media\";\n\nexport const IMAGE_MIME_TYPES = [\n \"image/png\",\n \"image/jpeg\",\n \"image/gif\",\n \"image/webp\",\n \"image/svg+xml\",\n \"image/avif\",\n] as const;\n\nexport const VIDEO_MIME_TYPES = [\"video/mp4\", \"video/webm\", \"video/ogg\"] as const;\n\nexport const AUDIO_MIME_TYPES = [\n \"audio/mpeg\",\n \"audio/ogg\",\n \"audio/wav\",\n \"audio/webm\",\n \"audio/aac\",\n \"audio/flac\",\n] as const;\n\nexport const MEDIA_MIME_TYPES = [...IMAGE_MIME_TYPES, ...VIDEO_MIME_TYPES, ...AUDIO_MIME_TYPES] as const;\n\nexport const IMAGE_EXTENSIONS = [\".png\", \".jpg\", \".jpeg\", \".gif\", \".webp\", \".svg\", \".avif\"] as const;\nexport const VIDEO_EXTENSIONS = [\".mp4\", \".webm\", \".ogv\"] as const;\nexport const AUDIO_EXTENSIONS = [\".mp3\", \".ogg\", \".wav\", \".m4a\", \".aac\", \".flac\"] as const;\n","/**\n * @context Core layer — MDX parser/serializer at src/core/parsers/parser-mdx.ts\n * @does Parses .mdx content into frontmatter + body, and serializes them back to MDX strings\n * @depends none (gray-matter is an external dep)\n * @do Add MDX transform steps here; both parse and serialize live here intentionally\n * @dont Access the filesystem; import from CLI or UI; handle JSON content\n */\n\nimport matter from \"gray-matter\";\n\nexport interface ParsedMdx {\n data: Record<string, unknown>;\n body: string;\n}\n\nexport function parseMdx(content: string): ParsedMdx {\n const { data, content: body } = matter(content);\n return { data, body: body.trim() };\n}\n\nexport function serializeMdx(data: Record<string, unknown>, body: string): string {\n return matter.stringify(body, data);\n}\n","/**\n * @context Core layer — JSON parser at src/core/parsers/parser-json.ts\n * @does Parses JSON content strings into typed ParsedJson results (array or object)\n * @depends none\n * @do Extend ParsedJson variants here if new JSON structures are supported\n * @dont Access the filesystem; import from CLI or UI; contain serialization logic\n */\n\nexport interface ParsedJsonArray {\n type: \"json-array\";\n entries: Record<string, unknown>[];\n}\n\nexport interface ParsedJsonObject {\n type: \"json-object\";\n data: Record<string, unknown>;\n}\n\nexport type ParsedJson = ParsedJsonArray | ParsedJsonObject;\n\nexport function parseJson(content: string): ParsedJson {\n const parsed: unknown = JSON.parse(content);\n\n if (Array.isArray(parsed)) {\n return {\n type: \"json-array\",\n entries: parsed as Record<string, unknown>[],\n };\n }\n\n if (typeof parsed === \"object\" && parsed !== null) {\n return {\n type: \"json-object\",\n data: parsed as Record<string, unknown>,\n };\n }\n\n throw new Error(\"JSON content must be an array or object\");\n}\n","/**\n * @context Core layer — schema inferrer at src/core/schema-inferrer.ts\n * @does Infers a CollectionSchema from actual content entries when no manual schema is defined\n * @depends src/shared/types.ts, src/shared/fields.ts\n * @do Add new type detection heuristics here (e.g. color, phone)\n * @dont Import from CLI or UI; access the filesystem; perform I/O\n */\n\nimport type { ContentEntry } from \"../shared/types.js\";\nimport type { CollectionSchema, FieldDefinition, SelectOption } from \"../shared/fields.js\";\n\n// Value detector patterns\nconst RE_ISO_DATE = /^\\d{4}-\\d{2}-\\d{2}$/;\nconst RE_ISO_DATETIME =\n /^\\d{4}-\\d{2}-\\d{2}[T ]\\d{2}:\\d{2}(:\\d{2}(\\.\\d+)?)?(Z|[+-]\\d{2}:?\\d{2})?$/;\nconst RE_EMAIL = /^[^\\s@]+@[^\\s@]+\\.[^\\s@]+$/;\nconst RE_URL = /^https?:\\/\\/.+/;\nconst LONG_TEXT_THRESHOLD = 200;\n\nfunction isISODate(value: string): boolean {\n return RE_ISO_DATE.test(value);\n}\n\nfunction isISODateTime(value: string): boolean {\n return RE_ISO_DATETIME.test(value);\n}\n\nfunction isEmail(value: string): boolean {\n return RE_EMAIL.test(value);\n}\n\nfunction isUrl(value: string): boolean {\n return RE_URL.test(value);\n}\n\nfunction inferStringField(name: string, strings: string[]): FieldDefinition {\n if (strings.every(isEmail)) return { name, type: \"email\" };\n if (strings.every(isUrl)) return { name, type: \"url\" };\n if (strings.every(isISODateTime)) return { name, type: \"date\", includeTime: true };\n if (strings.every(isISODate)) return { name, type: \"date\" };\n\n const isLong = strings.some((s) => s.length > LONG_TEXT_THRESHOLD || s.includes(\"\\n\"));\n return { name, type: isLong ? \"long-text\" : \"text\" };\n}\n\nfunction inferArrayField(name: string, items: unknown[]): FieldDefinition {\n if (items.length === 0) return { name, type: \"array\", itemFields: [] };\n\n if (items.every((item) => typeof item === \"string\")) {\n const unique = [...new Set(items as string[])].slice(0, 50);\n const options: SelectOption[] = unique.map((v) => ({ label: v, value: v }));\n return { name, type: \"multi-select\", options };\n }\n\n if (items.every((item) => typeof item === \"object\" && item !== null && !Array.isArray(item))) {\n return { name, type: \"array\", itemFields: inferFields(items as Record<string, unknown>[]) };\n }\n\n return { name, type: \"array\", itemFields: [] };\n}\n\nfunction inferFieldDefinition(name: string, values: unknown[]): FieldDefinition {\n const present = values.filter((v) => v !== null && v !== undefined);\n\n if (present.length === 0) return { name, type: \"text\" };\n if (present.every((v) => typeof v === \"boolean\")) return { name, type: \"boolean\" };\n\n if (present.every((v) => typeof v === \"number\")) {\n const format = present.every((v) => Number.isInteger(v)) ? \"integer\" : \"decimal\";\n return { name, type: \"number\", format };\n }\n\n if (present.every((v) => typeof v === \"string\")) {\n return inferStringField(name, present as string[]);\n }\n\n if (present.every((v) => Array.isArray(v))) {\n return inferArrayField(name, (present as unknown[][]).flat());\n }\n\n if (present.every((v) => typeof v === \"object\" && v !== null && !Array.isArray(v))) {\n return { name, type: \"object\", fields: inferFields(present as Record<string, unknown>[]) };\n }\n\n return { name, type: \"text\" };\n}\n\nfunction inferFields(rows: Record<string, unknown>[]): FieldDefinition[] {\n const keySet = new Set<string>(rows.flatMap((row) => Object.keys(row)));\n return Array.from(keySet).map((key) => inferFieldDefinition(key, rows.map((row) => row[key])));\n}\n\n/**\n * Infer a `CollectionSchema` from the data of a set of content entries.\n *\n * The result is a best-effort approximation — string fields that look like\n * emails, URLs, or ISO dates get the correct semantic type. Everything else\n * falls back to `text`.\n */\nexport function inferSchema(entries: ContentEntry[], collectionName: string): CollectionSchema {\n const rows = entries.map((entry) => entry.data as Record<string, unknown>);\n return { collection: collectionName, fields: inferFields(rows) };\n}\n","/**\n * @context Core layer — content store at src/core/content-store.ts\n * @does Manages a singleton ContentIndex; exposes loadContent() and getStore() for consumers\n * @depends src/core/indexer.ts, src/shared/fs-adapter.interface.ts, src/shared/types.ts\n * @do Use this as the single access point for in-memory indexed content\n * @dont Import from CLI or UI; instantiate FsAdapter here; contain parsing or I/O logic\n */\n\nimport type { IFsAdapter } from \"../shared/fs-adapter.interface.js\";\nimport type { StudioConfig } from \"../shared/types.js\";\nimport { ContentIndex } from \"./indexer.js\";\n\nlet store: ContentIndex | null = null;\n\nexport function getStore(): ContentIndex {\n if (!store) {\n throw new Error(\"Content not loaded. Call loadContent() before querying.\");\n }\n return store;\n}\n\nexport async function loadContent(\n fsAdapter: IFsAdapter,\n config?: StudioConfig,\n): Promise<ContentIndex> {\n const index = new ContentIndex(fsAdapter);\n await index.build(config);\n store = index;\n return index;\n}\n","/**\r\n * @context Core layer — studio initializer at src/core/init.ts\r\n * @does Convenience function to initialize the content store for Node.js environments\r\n * @depends src/core/content-store.ts, src/cli/adapters/fs-adapter.ts\r\n * @do Use this as the entry point for Next.js and other Node.js consumers\r\n * @dont Import from UI; use in browser environments\r\n */\r\n\r\nimport path from \"node:path\";\r\nimport { FsAdapter } from \"../cli/adapters/fs-adapter.js\";\r\nimport { loadContent, getStore } from \"./content-store.js\";\r\nimport type { StudioConfig } from \"../shared/types.js\";\r\n\r\n/**\r\n * Initialize the content store from the filesystem.\r\n * Call this once before using `queryCollection()`.\r\n *\r\n * @param contentsDir - Path to the contents directory. Defaults to `./contents` relative to cwd.\r\n * @param config - Optional studio config for schemas and scripts.\r\n *\r\n * @example\r\n * ```ts\r\n * import { initStudio, queryCollection } from \"nextjs-studio\";\r\n *\r\n * await initStudio();\r\n * const posts = queryCollection(\"posts\").all();\r\n * ```\r\n */\r\nexport async function initStudio(\r\n contentsDir?: string,\r\n config?: StudioConfig,\r\n): Promise<void> {\r\n const dir = contentsDir ?? path.join(process.cwd(), \"contents\");\r\n await loadContent(new FsAdapter(dir), config);\r\n}\r\n\r\n/**\r\n * Returns true if the content store has been initialized.\r\n */\r\nexport function isStudioInitialized(): boolean {\r\n try {\r\n getStore();\r\n return true;\r\n } catch {\r\n return false;\r\n }\r\n}\r\n\r\n/**\r\n * Ensures the content store is initialized. Safe to call multiple times — only\r\n * initializes once. Ideal for use at the top of Next.js server components and\r\n * `generateStaticParams` / `generateMetadata` functions.\r\n *\r\n * @param contentsDir - Path to the contents directory. Defaults to `./contents` relative to cwd.\r\n * @param config - Optional studio config for schemas and scripts.\r\n *\r\n * @example\r\n * ```ts\r\n * import { ensureContentLoaded, queryCollection } from \"nextjs-studio\";\r\n *\r\n * export default async function Page() {\r\n * await ensureContentLoaded();\r\n * const posts = queryCollection(\"posts\");\r\n * // ...\r\n * }\r\n * ```\r\n */\r\nexport async function ensureContentLoaded(\r\n contentsDir?: string,\r\n config?: StudioConfig,\r\n): Promise<void> {\r\n if (!isStudioInitialized()) {\r\n await initStudio(contentsDir, config);\r\n }\r\n}\r\n","/**\n * @context CLI layer — filesystem adapter at src/cli/adapters/fs-adapter.ts\n * @does Implements IFsAdapter; abstracts all file read/write/list operations behind a single interface\n * @depends src/shared/types.ts, src/shared/constants.ts, src/shared/fs-adapter.interface.ts\n * @do Add new I/O operations here; all file access must go through this adapter\n * @dont Import UI components, run HTTP requests, or contain business logic\n */\n\nimport fs from \"node:fs/promises\";\nimport path from \"node:path\";\nimport type { Dirent } from \"node:fs\";\nimport type { FileInfo, DirectoryFileEntry } from \"../../shared/types.js\";\nimport type { IFsAdapter } from \"../../shared/fs-adapter.interface.js\";\nimport { SUPPORTED_EXTENSIONS } from \"../../shared/constants.js\";\n\nexport class FsAdapter implements IFsAdapter {\n private readonly basePath: string;\n\n constructor(basePath: string) {\n this.basePath = path.resolve(basePath);\n }\n\n private resolve(...segments: string[]): string {\n return path.resolve(this.basePath, ...segments);\n }\n\n async readFile(filePath: string): Promise<string> {\n return fs.readFile(this.resolve(filePath), \"utf-8\");\n }\n\n async writeFile(filePath: string, content: string): Promise<void> {\n const fullPath = this.resolve(filePath);\n await fs.mkdir(path.dirname(fullPath), { recursive: true });\n await fs.writeFile(fullPath, content, \"utf-8\");\n }\n\n async deleteFile(filePath: string): Promise<void> {\n await fs.unlink(this.resolve(filePath));\n }\n\n async exists(filePath: string): Promise<boolean> {\n try {\n await fs.access(this.resolve(filePath));\n return true;\n } catch {\n return false;\n }\n }\n\n async getStats(filePath: string): Promise<FileInfo> {\n const fullPath = this.resolve(filePath);\n const stats = await fs.stat(fullPath);\n return { path: filePath, size: stats.size, modifiedAt: stats.mtime };\n }\n\n async listFiles(dirPath: string, extensions?: readonly string[]): Promise<string[]> {\n const fullPath = this.resolve(dirPath);\n const filterExts = extensions ?? SUPPORTED_EXTENSIONS;\n\n let entries: Dirent[];\n try {\n entries = await fs.readdir(fullPath, { withFileTypes: true });\n } catch {\n return [];\n }\n\n return entries\n .filter((entry) => entry.isFile() && filterExts.some((ext) => entry.name.endsWith(ext)))\n .map((entry) => this.join(dirPath, entry.name));\n }\n\n async listDirectories(dirPath: string): Promise<string[]> {\n const fullPath = this.resolve(dirPath);\n\n let entries: Dirent[];\n try {\n entries = await fs.readdir(fullPath, { withFileTypes: true });\n } catch {\n return [];\n }\n\n return entries\n .filter((entry) => entry.isDirectory())\n .map((entry) => this.join(dirPath, entry.name));\n }\n\n async readBuffer(filePath: string): Promise<Buffer> {\n return fs.readFile(this.resolve(filePath));\n }\n\n async writeBuffer(filePath: string, data: Buffer): Promise<void> {\n const fullPath = this.resolve(filePath);\n await fs.mkdir(path.dirname(fullPath), { recursive: true });\n await fs.writeFile(fullPath, data);\n }\n\n async listAllFiles(dirPath: string): Promise<DirectoryFileEntry[]> {\n const fullPath = this.resolve(dirPath);\n\n let entries: Dirent[];\n try {\n entries = await fs.readdir(fullPath, { withFileTypes: true });\n } catch {\n return [];\n }\n\n const results: DirectoryFileEntry[] = [];\n for (const entry of entries) {\n if (!entry.isFile()) continue;\n const relativePath = this.join(dirPath, entry.name);\n const stats = await fs.stat(this.resolve(relativePath));\n results.push({ name: entry.name, relativePath, size: stats.size, modifiedAt: stats.mtime });\n }\n return results;\n }\n\n join(...segments: string[]): string {\n return path.join(...segments);\n }\n\n basename(filePath: string): string {\n return path.basename(filePath);\n }\n\n extname(filePath: string): string {\n return path.extname(filePath);\n }\n\n relative(from: string, to: string): string {\n return path.relative(from, to);\n }\n\n normalizeSlug(relativePath: string, ext: string): string {\n return relativePath.replace(ext, \"\").split(path.sep).join(\"/\");\n }\n}\n","/**\n * @context Shared layer — field label utilities at src/shared/field-utils.ts\n * @does Resolves human-readable labels for field definitions and raw key strings\n * @depends src/shared/fields.ts\n * @do Add field-related utility functions here\n * @dont Import from CLI or UI; contain field type definitions or schema logic\n */\n\nimport type { BaseField } from \"./fields.js\";\n\n/**\n * Resolve the human-readable label for a field.\n *\n * When the field definition has an explicit `label`, that is returned as-is.\n * Otherwise the `name` (camelCase / kebab-case / snake_case) is converted to Title Case:\n *\n * @example\n * fieldLabel({ name: \"siteName\", type: \"text\" }) // \"Site Name\"\n * fieldLabel({ name: \"created_at\", type: \"date\" }) // \"Created At\"\n * fieldLabel({ name: \"bio\", type: \"long-text\", label: \"About\" }) // \"About\"\n */\nexport function fieldLabel(field: Pick<BaseField, \"name\" | \"label\">): string {\n if (field.label) return field.label;\n return field.name\n .replace(/[-_](.)/g, (_, c: string) => ` ${c.toUpperCase()}`)\n .replace(/([A-Z])/g, \" $1\")\n .trim()\n .replace(/\\b\\w/g, (c) => c.toUpperCase());\n}\n\n/**\n * Resolve the label for a raw key string (no field definition available).\n * Useful for dynamic keys that have no schema entry.\n */\nexport function keyLabel(name: string): string {\n return fieldLabel({ name });\n}\n"],"mappings":";AAQA,SAAS,QAAQ,SAAS,KAAK,aAAa;;;ACA5C,OAAO,aAAa;;;ACGb,IAAM,uBAAuB,CAAC,QAAQ,OAAO;AAC7C,IAAM,wBAAwB;AAI9B,IAAM,mBAAmB;AAAA,EAC9B;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AACF;AAEO,IAAM,mBAAmB,CAAC,aAAa,cAAc,WAAW;AAEhE,IAAM,mBAAmB;AAAA,EAC9B;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AACF;AAEO,IAAM,mBAAmB,CAAC,GAAG,kBAAkB,GAAG,kBAAkB,GAAG,gBAAgB;;;AC5B9F,OAAO,YAAY;AAOZ,SAAS,SAAS,SAA4B;AACnD,QAAM,EAAE,MAAM,SAAS,KAAK,IAAI,OAAO,OAAO;AAC9C,SAAO,EAAE,MAAM,MAAM,KAAK,KAAK,EAAE;AACnC;;;ACEO,SAAS,UAAU,SAA6B;AACrD,QAAM,SAAkB,KAAK,MAAM,OAAO;AAE1C,MAAI,MAAM,QAAQ,MAAM,GAAG;AACzB,WAAO;AAAA,MACL,MAAM;AAAA,MACN,SAAS;AAAA,IACX;AAAA,EACF;AAEA,MAAI,OAAO,WAAW,YAAY,WAAW,MAAM;AACjD,WAAO;AAAA,MACL,MAAM;AAAA,MACN,MAAM;AAAA,IACR;AAAA,EACF;AAEA,QAAM,IAAI,MAAM,yCAAyC;AAC3D;;;AC1BA,IAAM,cAAc;AACpB,IAAM,kBACJ;AACF,IAAM,WAAW;AACjB,IAAM,SAAS;AACf,IAAM,sBAAsB;AAE5B,SAAS,UAAU,OAAwB;AACzC,SAAO,YAAY,KAAK,KAAK;AAC/B;AAEA,SAAS,cAAc,OAAwB;AAC7C,SAAO,gBAAgB,KAAK,KAAK;AACnC;AAEA,SAAS,QAAQ,OAAwB;AACvC,SAAO,SAAS,KAAK,KAAK;AAC5B;AAEA,SAAS,MAAM,OAAwB;AACrC,SAAO,OAAO,KAAK,KAAK;AAC1B;AAEA,SAAS,iBAAiB,MAAc,SAAoC;AAC1E,MAAI,QAAQ,MAAM,OAAO,EAAG,QAAO,EAAE,MAAM,MAAM,QAAQ;AACzD,MAAI,QAAQ,MAAM,KAAK,EAAG,QAAO,EAAE,MAAM,MAAM,MAAM;AACrD,MAAI,QAAQ,MAAM,aAAa,EAAG,QAAO,EAAE,MAAM,MAAM,QAAQ,aAAa,KAAK;AACjF,MAAI,QAAQ,MAAM,SAAS,EAAG,QAAO,EAAE,MAAM,MAAM,OAAO;AAE1D,QAAM,SAAS,QAAQ,KAAK,CAAC,MAAM,EAAE,SAAS,uBAAuB,EAAE,SAAS,IAAI,CAAC;AACrF,SAAO,EAAE,MAAM,MAAM,SAAS,cAAc,OAAO;AACrD;AAEA,SAAS,gBAAgB,MAAc,OAAmC;AACxE,MAAI,MAAM,WAAW,EAAG,QAAO,EAAE,MAAM,MAAM,SAAS,YAAY,CAAC,EAAE;AAErE,MAAI,MAAM,MAAM,CAAC,SAAS,OAAO,SAAS,QAAQ,GAAG;AACnD,UAAM,SAAS,CAAC,GAAG,IAAI,IAAI,KAAiB,CAAC,EAAE,MAAM,GAAG,EAAE;AAC1D,UAAM,UAA0B,OAAO,IAAI,CAAC,OAAO,EAAE,OAAO,GAAG,OAAO,EAAE,EAAE;AAC1E,WAAO,EAAE,MAAM,MAAM,gBAAgB,QAAQ;AAAA,EAC/C;AAEA,MAAI,MAAM,MAAM,CAAC,SAAS,OAAO,SAAS,YAAY,SAAS,QAAQ,CAAC,MAAM,QAAQ,IAAI,CAAC,GAAG;AAC5F,WAAO,EAAE,MAAM,MAAM,SAAS,YAAY,YAAY,KAAkC,EAAE;AAAA,EAC5F;AAEA,SAAO,EAAE,MAAM,MAAM,SAAS,YAAY,CAAC,EAAE;AAC/C;AAEA,SAAS,qBAAqB,MAAc,QAAoC;AAC9E,QAAM,UAAU,OAAO,OAAO,CAAC,MAAM,MAAM,QAAQ,MAAM,MAAS;AAElE,MAAI,QAAQ,WAAW,EAAG,QAAO,EAAE,MAAM,MAAM,OAAO;AACtD,MAAI,QAAQ,MAAM,CAAC,MAAM,OAAO,MAAM,SAAS,EAAG,QAAO,EAAE,MAAM,MAAM,UAAU;AAEjF,MAAI,QAAQ,MAAM,CAAC,MAAM,OAAO,MAAM,QAAQ,GAAG;AAC/C,UAAM,SAAS,QAAQ,MAAM,CAAC,MAAM,OAAO,UAAU,CAAC,CAAC,IAAI,YAAY;AACvE,WAAO,EAAE,MAAM,MAAM,UAAU,OAAO;AAAA,EACxC;AAEA,MAAI,QAAQ,MAAM,CAAC,MAAM,OAAO,MAAM,QAAQ,GAAG;AAC/C,WAAO,iBAAiB,MAAM,OAAmB;AAAA,EACnD;AAEA,MAAI,QAAQ,MAAM,CAAC,MAAM,MAAM,QAAQ,CAAC,CAAC,GAAG;AAC1C,WAAO,gBAAgB,MAAO,QAAwB,KAAK,CAAC;AAAA,EAC9D;AAEA,MAAI,QAAQ,MAAM,CAAC,MAAM,OAAO,MAAM,YAAY,MAAM,QAAQ,CAAC,MAAM,QAAQ,CAAC,CAAC,GAAG;AAClF,WAAO,EAAE,MAAM,MAAM,UAAU,QAAQ,YAAY,OAAoC,EAAE;AAAA,EAC3F;AAEA,SAAO,EAAE,MAAM,MAAM,OAAO;AAC9B;AAEA,SAAS,YAAY,MAAoD;AACvE,QAAM,SAAS,IAAI,IAAY,KAAK,QAAQ,CAAC,QAAQ,OAAO,KAAK,GAAG,CAAC,CAAC;AACtE,SAAO,MAAM,KAAK,MAAM,EAAE,IAAI,CAAC,QAAQ,qBAAqB,KAAK,KAAK,IAAI,CAAC,QAAQ,IAAI,GAAG,CAAC,CAAC,CAAC;AAC/F;AASO,SAAS,YAAY,SAAyB,gBAA0C;AAC7F,QAAM,OAAO,QAAQ,IAAI,CAAC,UAAU,MAAM,IAA+B;AACzE,SAAO,EAAE,YAAY,gBAAgB,QAAQ,YAAY,IAAI,EAAE;AACjE;;;AJrFO,IAAM,eAAN,MAAmB;AAAA,EACP,UAAU,oBAAI,IAA4B;AAAA,EAC1C,cAAc,oBAAI,IAAwB;AAAA,EAC1C;AAAA,EAEjB,YAAY,WAAuB;AACjC,SAAK,KAAK;AAAA,EACZ;AAAA,EAEA,MAAM,MAAM,QAAsC;AAChD,SAAK,MAAM;AACX,UAAM,OAAO,MAAM,KAAK,GAAG,gBAAgB,GAAG;AAE9C,eAAW,OAAO,MAAM;AACtB,YAAM,UAAU,KAAK,GAAG,SAAS,GAAG;AACpC,YAAM,iBAAiB,QAAQ,OAAO;AACtC,YAAM,mBAAmB,QAAQ,cAAc,cAAc;AAC7D,YAAM,KAAK,gBAAgB,SAAS,gBAAgB,kBAAkB,MAAM;AAAA,IAC9E;AAAA,EACF;AAAA,EAEA,cAAc,MAA8B;AAC1C,WAAO,KAAK,QAAQ,IAAI,IAAI,KAAK,CAAC;AAAA,EACpC;AAAA,EAEA,iBAA+B;AAC7B,WAAO,MAAM,KAAK,KAAK,YAAY,OAAO,CAAC;AAAA,EAC7C;AAAA,EAEA,QAAc;AACZ,SAAK,QAAQ,MAAM;AACnB,SAAK,YAAY,MAAM;AAAA,EACzB;AAAA,EAEA,MAAc,gBACZ,SACA,gBACA,cACe;AACf,UAAM,UAA0B,CAAC;AACjC,UAAM,KAAK,QAAQ,SAAS,gBAAgB,SAAS,OAAO;AAE5D,UAAM,YAAY,KAAK,GAAG,KAAK,SAAS,qBAAqB;AAC7D,UAAM,WAAW,MAAM,KAAK,aAAa,SAAS;AAClD,QAAI,UAAU;AACZ,WAAK,cAAc,SAAS,QAAQ;AAAA,IACtC;AAEA,UAAM,SAAS,gBAAgB,YAAY,SAAS,cAAc;AAElE,SAAK,QAAQ,IAAI,gBAAgB,OAAO;AACxC,SAAK,YAAY,IAAI,gBAAgB;AAAA,MACnC,MAAM;AAAA,MACN,MAAM,KAAK,qBAAqB,OAAO;AAAA,MACvC,OAAO,QAAQ;AAAA,MACf,UAAU;AAAA,MACV;AAAA,IACF,CAAC;AAAA,EACH;AAAA,EAEA,MAAc,QACZ,SACA,gBACA,SACA,SACe;AACf,UAAM,UAAU,MAAM,KAAK,GAAG,gBAAgB,OAAO;AACrD,eAAW,UAAU,SAAS;AAC5B,YAAM,KAAK,QAAQ,SAAS,gBAAgB,QAAQ,OAAO;AAAA,IAC7D;AAEA,UAAM,QAAQ,MAAM,KAAK,GAAG,UAAU,OAAO;AAC7C,eAAW,YAAY,OAAO;AAC5B,YAAM,WAAW,KAAK,GAAG,SAAS,QAAQ;AAC1C,UAAI,aAAa,sBAAuB;AAExC,YAAM,MAAM,KAAK,GAAG,QAAQ,QAAQ;AACpC,YAAM,UAAU,MAAM,KAAK,GAAG,SAAS,QAAQ;AAC/C,YAAM,eAAe,KAAK,GAAG,SAAS,SAAS,QAAQ;AACvD,YAAM,OAAO,KAAK,GACf,cAAc,cAAc,GAAG,EAC/B,MAAM,GAAG,EACT,IAAI,CAAC,YAAY,QAAQ,OAAO,CAAC,EACjC,KAAK,GAAG;AAEX,UAAI,QAAQ,QAAQ;AAClB,gBAAQ,KAAK,KAAK,cAAc,gBAAgB,MAAM,OAAO,CAAC;AAAA,MAChE,WAAW,QAAQ,SAAS;AAC1B,gBAAQ,KAAK,GAAG,KAAK,iBAAiB,gBAAgB,MAAM,OAAO,CAAC;AAAA,MACtE;AAAA,IACF;AAAA,EACF;AAAA,EAEQ,cAAc,gBAAwB,MAAc,SAA+B;AACzF,UAAM,SAAS,SAAS,OAAO;AAC/B,WAAO;AAAA,MACL,YAAY;AAAA,MACZ;AAAA,MACA,MAAM,IAAI,cAAc,IAAI,IAAI;AAAA,MAChC,MAAM,OAAO;AAAA,MACb,MAAM,OAAO;AAAA,IACf;AAAA,EACF;AAAA,EAEQ,iBAAiB,gBAAwB,MAAc,SAAiC;AAC9F,UAAM,SAAS,UAAU,OAAO;AAEhC,QAAI,OAAO,SAAS,cAAc;AAChC,aAAO,OAAO,QAAQ,IAAI,CAAC,MAAM,UAAU;AACzC,cAAM,YACJ,OAAO,KAAK,MAAM,MAAM,WAAW,QAAQ,KAAK,MAAM,CAAC,IAAI,GAAG,IAAI,IAAI,KAAK;AAC7E,eAAO;AAAA,UACL,YAAY;AAAA,UACZ,MAAM;AAAA,UACN,MAAM,IAAI,cAAc,IAAI,SAAS;AAAA,UACrC;AAAA,QACF;AAAA,MACF,CAAC;AAAA,IACH;AAEA,WAAO,CAAC,EAAE,YAAY,gBAAgB,MAAM,MAAM,IAAI,cAAc,IAAI,IAAI,IAAI,MAAM,OAAO,KAAK,CAAC;AAAA,EACrG;AAAA,EAEA,MAAc,aAAa,WAA6C;AACtE,QAAI,CAAE,MAAM,KAAK,GAAG,OAAO,SAAS,EAAI,QAAO;AAE/C,QAAI;AACF,YAAM,UAAU,MAAM,KAAK,GAAG,SAAS,SAAS;AAChD,YAAM,SAAkB,KAAK,MAAM,OAAO;AAC1C,UAAI,MAAM,QAAQ,MAAM,EAAG,QAAO;AAAA,IACpC,SAAS,OAAO;AACd,cAAQ,KAAK,kDAAkD,SAAS,IAAI,KAAK;AAAA,IACnF;AACA,WAAO;AAAA,EACT;AAAA,EAEQ,cAAc,SAAyB,UAA0B;AACvE,UAAM,WAAW,IAAI,IAAI,SAAS,IAAI,CAAC,MAAM,UAAU,CAAC,MAAM,KAAK,CAAC,CAAC;AACrE,YAAQ,KAAK,CAAC,GAAG,MAAM;AACrB,YAAM,SAAS,SAAS,IAAI,EAAE,IAAI,KAAK;AACvC,YAAM,SAAS,SAAS,IAAI,EAAE,IAAI,KAAK;AACvC,aAAO,SAAS;AAAA,IAClB,CAAC;AAAA,EACH;AAAA,EAEQ,qBAAqB,SAA6C;AACxE,QAAI,QAAQ,WAAW,EAAG,QAAO;AACjC,UAAM,QAAQ,QAAQ,CAAC;AACvB,QAAI,MAAM,SAAS,OAAW,QAAO;AACrC,QAAI,QAAQ,WAAW,KAAK,CAAC,MAAM,KAAK,SAAS,GAAG,EAAG,QAAO;AAC9D,WAAO;AAAA,EACT;AACF;;;AK7JA,IAAI,QAA6B;AAE1B,SAAS,WAAyB;AACvC,MAAI,CAAC,OAAO;AACV,UAAM,IAAI,MAAM,yDAAyD;AAAA,EAC3E;AACA,SAAO;AACT;AAEA,eAAsB,YACpB,WACA,QACuB;AACvB,QAAM,QAAQ,IAAI,aAAa,SAAS;AACxC,QAAM,MAAM,MAAM,MAAM;AACxB,UAAQ;AACR,SAAO;AACT;;;ANAO,IAAM,eAAN,MAAgD;AAAA,EACpC;AAAA,EACT,UAAwB,CAAC;AAAA,EAEjC,YAAY,YAAoB;AAC9B,SAAK,iBAAiB;AAAA,EACxB;AAAA,EAEA,MAAM,YAA2C;AAC/C,SAAK,QAAQ,QAAQ,EAAE,GAAG,KAAK,QAAQ,OAAO,GAAG,WAAW;AAC5D,WAAO;AAAA,EACT;AAAA,EAEA,KAAK,OAAe,QAAwB,OAAa;AACvD,SAAK,QAAQ,OAAO,EAAE,OAAO,MAAM;AACnC,WAAO;AAAA,EACT;AAAA,EAEA,MAAM,OAAqB;AACzB,SAAK,QAAQ,QAAQ;AACrB,WAAO;AAAA,EACT;AAAA,EAEA,OAAO,OAAqB;AAC1B,SAAK,QAAQ,SAAS;AACtB,WAAO;AAAA,EACT;AAAA,EAEA,MAAW;AACT,QAAI,UAAU,CAAC,GAAG,SAAS,EAAE,cAAc,KAAK,cAAc,CAAC;AAE/D,QAAI,KAAK,QAAQ,OAAO;AACtB,YAAM,aAAa,KAAK,QAAQ;AAChC,gBAAU;AAAA,QAAO;AAAA,QAAS,CAAC,UACzB,OAAO,QAAQ,UAAU,EAAE,MAAM,CAAC,CAAC,KAAK,KAAK,MAAM,IAAI,MAAM,MAAM,GAAG,MAAM,KAAK;AAAA,MACnF;AAAA,IACF;AAEA,QAAI,KAAK,QAAQ,MAAM;AACrB,YAAM,EAAE,OAAO,MAAM,IAAI,KAAK,QAAQ;AACtC,gBAAU,QAAQ,SAAS,CAAC,CAAC,UAAU,IAAI,MAAM,MAAM,KAAK,CAAC,GAAG,CAAC,KAAK,CAAC;AAAA,IACzE;AAEA,UAAM,QAAQ,KAAK,QAAQ,UAAU;AACrC,UAAM,MAAM,KAAK,QAAQ,QAAQ,QAAQ,KAAK,QAAQ,QAAQ;AAC9D,WAAO,MAAM,SAAS,OAAO,GAAG,EAAE,IAAI,CAAC,MAAM,EAAE,IAAoB;AAAA,EACrE;AAAA,EAEA,QAAuB;AACrB,WAAO,KAAK,MAAM,CAAC,EAAE,IAAI,EAAE,CAAC;AAAA,EAC9B;AAAA,EAEA,QAAgB;AACd,WAAO,KAAK,IAAI,EAAE;AAAA,EACpB;AACF;AASO,SAAS,gBAAgB,MAA4B;AAC1D,SAAO,IAAI,aAAa,IAAI;AAC9B;;;AOvFA,OAAOA,WAAU;;;ACAjB,OAAO,QAAQ;AACf,OAAO,UAAU;AAMV,IAAM,YAAN,MAAsC;AAAA,EAC1B;AAAA,EAEjB,YAAY,UAAkB;AAC5B,SAAK,WAAW,KAAK,QAAQ,QAAQ;AAAA,EACvC;AAAA,EAEQ,WAAW,UAA4B;AAC7C,WAAO,KAAK,QAAQ,KAAK,UAAU,GAAG,QAAQ;AAAA,EAChD;AAAA,EAEA,MAAM,SAAS,UAAmC;AAChD,WAAO,GAAG,SAAS,KAAK,QAAQ,QAAQ,GAAG,OAAO;AAAA,EACpD;AAAA,EAEA,MAAM,UAAU,UAAkB,SAAgC;AAChE,UAAM,WAAW,KAAK,QAAQ,QAAQ;AACtC,UAAM,GAAG,MAAM,KAAK,QAAQ,QAAQ,GAAG,EAAE,WAAW,KAAK,CAAC;AAC1D,UAAM,GAAG,UAAU,UAAU,SAAS,OAAO;AAAA,EAC/C;AAAA,EAEA,MAAM,WAAW,UAAiC;AAChD,UAAM,GAAG,OAAO,KAAK,QAAQ,QAAQ,CAAC;AAAA,EACxC;AAAA,EAEA,MAAM,OAAO,UAAoC;AAC/C,QAAI;AACF,YAAM,GAAG,OAAO,KAAK,QAAQ,QAAQ,CAAC;AACtC,aAAO;AAAA,IACT,QAAQ;AACN,aAAO;AAAA,IACT;AAAA,EACF;AAAA,EAEA,MAAM,SAAS,UAAqC;AAClD,UAAM,WAAW,KAAK,QAAQ,QAAQ;AACtC,UAAM,QAAQ,MAAM,GAAG,KAAK,QAAQ;AACpC,WAAO,EAAE,MAAM,UAAU,MAAM,MAAM,MAAM,YAAY,MAAM,MAAM;AAAA,EACrE;AAAA,EAEA,MAAM,UAAU,SAAiB,YAAmD;AAClF,UAAM,WAAW,KAAK,QAAQ,OAAO;AACrC,UAAM,aAAa,cAAc;AAEjC,QAAI;AACJ,QAAI;AACF,gBAAU,MAAM,GAAG,QAAQ,UAAU,EAAE,eAAe,KAAK,CAAC;AAAA,IAC9D,QAAQ;AACN,aAAO,CAAC;AAAA,IACV;AAEA,WAAO,QACJ,OAAO,CAAC,UAAU,MAAM,OAAO,KAAK,WAAW,KAAK,CAAC,QAAQ,MAAM,KAAK,SAAS,GAAG,CAAC,CAAC,EACtF,IAAI,CAAC,UAAU,KAAK,KAAK,SAAS,MAAM,IAAI,CAAC;AAAA,EAClD;AAAA,EAEA,MAAM,gBAAgB,SAAoC;AACxD,UAAM,WAAW,KAAK,QAAQ,OAAO;AAErC,QAAI;AACJ,QAAI;AACF,gBAAU,MAAM,GAAG,QAAQ,UAAU,EAAE,eAAe,KAAK,CAAC;AAAA,IAC9D,QAAQ;AACN,aAAO,CAAC;AAAA,IACV;AAEA,WAAO,QACJ,OAAO,CAAC,UAAU,MAAM,YAAY,CAAC,EACrC,IAAI,CAAC,UAAU,KAAK,KAAK,SAAS,MAAM,IAAI,CAAC;AAAA,EAClD;AAAA,EAEA,MAAM,WAAW,UAAmC;AAClD,WAAO,GAAG,SAAS,KAAK,QAAQ,QAAQ,CAAC;AAAA,EAC3C;AAAA,EAEA,MAAM,YAAY,UAAkB,MAA6B;AAC/D,UAAM,WAAW,KAAK,QAAQ,QAAQ;AACtC,UAAM,GAAG,MAAM,KAAK,QAAQ,QAAQ,GAAG,EAAE,WAAW,KAAK,CAAC;AAC1D,UAAM,GAAG,UAAU,UAAU,IAAI;AAAA,EACnC;AAAA,EAEA,MAAM,aAAa,SAAgD;AACjE,UAAM,WAAW,KAAK,QAAQ,OAAO;AAErC,QAAI;AACJ,QAAI;AACF,gBAAU,MAAM,GAAG,QAAQ,UAAU,EAAE,eAAe,KAAK,CAAC;AAAA,IAC9D,QAAQ;AACN,aAAO,CAAC;AAAA,IACV;AAEA,UAAM,UAAgC,CAAC;AACvC,eAAW,SAAS,SAAS;AAC3B,UAAI,CAAC,MAAM,OAAO,EAAG;AACrB,YAAM,eAAe,KAAK,KAAK,SAAS,MAAM,IAAI;AAClD,YAAM,QAAQ,MAAM,GAAG,KAAK,KAAK,QAAQ,YAAY,CAAC;AACtD,cAAQ,KAAK,EAAE,MAAM,MAAM,MAAM,cAAc,MAAM,MAAM,MAAM,YAAY,MAAM,MAAM,CAAC;AAAA,IAC5F;AACA,WAAO;AAAA,EACT;AAAA,EAEA,QAAQ,UAA4B;AAClC,WAAO,KAAK,KAAK,GAAG,QAAQ;AAAA,EAC9B;AAAA,EAEA,SAAS,UAA0B;AACjC,WAAO,KAAK,SAAS,QAAQ;AAAA,EAC/B;AAAA,EAEA,QAAQ,UAA0B;AAChC,WAAO,KAAK,QAAQ,QAAQ;AAAA,EAC9B;AAAA,EAEA,SAAS,MAAc,IAAoB;AACzC,WAAO,KAAK,SAAS,MAAM,EAAE;AAAA,EAC/B;AAAA,EAEA,cAAc,cAAsB,KAAqB;AACvD,WAAO,aAAa,QAAQ,KAAK,EAAE,EAAE,MAAM,KAAK,GAAG,EAAE,KAAK,GAAG;AAAA,EAC/D;AACF;;;AD3GA,eAAsB,WACpB,aACA,QACe;AACf,QAAM,MAAM,eAAeC,MAAK,KAAK,QAAQ,IAAI,GAAG,UAAU;AAC9D,QAAM,YAAY,IAAI,UAAU,GAAG,GAAG,MAAM;AAC9C;AAKO,SAAS,sBAA+B;AAC7C,MAAI;AACF,aAAS;AACT,WAAO;AAAA,EACT,QAAQ;AACN,WAAO;AAAA,EACT;AACF;AAqBA,eAAsB,oBACpB,aACA,QACe;AACf,MAAI,CAAC,oBAAoB,GAAG;AAC1B,UAAM,WAAW,aAAa,MAAM;AAAA,EACtC;AACF;;;AErDO,SAAS,WAAW,OAAkD;AAC3E,MAAI,MAAM,MAAO,QAAO,MAAM;AAC9B,SAAO,MAAM,KACV,QAAQ,YAAY,CAAC,GAAG,MAAc,IAAI,EAAE,YAAY,CAAC,EAAE,EAC3D,QAAQ,YAAY,KAAK,EACzB,KAAK,EACL,QAAQ,SAAS,CAAC,MAAM,EAAE,YAAY,CAAC;AAC5C;AAMO,SAAS,SAAS,MAAsB;AAC7C,SAAO,WAAW,EAAE,KAAK,CAAC;AAC5B;","names":["path","path"]}
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "nextjs-studio",
3
- "version": "0.2.0",
3
+ "version": "0.4.0",
4
4
  "description": "A Git-based, local-first CMS for Next.js projects",
5
5
  "keywords": [
6
6
  "nextjs",
@@ -14,7 +14,7 @@
14
14
  "homepage": "https://github.com/TiagoDanin/Nextjs-Studio",
15
15
  "repository": {
16
16
  "type": "git",
17
- "url": "https://github.com/TiagoDanin/Nextjs-Studio.git"
17
+ "url": "git+https://github.com/TiagoDanin/Nextjs-Studio.git"
18
18
  },
19
19
  "license": "MIT",
20
20
  "author": "Tiago Danin",
@@ -27,7 +27,9 @@
27
27
  },
28
28
  "main": "./dist/core/index.js",
29
29
  "types": "./dist/core/index.d.ts",
30
- "bin": "./dist/bin/nextjs-studio.js",
30
+ "bin": {
31
+ "nextjs-studio": "dist/bin/nextjs-studio.js"
32
+ },
31
33
  "files": [
32
34
  "dist",
33
35
  "README.md",