@f0rbit/corpus 0.1.2 → 0.1.4

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (43) hide show
  1. package/dist/backend/cloudflare.d.ts +40 -0
  2. package/dist/backend/cloudflare.d.ts.map +1 -1
  3. package/dist/backend/cloudflare.js +40 -0
  4. package/dist/backend/file.d.ts +35 -0
  5. package/dist/backend/file.d.ts.map +1 -1
  6. package/dist/backend/file.js +35 -0
  7. package/dist/backend/layered.d.ts +46 -0
  8. package/dist/backend/layered.d.ts.map +1 -0
  9. package/dist/backend/layered.js +193 -0
  10. package/dist/backend/memory.d.ts +30 -0
  11. package/dist/backend/memory.d.ts.map +1 -1
  12. package/dist/backend/memory.js +30 -0
  13. package/dist/backends.d.ts +11 -0
  14. package/dist/backends.d.ts.map +1 -0
  15. package/dist/backends.js +9 -0
  16. package/dist/cloudflare.d.ts +2 -5
  17. package/dist/cloudflare.d.ts.map +1 -1
  18. package/dist/cloudflare.js +2 -5
  19. package/dist/codecs.d.ts +8 -0
  20. package/dist/codecs.d.ts.map +1 -0
  21. package/dist/codecs.js +6 -0
  22. package/dist/core.d.ts +9 -0
  23. package/dist/core.d.ts.map +1 -0
  24. package/dist/core.js +7 -0
  25. package/dist/corpus.d.ts +68 -1
  26. package/dist/corpus.d.ts.map +1 -1
  27. package/dist/corpus.js +194 -1
  28. package/dist/index.d.ts +4 -6
  29. package/dist/index.d.ts.map +1 -1
  30. package/dist/index.js +3 -5
  31. package/dist/schema.d.ts +27 -0
  32. package/dist/schema.d.ts.map +1 -1
  33. package/dist/schema.js +27 -0
  34. package/dist/sst.d.ts +38 -0
  35. package/dist/sst.d.ts.map +1 -1
  36. package/dist/sst.js +38 -0
  37. package/dist/types.d.ts +229 -1
  38. package/dist/types.d.ts.map +1 -1
  39. package/dist/types.js +91 -2
  40. package/dist/utils.d.ts +133 -0
  41. package/dist/utils.d.ts.map +1 -0
  42. package/dist/utils.js +174 -0
  43. package/package.json +5 -2
@@ -1,3 +1,7 @@
1
+ /**
2
+ * @module Backends
3
+ * @description Cloudflare Workers storage backend using D1 and R2.
4
+ */
1
5
  import type { Backend, EventHandler } from '../types';
2
6
  type D1Database = {
3
7
  prepare: (sql: string) => unknown;
@@ -18,6 +22,42 @@ export type CloudflareBackendConfig = {
18
22
  r2: R2Bucket;
19
23
  on_event?: EventHandler;
20
24
  };
25
+ /**
26
+ * Creates a Cloudflare Workers storage backend using D1 and R2.
27
+ * @category Backends
28
+ * @group Storage Backends
29
+ *
30
+ * Uses D1 (SQLite) for metadata storage and R2 (object storage) for binary data.
31
+ * Requires running `CORPUS_MIGRATION_SQL` on the D1 database before first use.
32
+ *
33
+ * This backend is designed for production use in Cloudflare Workers environments,
34
+ * providing durable, globally distributed storage.
35
+ *
36
+ * @param config - Configuration with `d1` (D1 database), `r2` (R2 bucket), and optional `on_event` handler
37
+ * @returns A Backend instance using Cloudflare D1 + R2
38
+ *
39
+ * @example
40
+ * ```ts
41
+ * // In a Cloudflare Worker
42
+ * export default {
43
+ * async fetch(request: Request, env: Env) {
44
+ * const backend = create_cloudflare_backend({
45
+ * d1: env.CORPUS_DB,
46
+ * r2: env.CORPUS_BUCKET
47
+ * })
48
+ *
49
+ * const corpus = create_corpus()
50
+ * .with_backend(backend)
51
+ * .with_store(define_store('cache', json_codec(CacheSchema)))
52
+ * .build()
53
+ *
54
+ * // Use corpus...
55
+ * }
56
+ * }
57
+ * ```
58
+ *
59
+ * @see CORPUS_MIGRATION_SQL for required database setup
60
+ */
21
61
  export declare function create_cloudflare_backend(config: CloudflareBackendConfig): Backend;
22
62
  export {};
23
63
  //# sourceMappingURL=cloudflare.d.ts.map
@@ -1 +1 @@
1
- {"version":3,"file":"cloudflare.d.ts","sourceRoot":"","sources":["../../backend/cloudflare.ts"],"names":[],"mappings":"AAEA,OAAO,KAAK,EAAE,OAAO,EAAwF,YAAY,EAAE,MAAM,UAAU,CAAA;AAI3I,KAAK,UAAU,GAAG;IAAE,OAAO,EAAE,CAAC,GAAG,EAAE,MAAM,KAAK,OAAO,CAAA;CAAE,CAAA;AACvD,KAAK,QAAQ,GAAG;IACd,GAAG,EAAE,CAAC,GAAG,EAAE,MAAM,KAAK,OAAO,CAAC;QAAE,IAAI,EAAE,cAAc,CAAC,UAAU,CAAC,CAAC;QAAC,WAAW,EAAE,MAAM,OAAO,CAAC,WAAW,CAAC,CAAA;KAAE,GAAG,IAAI,CAAC,CAAA;IACnH,GAAG,EAAE,CAAC,GAAG,EAAE,MAAM,EAAE,IAAI,EAAE,cAAc,CAAC,UAAU,CAAC,GAAG,UAAU,KAAK,OAAO,CAAC,IAAI,CAAC,CAAA;IAClF,MAAM,EAAE,CAAC,GAAG,EAAE,MAAM,KAAK,OAAO,CAAC,IAAI,CAAC,CAAA;IACtC,IAAI,EAAE,CAAC,GAAG,EAAE,MAAM,KAAK,OAAO,CAAC;QAAE,GAAG,EAAE,MAAM,CAAA;KAAE,GAAG,IAAI,CAAC,CAAA;CACvD,CAAA;AAED,MAAM,MAAM,uBAAuB,GAAG;IACpC,EAAE,EAAE,UAAU,CAAA;IACd,EAAE,EAAE,QAAQ,CAAA;IACZ,QAAQ,CAAC,EAAE,YAAY,CAAA;CACxB,CAAA;AAED,wBAAgB,yBAAyB,CAAC,MAAM,EAAE,uBAAuB,GAAG,OAAO,CA8PlF"}
1
+ {"version":3,"file":"cloudflare.d.ts","sourceRoot":"","sources":["../../backend/cloudflare.ts"],"names":[],"mappings":"AAAA;;;GAGG;AAIH,OAAO,KAAK,EAAE,OAAO,EAAwF,YAAY,EAAE,MAAM,UAAU,CAAA;AAI3I,KAAK,UAAU,GAAG;IAAE,OAAO,EAAE,CAAC,GAAG,EAAE,MAAM,KAAK,OAAO,CAAA;CAAE,CAAA;AACvD,KAAK,QAAQ,GAAG;IACd,GAAG,EAAE,CAAC,GAAG,EAAE,MAAM,KAAK,OAAO,CAAC;QAAE,IAAI,EAAE,cAAc,CAAC,UAAU,CAAC,CAAC;QAAC,WAAW,EAAE,MAAM,OAAO,CAAC,WAAW,CAAC,CAAA;KAAE,GAAG,IAAI,CAAC,CAAA;IACnH,GAAG,EAAE,CAAC,GAAG,EAAE,MAAM,EAAE,IAAI,EAAE,cAAc,CAAC,UAAU,CAAC,GAAG,UAAU,KAAK,OAAO,CAAC,IAAI,CAAC,CAAA;IAClF,MAAM,EAAE,CAAC,GAAG,EAAE,MAAM,KAAK,OAAO,CAAC,IAAI,CAAC,CAAA;IACtC,IAAI,EAAE,CAAC,GAAG,EAAE,MAAM,KAAK,OAAO,CAAC;QAAE,GAAG,EAAE,MAAM,CAAA;KAAE,GAAG,IAAI,CAAC,CAAA;CACvD,CAAA;AAED,MAAM,MAAM,uBAAuB,GAAG;IACpC,EAAE,EAAE,UAAU,CAAA;IACd,EAAE,EAAE,QAAQ,CAAA;IACZ,QAAQ,CAAC,EAAE,YAAY,CAAA;CACxB,CAAA;AAED;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;GAmCG;AACH,wBAAgB,yBAAyB,CAAC,MAAM,EAAE,uBAAuB,GAAG,OAAO,CA8PlF"}
@@ -1,7 +1,47 @@
1
+ /**
2
+ * @module Backends
3
+ * @description Cloudflare Workers storage backend using D1 and R2.
4
+ */
1
5
  import { eq, and, desc, lt, gt, like, sql } from 'drizzle-orm';
2
6
  import { drizzle } from 'drizzle-orm/d1';
3
7
  import { ok, err } from '../types';
4
8
  import { corpus_snapshots } from '../schema';
9
+ /**
10
+ * Creates a Cloudflare Workers storage backend using D1 and R2.
11
+ * @category Backends
12
+ * @group Storage Backends
13
+ *
14
+ * Uses D1 (SQLite) for metadata storage and R2 (object storage) for binary data.
15
+ * Requires running `CORPUS_MIGRATION_SQL` on the D1 database before first use.
16
+ *
17
+ * This backend is designed for production use in Cloudflare Workers environments,
18
+ * providing durable, globally distributed storage.
19
+ *
20
+ * @param config - Configuration with `d1` (D1 database), `r2` (R2 bucket), and optional `on_event` handler
21
+ * @returns A Backend instance using Cloudflare D1 + R2
22
+ *
23
+ * @example
24
+ * ```ts
25
+ * // In a Cloudflare Worker
26
+ * export default {
27
+ * async fetch(request: Request, env: Env) {
28
+ * const backend = create_cloudflare_backend({
29
+ * d1: env.CORPUS_DB,
30
+ * r2: env.CORPUS_BUCKET
31
+ * })
32
+ *
33
+ * const corpus = create_corpus()
34
+ * .with_backend(backend)
35
+ * .with_store(define_store('cache', json_codec(CacheSchema)))
36
+ * .build()
37
+ *
38
+ * // Use corpus...
39
+ * }
40
+ * }
41
+ * ```
42
+ *
43
+ * @see CORPUS_MIGRATION_SQL for required database setup
44
+ */
5
45
  export function create_cloudflare_backend(config) {
6
46
  const db = drizzle(config.d1);
7
47
  const { r2, on_event } = config;
@@ -1,7 +1,42 @@
1
+ /**
2
+ * @module Backends
3
+ * @description File-system storage backend for local persistence.
4
+ */
1
5
  import type { Backend, EventHandler } from '../types';
2
6
  export type FileBackendConfig = {
3
7
  base_path: string;
4
8
  on_event?: EventHandler;
5
9
  };
10
+ /**
11
+ * Creates a file-system storage backend for local persistence.
12
+ * @category Backends
13
+ * @group Storage Backends
14
+ *
15
+ * Uses Bun's file APIs for efficient I/O. Metadata is stored as JSON files
16
+ * per store, and data is stored as binary files in a shared `_data` directory.
17
+ *
18
+ * Directory structure:
19
+ * ```
20
+ * base_path/
21
+ * <store_id>/_meta.json # Metadata for each store
22
+ * _data/<store_id>_<hash>.bin # Binary data files
23
+ * ```
24
+ *
25
+ * @param config - Configuration with `base_path` (root directory) and optional `on_event` handler
26
+ * @returns A Backend instance using file-system storage
27
+ *
28
+ * @example
29
+ * ```ts
30
+ * const backend = create_file_backend({
31
+ * base_path: './data/corpus',
32
+ * on_event: (e) => console.log(e.type)
33
+ * })
34
+ *
35
+ * const corpus = create_corpus()
36
+ * .with_backend(backend)
37
+ * .with_store(define_store('documents', json_codec(DocSchema)))
38
+ * .build()
39
+ * ```
40
+ */
6
41
  export declare function create_file_backend(config: FileBackendConfig): Backend;
7
42
  //# sourceMappingURL=file.d.ts.map
@@ -1 +1 @@
1
- {"version":3,"file":"file.d.ts","sourceRoot":"","sources":["../../backend/file.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,OAAO,EAAwF,YAAY,EAAE,MAAM,UAAU,CAAA;AAK3I,MAAM,MAAM,iBAAiB,GAAG;IAC9B,SAAS,EAAE,MAAM,CAAA;IACjB,QAAQ,CAAC,EAAE,YAAY,CAAA;CACxB,CAAA;AAED,wBAAgB,mBAAmB,CAAC,MAAM,EAAE,iBAAiB,GAAG,OAAO,CAqMtE"}
1
+ {"version":3,"file":"file.d.ts","sourceRoot":"","sources":["../../backend/file.ts"],"names":[],"mappings":"AAAA;;;GAGG;AAEH,OAAO,KAAK,EAAE,OAAO,EAAwF,YAAY,EAAE,MAAM,UAAU,CAAA;AAK3I,MAAM,MAAM,iBAAiB,GAAG;IAC9B,SAAS,EAAE,MAAM,CAAA;IACjB,QAAQ,CAAC,EAAE,YAAY,CAAA;CACxB,CAAA;AAED;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;GA8BG;AACH,wBAAgB,mBAAmB,CAAC,MAAM,EAAE,iBAAiB,GAAG,OAAO,CAqMtE"}
@@ -1,6 +1,41 @@
1
+ /**
2
+ * @module Backends
3
+ * @description File-system storage backend for local persistence.
4
+ */
1
5
  import { ok, err } from '../types';
2
6
  import { mkdir, readdir } from 'node:fs/promises';
3
7
  import { join, dirname } from 'node:path';
8
+ /**
9
+ * Creates a file-system storage backend for local persistence.
10
+ * @category Backends
11
+ * @group Storage Backends
12
+ *
13
+ * Uses Bun's file APIs for efficient I/O. Metadata is stored as JSON files
14
+ * per store, and data is stored as binary files in a shared `_data` directory.
15
+ *
16
+ * Directory structure:
17
+ * ```
18
+ * base_path/
19
+ * <store_id>/_meta.json # Metadata for each store
20
+ * _data/<store_id>_<hash>.bin # Binary data files
21
+ * ```
22
+ *
23
+ * @param config - Configuration with `base_path` (root directory) and optional `on_event` handler
24
+ * @returns A Backend instance using file-system storage
25
+ *
26
+ * @example
27
+ * ```ts
28
+ * const backend = create_file_backend({
29
+ * base_path: './data/corpus',
30
+ * on_event: (e) => console.log(e.type)
31
+ * })
32
+ *
33
+ * const corpus = create_corpus()
34
+ * .with_backend(backend)
35
+ * .with_store(define_store('documents', json_codec(DocSchema)))
36
+ * .build()
37
+ * ```
38
+ */
4
39
  export function create_file_backend(config) {
5
40
  const { base_path, on_event } = config;
6
41
  function emit(event) {
@@ -0,0 +1,46 @@
1
+ /**
2
+ * @module Backends
3
+ * @description Layered backend for caching and replication strategies.
4
+ */
5
+ import type { Backend } from '../types';
6
+ export type LayeredBackendOptions = {
7
+ read: Backend[];
8
+ write: Backend[];
9
+ list_strategy?: 'merge' | 'first';
10
+ };
11
+ /**
12
+ * Creates a layered backend that combines multiple backends with read/write separation.
13
+ * @category Backends
14
+ * @group Composite Backends
15
+ *
16
+ * Read operations use fallback: tries each read backend in order until one succeeds.
17
+ * Write operations use fanout: writes to all write backends (fails if any fail).
18
+ *
19
+ * Common use cases:
20
+ * - **Caching**: Memory backend first for reads, file backend for persistence
21
+ * - **Replication**: Write to multiple backends for redundancy
22
+ * - **Migration**: Read from old + new backends, write only to new
23
+ *
24
+ * @param options - Configuration with `read` backends (tried in order), `write` backends (all receive writes), and optional `list_strategy` ('merge' or 'first')
25
+ * @returns A Backend that delegates to the configured backends
26
+ *
27
+ * @example
28
+ * ```ts
29
+ * // Caching layer: memory cache with file persistence
30
+ * const cache = create_memory_backend()
31
+ * const storage = create_file_backend({ base_path: './data' })
32
+ *
33
+ * const backend = create_layered_backend({
34
+ * read: [cache, storage], // Try cache first, fall back to disk
35
+ * write: [cache, storage], // Write to both
36
+ * })
37
+ *
38
+ * // Migration: read from old and new, write only to new
39
+ * const backend = create_layered_backend({
40
+ * read: [newBackend, oldBackend],
41
+ * write: [newBackend],
42
+ * })
43
+ * ```
44
+ */
45
+ export declare function create_layered_backend(options: LayeredBackendOptions): Backend;
46
+ //# sourceMappingURL=layered.d.ts.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"layered.d.ts","sourceRoot":"","sources":["../../backend/layered.ts"],"names":[],"mappings":"AAAA;;;GAGG;AAEH,OAAO,KAAK,EAAE,OAAO,EAA6E,MAAM,UAAU,CAAA;AAGlH,MAAM,MAAM,qBAAqB,GAAG;IAClC,IAAI,EAAE,OAAO,EAAE,CAAA;IACf,KAAK,EAAE,OAAO,EAAE,CAAA;IAChB,aAAa,CAAC,EAAE,OAAO,GAAG,OAAO,CAAA;CAClC,CAAA;AAED;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;GAiCG;AACH,wBAAgB,sBAAsB,CAAC,OAAO,EAAE,qBAAqB,GAAG,OAAO,CA4I9E"}
@@ -0,0 +1,193 @@
1
+ /**
2
+ * @module Backends
3
+ * @description Layered backend for caching and replication strategies.
4
+ */
5
+ import { ok, err } from '../types';
6
+ /**
7
+ * Creates a layered backend that combines multiple backends with read/write separation.
8
+ * @category Backends
9
+ * @group Composite Backends
10
+ *
11
+ * Read operations use fallback: tries each read backend in order until one succeeds.
12
+ * Write operations use fanout: writes to all write backends (fails if any fail).
13
+ *
14
+ * Common use cases:
15
+ * - **Caching**: Memory backend first for reads, file backend for persistence
16
+ * - **Replication**: Write to multiple backends for redundancy
17
+ * - **Migration**: Read from old + new backends, write only to new
18
+ *
19
+ * @param options - Configuration with `read` backends (tried in order), `write` backends (all receive writes), and optional `list_strategy` ('merge' or 'first')
20
+ * @returns A Backend that delegates to the configured backends
21
+ *
22
+ * @example
23
+ * ```ts
24
+ * // Caching layer: memory cache with file persistence
25
+ * const cache = create_memory_backend()
26
+ * const storage = create_file_backend({ base_path: './data' })
27
+ *
28
+ * const backend = create_layered_backend({
29
+ * read: [cache, storage], // Try cache first, fall back to disk
30
+ * write: [cache, storage], // Write to both
31
+ * })
32
+ *
33
+ * // Migration: read from old and new, write only to new
34
+ * const backend = create_layered_backend({
35
+ * read: [newBackend, oldBackend],
36
+ * write: [newBackend],
37
+ * })
38
+ * ```
39
+ */
40
+ export function create_layered_backend(options) {
41
+ const { read, write, list_strategy = 'merge' } = options;
42
+ const metadata = {
43
+ async get(store_id, version) {
44
+ for (const backend of read) {
45
+ const result = await backend.metadata.get(store_id, version);
46
+ if (result.ok)
47
+ return result;
48
+ if (result.error.kind !== 'not_found')
49
+ return result;
50
+ }
51
+ return err({ kind: 'not_found', store_id, version });
52
+ },
53
+ async put(meta) {
54
+ for (const backend of write) {
55
+ const result = await backend.metadata.put(meta);
56
+ if (!result.ok)
57
+ return result;
58
+ }
59
+ return ok(undefined);
60
+ },
61
+ async delete(store_id, version) {
62
+ for (const backend of write) {
63
+ const result = await backend.metadata.delete(store_id, version);
64
+ if (!result.ok && result.error.kind !== 'not_found')
65
+ return result;
66
+ }
67
+ return ok(undefined);
68
+ },
69
+ async *list(store_id, opts) {
70
+ if (read.length === 0)
71
+ return;
72
+ if (list_strategy === 'first') {
73
+ yield* read[0].metadata.list(store_id, opts);
74
+ return;
75
+ }
76
+ const seen = new Set();
77
+ const all = [];
78
+ for (const backend of read) {
79
+ for await (const meta of backend.metadata.list(store_id, opts)) {
80
+ if (seen.has(meta.version))
81
+ continue;
82
+ seen.add(meta.version);
83
+ all.push(meta);
84
+ }
85
+ }
86
+ all.sort((a, b) => b.created_at.getTime() - a.created_at.getTime());
87
+ const limit = opts?.limit ?? Infinity;
88
+ for (const meta of all.slice(0, limit)) {
89
+ yield meta;
90
+ }
91
+ },
92
+ async get_latest(store_id) {
93
+ let latest = null;
94
+ for (const backend of read) {
95
+ const result = await backend.metadata.get_latest(store_id);
96
+ if (!result.ok) {
97
+ if (result.error.kind !== 'not_found')
98
+ return result;
99
+ continue;
100
+ }
101
+ if (!latest || result.value.created_at > latest.created_at) {
102
+ latest = result.value;
103
+ }
104
+ }
105
+ if (!latest) {
106
+ return err({ kind: 'not_found', store_id, version: 'latest' });
107
+ }
108
+ return ok(latest);
109
+ },
110
+ async *get_children(parent_store_id, parent_version) {
111
+ const seen = new Set();
112
+ for (const backend of read) {
113
+ for await (const meta of backend.metadata.get_children(parent_store_id, parent_version)) {
114
+ const key = `${meta.store_id}:${meta.version}`;
115
+ if (seen.has(key))
116
+ continue;
117
+ seen.add(key);
118
+ yield meta;
119
+ }
120
+ }
121
+ },
122
+ async find_by_hash(store_id, content_hash) {
123
+ for (const backend of read) {
124
+ const result = await backend.metadata.find_by_hash(store_id, content_hash);
125
+ if (result)
126
+ return result;
127
+ }
128
+ return null;
129
+ },
130
+ };
131
+ const data = {
132
+ async get(data_key) {
133
+ for (const backend of read) {
134
+ const result = await backend.data.get(data_key);
135
+ if (result.ok)
136
+ return result;
137
+ if (result.error.kind !== 'not_found')
138
+ return result;
139
+ }
140
+ return err({ kind: 'not_found', store_id: data_key, version: '' });
141
+ },
142
+ async put(data_key, data) {
143
+ if (write.length === 0)
144
+ return ok(undefined);
145
+ if (write.length === 1) {
146
+ return write[0].data.put(data_key, data);
147
+ }
148
+ const bytes = await to_bytes(data);
149
+ for (const backend of write) {
150
+ const result = await backend.data.put(data_key, bytes);
151
+ if (!result.ok)
152
+ return result;
153
+ }
154
+ return ok(undefined);
155
+ },
156
+ async delete(data_key) {
157
+ for (const backend of write) {
158
+ const result = await backend.data.delete(data_key);
159
+ if (!result.ok && result.error.kind !== 'not_found')
160
+ return result;
161
+ }
162
+ return ok(undefined);
163
+ },
164
+ async exists(data_key) {
165
+ for (const backend of read) {
166
+ if (await backend.data.exists(data_key))
167
+ return true;
168
+ }
169
+ return false;
170
+ },
171
+ };
172
+ return { metadata, data };
173
+ }
174
+ async function to_bytes(data) {
175
+ if (data instanceof Uint8Array)
176
+ return data;
177
+ const chunks = [];
178
+ const reader = data.getReader();
179
+ while (true) {
180
+ const { done, value } = await reader.read();
181
+ if (done)
182
+ break;
183
+ chunks.push(value);
184
+ }
185
+ const total = chunks.reduce((sum, c) => sum + c.length, 0);
186
+ const result = new Uint8Array(total);
187
+ let offset = 0;
188
+ for (const chunk of chunks) {
189
+ result.set(chunk, offset);
190
+ offset += chunk.length;
191
+ }
192
+ return result;
193
+ }
@@ -1,6 +1,36 @@
1
+ /**
2
+ * @module Backends
3
+ * @description In-memory storage backend for testing and development.
4
+ */
1
5
  import type { Backend, EventHandler } from '../types';
2
6
  export type MemoryBackendOptions = {
3
7
  on_event?: EventHandler;
4
8
  };
9
+ /**
10
+ * Creates an in-memory storage backend.
11
+ * @category Backends
12
+ * @group Storage Backends
13
+ *
14
+ * Ideal for testing, development, and ephemeral storage scenarios.
15
+ * All data is lost when the process ends.
16
+ *
17
+ * @param options - Optional configuration with `on_event` handler for observability
18
+ * @returns A Backend instance using in-memory storage
19
+ *
20
+ * @example
21
+ * ```ts
22
+ * // Basic usage for testing
23
+ * const backend = create_memory_backend()
24
+ * const corpus = create_corpus()
25
+ * .with_backend(backend)
26
+ * .with_store(define_store('test', text_codec()))
27
+ * .build()
28
+ *
29
+ * // With event logging
30
+ * const backend = create_memory_backend({
31
+ * on_event: (e) => console.log(`[${e.type}]`, e)
32
+ * })
33
+ * ```
34
+ */
5
35
  export declare function create_memory_backend(options?: MemoryBackendOptions): Backend;
6
36
  //# sourceMappingURL=memory.d.ts.map
@@ -1 +1 @@
1
- {"version":3,"file":"memory.d.ts","sourceRoot":"","sources":["../../backend/memory.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,OAAO,EAAwF,YAAY,EAAE,MAAM,UAAU,CAAA;AAG3I,MAAM,MAAM,oBAAoB,GAAG;IACjC,QAAQ,CAAC,EAAE,YAAY,CAAA;CACxB,CAAA;AAED,wBAAgB,qBAAqB,CAAC,OAAO,CAAC,EAAE,oBAAoB,GAAG,OAAO,CAiJ7E"}
1
+ {"version":3,"file":"memory.d.ts","sourceRoot":"","sources":["../../backend/memory.ts"],"names":[],"mappings":"AAAA;;;GAGG;AAEH,OAAO,KAAK,EAAE,OAAO,EAAwF,YAAY,EAAE,MAAM,UAAU,CAAA;AAG3I,MAAM,MAAM,oBAAoB,GAAG;IACjC,QAAQ,CAAC,EAAE,YAAY,CAAA;CACxB,CAAA;AAED;;;;;;;;;;;;;;;;;;;;;;;;;GAyBG;AACH,wBAAgB,qBAAqB,CAAC,OAAO,CAAC,EAAE,oBAAoB,GAAG,OAAO,CAiJ7E"}
@@ -1,4 +1,34 @@
1
+ /**
2
+ * @module Backends
3
+ * @description In-memory storage backend for testing and development.
4
+ */
1
5
  import { ok, err } from '../types';
6
+ /**
7
+ * Creates an in-memory storage backend.
8
+ * @category Backends
9
+ * @group Storage Backends
10
+ *
11
+ * Ideal for testing, development, and ephemeral storage scenarios.
12
+ * All data is lost when the process ends.
13
+ *
14
+ * @param options - Optional configuration with `on_event` handler for observability
15
+ * @returns A Backend instance using in-memory storage
16
+ *
17
+ * @example
18
+ * ```ts
19
+ * // Basic usage for testing
20
+ * const backend = create_memory_backend()
21
+ * const corpus = create_corpus()
22
+ * .with_backend(backend)
23
+ * .with_store(define_store('test', text_codec()))
24
+ * .build()
25
+ *
26
+ * // With event logging
27
+ * const backend = create_memory_backend({
28
+ * on_event: (e) => console.log(`[${e.type}]`, e)
29
+ * })
30
+ * ```
31
+ */
2
32
  export function create_memory_backend(options) {
3
33
  const meta_store = new Map();
4
34
  const data_store = new Map();
@@ -0,0 +1,11 @@
1
+ /**
2
+ * Storage backend implementations for different environments.
3
+ * @module Backends
4
+ * @packageDocumentation
5
+ */
6
+ export { create_memory_backend, type MemoryBackendOptions } from './backend/memory';
7
+ export { create_file_backend, type FileBackendConfig } from './backend/file';
8
+ export { create_cloudflare_backend, type CloudflareBackendConfig } from './backend/cloudflare';
9
+ export { create_layered_backend, type LayeredBackendOptions } from './backend/layered';
10
+ export type { Backend, MetadataClient, DataClient, DataHandle, EventHandler, CorpusEvent } from './types';
11
+ //# sourceMappingURL=backends.d.ts.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"backends.d.ts","sourceRoot":"","sources":["../backends.ts"],"names":[],"mappings":"AAAA;;;;GAIG;AAEH,OAAO,EAAE,qBAAqB,EAAE,KAAK,oBAAoB,EAAE,MAAM,kBAAkB,CAAA;AACnF,OAAO,EAAE,mBAAmB,EAAE,KAAK,iBAAiB,EAAE,MAAM,gBAAgB,CAAA;AAC5E,OAAO,EAAE,yBAAyB,EAAE,KAAK,uBAAuB,EAAE,MAAM,sBAAsB,CAAA;AAC9F,OAAO,EAAE,sBAAsB,EAAE,KAAK,qBAAqB,EAAE,MAAM,mBAAmB,CAAA;AACtF,YAAY,EAAE,OAAO,EAAE,cAAc,EAAE,UAAU,EAAE,UAAU,EAAE,YAAY,EAAE,WAAW,EAAE,MAAM,SAAS,CAAA"}
@@ -0,0 +1,9 @@
1
+ /**
2
+ * Storage backend implementations for different environments.
3
+ * @module Backends
4
+ * @packageDocumentation
5
+ */
6
+ export { create_memory_backend } from './backend/memory';
7
+ export { create_file_backend } from './backend/file';
8
+ export { create_cloudflare_backend } from './backend/cloudflare';
9
+ export { create_layered_backend } from './backend/layered';
@@ -2,14 +2,11 @@
2
2
  * Cloudflare Workers compatible exports
3
3
  * This entry point excludes the file backend which uses Node.js APIs
4
4
  */
5
- export { create_corpus } from './corpus';
6
- export { create_store } from './store';
5
+ export { create_corpus, create_store } from './corpus';
7
6
  export { create_memory_backend, type MemoryBackendOptions } from './backend/memory';
8
7
  export { create_cloudflare_backend, type CloudflareBackendConfig } from './backend/cloudflare';
9
- export { json_codec, text_codec, binary_codec } from './codec';
8
+ export { json_codec, text_codec, binary_codec, compute_hash, generate_version } from './utils';
10
9
  export { corpus_snapshots, type CorpusSnapshotRow, type CorpusSnapshotInsert } from './schema';
11
- export { compute_hash } from './hash';
12
- export { generate_version } from './version';
13
10
  export type { ContentType, ParentRef, SnapshotMeta, Snapshot, DataHandle, MetadataClient, DataClient, ListOpts, Backend, Codec, Store, StoreDefinition, PutOpts, CorpusBuilder, Corpus, CorpusError, Result, CorpusEvent, EventHandler, } from './types';
14
11
  export { ok, err, define_store } from './types';
15
12
  //# sourceMappingURL=cloudflare.d.ts.map
@@ -1 +1 @@
1
- {"version":3,"file":"cloudflare.d.ts","sourceRoot":"","sources":["../cloudflare.ts"],"names":[],"mappings":"AAAA;;;GAGG;AAEH,OAAO,EAAE,aAAa,EAAE,MAAM,UAAU,CAAA;AACxC,OAAO,EAAE,YAAY,EAAE,MAAM,SAAS,CAAA;AAEtC,OAAO,EAAE,qBAAqB,EAAE,KAAK,oBAAoB,EAAE,MAAM,kBAAkB,CAAA;AACnF,OAAO,EAAE,yBAAyB,EAAE,KAAK,uBAAuB,EAAE,MAAM,sBAAsB,CAAA;AAE9F,OAAO,EAAE,UAAU,EAAE,UAAU,EAAE,YAAY,EAAE,MAAM,SAAS,CAAA;AAE9D,OAAO,EAAE,gBAAgB,EAAE,KAAK,iBAAiB,EAAE,KAAK,oBAAoB,EAAE,MAAM,UAAU,CAAA;AAE9F,OAAO,EAAE,YAAY,EAAE,MAAM,QAAQ,CAAA;AACrC,OAAO,EAAE,gBAAgB,EAAE,MAAM,WAAW,CAAA;AAE5C,YAAY,EACV,WAAW,EACX,SAAS,EACT,YAAY,EACZ,QAAQ,EACR,UAAU,EACV,cAAc,EACd,UAAU,EACV,QAAQ,EACR,OAAO,EACP,KAAK,EACL,KAAK,EACL,eAAe,EACf,OAAO,EACP,aAAa,EACb,MAAM,EACN,WAAW,EACX,MAAM,EACN,WAAW,EACX,YAAY,GACb,MAAM,SAAS,CAAA;AAEhB,OAAO,EAAE,EAAE,EAAE,GAAG,EAAE,YAAY,EAAE,MAAM,SAAS,CAAA"}
1
+ {"version":3,"file":"cloudflare.d.ts","sourceRoot":"","sources":["../cloudflare.ts"],"names":[],"mappings":"AAAA;;;GAGG;AAEH,OAAO,EAAE,aAAa,EAAE,YAAY,EAAE,MAAM,UAAU,CAAA;AAEtD,OAAO,EAAE,qBAAqB,EAAE,KAAK,oBAAoB,EAAE,MAAM,kBAAkB,CAAA;AACnF,OAAO,EAAE,yBAAyB,EAAE,KAAK,uBAAuB,EAAE,MAAM,sBAAsB,CAAA;AAE9F,OAAO,EAAE,UAAU,EAAE,UAAU,EAAE,YAAY,EAAE,YAAY,EAAE,gBAAgB,EAAE,MAAM,SAAS,CAAA;AAE9F,OAAO,EAAE,gBAAgB,EAAE,KAAK,iBAAiB,EAAE,KAAK,oBAAoB,EAAE,MAAM,UAAU,CAAA;AAE9F,YAAY,EACV,WAAW,EACX,SAAS,EACT,YAAY,EACZ,QAAQ,EACR,UAAU,EACV,cAAc,EACd,UAAU,EACV,QAAQ,EACR,OAAO,EACP,KAAK,EACL,KAAK,EACL,eAAe,EACf,OAAO,EACP,aAAa,EACb,MAAM,EACN,WAAW,EACX,MAAM,EACN,WAAW,EACX,YAAY,GACb,MAAM,SAAS,CAAA;AAEhB,OAAO,EAAE,EAAE,EAAE,GAAG,EAAE,YAAY,EAAE,MAAM,SAAS,CAAA"}
@@ -2,12 +2,9 @@
2
2
  * Cloudflare Workers compatible exports
3
3
  * This entry point excludes the file backend which uses Node.js APIs
4
4
  */
5
- export { create_corpus } from './corpus';
6
- export { create_store } from './store';
5
+ export { create_corpus, create_store } from './corpus';
7
6
  export { create_memory_backend } from './backend/memory';
8
7
  export { create_cloudflare_backend } from './backend/cloudflare';
9
- export { json_codec, text_codec, binary_codec } from './codec';
8
+ export { json_codec, text_codec, binary_codec, compute_hash, generate_version } from './utils';
10
9
  export { corpus_snapshots } from './schema';
11
- export { compute_hash } from './hash';
12
- export { generate_version } from './version';
13
10
  export { ok, err, define_store } from './types';
@@ -0,0 +1,8 @@
1
+ /**
2
+ * Codec implementations for serializing and deserializing data.
3
+ * @module Codecs
4
+ * @packageDocumentation
5
+ */
6
+ export { json_codec, text_codec, binary_codec } from './utils';
7
+ export type { Codec, ContentType } from './types';
8
+ //# sourceMappingURL=codecs.d.ts.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"codecs.d.ts","sourceRoot":"","sources":["../codecs.ts"],"names":[],"mappings":"AAAA;;;;GAIG;AAEH,OAAO,EAAE,UAAU,EAAE,UAAU,EAAE,YAAY,EAAE,MAAM,SAAS,CAAA;AAC9D,YAAY,EAAE,KAAK,EAAE,WAAW,EAAE,MAAM,SAAS,CAAA"}
package/dist/codecs.js ADDED
@@ -0,0 +1,6 @@
1
+ /**
2
+ * Codec implementations for serializing and deserializing data.
3
+ * @module Codecs
4
+ * @packageDocumentation
5
+ */
6
+ export { json_codec, text_codec, binary_codec } from './utils';
package/dist/core.d.ts ADDED
@@ -0,0 +1,9 @@
1
+ /**
2
+ * Core corpus functionality for creating and managing versioned data stores.
3
+ * @module Core
4
+ * @packageDocumentation
5
+ */
6
+ export { create_corpus, create_store } from './corpus';
7
+ export { define_store, ok, err } from './types';
8
+ export type { Corpus, CorpusBuilder, Store, StoreDefinition, Result, CorpusError, PutOpts, } from './types';
9
+ //# sourceMappingURL=core.d.ts.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"core.d.ts","sourceRoot":"","sources":["../core.ts"],"names":[],"mappings":"AAAA;;;;GAIG;AAEH,OAAO,EAAE,aAAa,EAAE,YAAY,EAAE,MAAM,UAAU,CAAA;AACtD,OAAO,EAAE,YAAY,EAAE,EAAE,EAAE,GAAG,EAAE,MAAM,SAAS,CAAA;AAC/C,YAAY,EACV,MAAM,EACN,aAAa,EACb,KAAK,EACL,eAAe,EACf,MAAM,EACN,WAAW,EACX,OAAO,GACR,MAAM,SAAS,CAAA"}
package/dist/core.js ADDED
@@ -0,0 +1,7 @@
1
+ /**
2
+ * Core corpus functionality for creating and managing versioned data stores.
3
+ * @module Core
4
+ * @packageDocumentation
5
+ */
6
+ export { create_corpus, create_store } from './corpus';
7
+ export { define_store, ok, err } from './types';