rvlite 0.2.1
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +270 -0
- package/bin/cli.js +1685 -0
- package/dist/wasm/README.md +216 -0
- package/dist/wasm/attention/LICENSE +21 -0
- package/dist/wasm/attention/README.md +193 -0
- package/dist/wasm/attention/package.json +17 -0
- package/dist/wasm/attention/ruvector_attention_wasm.d.ts +334 -0
- package/dist/wasm/attention/ruvector_attention_wasm.js +1470 -0
- package/dist/wasm/attention/ruvector_attention_wasm_bg.wasm +0 -0
- package/dist/wasm/attention/ruvector_attention_wasm_bg.wasm.d.ts +71 -0
- package/dist/wasm/package.json +24 -0
- package/dist/wasm/rvlite.d.ts +276 -0
- package/dist/wasm/rvlite.js +1504 -0
- package/dist/wasm/rvlite_bg.wasm +0 -0
- package/dist/wasm/rvlite_bg.wasm.d.ts +56 -0
- package/dist/wasm/sona/LICENSE-APACHE +103 -0
- package/dist/wasm/sona/LICENSE-MIT +21 -0
- package/dist/wasm/sona/README.md +1513 -0
- package/dist/wasm/sona/package.json +36 -0
- package/dist/wasm/sona/ruvector_sona.d.ts +513 -0
- package/dist/wasm/sona/ruvector_sona.js +1286 -0
- package/dist/wasm/sona/ruvector_sona_bg.wasm +0 -0
- package/dist/wasm/sona/ruvector_sona_bg.wasm.d.ts +53 -0
- package/dist/wasm/sona/sona.d.ts +281 -0
- package/dist/wasm/sona/sona.js +685 -0
- package/dist/wasm/sona/sona_bg.wasm +0 -0
- package/dist/wasm/sona/sona_bg.wasm.d.ts +26 -0
- package/package.json +81 -0
|
Binary file
|
|
@@ -0,0 +1,53 @@
|
|
|
1
|
+
/* tslint:disable */
|
|
2
|
+
/* eslint-disable */
|
|
3
|
+
export const memory: WebAssembly.Memory;
|
|
4
|
+
export const __wbg_wasmephemeralagent_free: (a: number, b: number) => void;
|
|
5
|
+
export const __wbg_wasmfederatedcoordinator_free: (a: number, b: number) => void;
|
|
6
|
+
export const __wbg_wasmsonaengine_free: (a: number, b: number) => void;
|
|
7
|
+
export const wasmephemeralagent_averageQuality: (a: number) => number;
|
|
8
|
+
export const wasmephemeralagent_clear: (a: number) => void;
|
|
9
|
+
export const wasmephemeralagent_exportState: (a: number) => number;
|
|
10
|
+
export const wasmephemeralagent_forceLearn: (a: number, b: number) => void;
|
|
11
|
+
export const wasmephemeralagent_getPatterns: (a: number) => number;
|
|
12
|
+
export const wasmephemeralagent_getStats: (a: number) => number;
|
|
13
|
+
export const wasmephemeralagent_new: (a: number, b: number, c: number) => void;
|
|
14
|
+
export const wasmephemeralagent_processTask: (a: number, b: number, c: number, d: number) => void;
|
|
15
|
+
export const wasmephemeralagent_processTaskWithRoute: (a: number, b: number, c: number, d: number, e: number, f: number) => void;
|
|
16
|
+
export const wasmephemeralagent_trajectoryCount: (a: number) => number;
|
|
17
|
+
export const wasmephemeralagent_uptimeSeconds: (a: number) => bigint;
|
|
18
|
+
export const wasmephemeralagent_withConfig: (a: number, b: number, c: number, d: number) => void;
|
|
19
|
+
export const wasmfederatedcoordinator_agentCount: (a: number) => number;
|
|
20
|
+
export const wasmfederatedcoordinator_aggregate: (a: number, b: number) => number;
|
|
21
|
+
export const wasmfederatedcoordinator_applyLora: (a: number, b: number, c: number, d: number) => void;
|
|
22
|
+
export const wasmfederatedcoordinator_clear: (a: number) => void;
|
|
23
|
+
export const wasmfederatedcoordinator_consolidate: (a: number, b: number) => void;
|
|
24
|
+
export const wasmfederatedcoordinator_findPatterns: (a: number, b: number, c: number, d: number) => number;
|
|
25
|
+
export const wasmfederatedcoordinator_getPatterns: (a: number) => number;
|
|
26
|
+
export const wasmfederatedcoordinator_getStats: (a: number) => number;
|
|
27
|
+
export const wasmfederatedcoordinator_new: (a: number, b: number, c: number) => void;
|
|
28
|
+
export const wasmfederatedcoordinator_setQualityThreshold: (a: number, b: number) => void;
|
|
29
|
+
export const wasmfederatedcoordinator_totalTrajectories: (a: number) => number;
|
|
30
|
+
export const wasmfederatedcoordinator_withConfig: (a: number, b: number, c: number, d: number) => void;
|
|
31
|
+
export const wasmsonaengine_applyLora: (a: number, b: number, c: number, d: number) => void;
|
|
32
|
+
export const wasmsonaengine_applyLoraLayer: (a: number, b: number, c: number, d: number, e: number) => void;
|
|
33
|
+
export const wasmsonaengine_endTrajectory: (a: number, b: bigint, c: number) => void;
|
|
34
|
+
export const wasmsonaengine_findPatterns: (a: number, b: number, c: number, d: number) => number;
|
|
35
|
+
export const wasmsonaengine_forceLearn: (a: number, b: number) => void;
|
|
36
|
+
export const wasmsonaengine_getConfig: (a: number) => number;
|
|
37
|
+
export const wasmsonaengine_getStats: (a: number) => number;
|
|
38
|
+
export const wasmsonaengine_isEnabled: (a: number) => number;
|
|
39
|
+
export const wasmsonaengine_learnFromFeedback: (a: number, b: number, c: number, d: number) => void;
|
|
40
|
+
export const wasmsonaengine_new: (a: number, b: number) => void;
|
|
41
|
+
export const wasmsonaengine_recordStep: (a: number, b: bigint, c: number, d: number, e: bigint) => void;
|
|
42
|
+
export const wasmsonaengine_runInstantCycle: (a: number) => void;
|
|
43
|
+
export const wasmsonaengine_setEnabled: (a: number, b: number) => void;
|
|
44
|
+
export const wasmsonaengine_startTrajectory: (a: number, b: number, c: number) => bigint;
|
|
45
|
+
export const wasmsonaengine_tick: (a: number) => number;
|
|
46
|
+
export const wasmsonaengine_withConfig: (a: number, b: number) => void;
|
|
47
|
+
export const wasm_init: () => void;
|
|
48
|
+
export const __wbindgen_export: (a: number, b: number) => number;
|
|
49
|
+
export const __wbindgen_export2: (a: number, b: number, c: number, d: number) => number;
|
|
50
|
+
export const __wbindgen_export3: (a: number) => void;
|
|
51
|
+
export const __wbindgen_export4: (a: number, b: number, c: number) => void;
|
|
52
|
+
export const __wbindgen_add_to_stack_pointer: (a: number) => number;
|
|
53
|
+
export const __wbindgen_start: () => void;
|
|
@@ -0,0 +1,281 @@
|
|
|
1
|
+
/* tslint:disable */
|
|
2
|
+
/* eslint-disable */
|
|
3
|
+
/**
|
|
4
|
+
* Initialize WASM module (called automatically)
|
|
5
|
+
*/
|
|
6
|
+
export function wasm_init(): void;
|
|
7
|
+
/**
|
|
8
|
+
* WASM-compatible SONA Engine wrapper
|
|
9
|
+
*
|
|
10
|
+
* Provides JavaScript bindings for the SONA adaptive learning system.
|
|
11
|
+
*/
|
|
12
|
+
export class WasmSonaEngine {
|
|
13
|
+
free(): void;
|
|
14
|
+
[Symbol.dispose](): void;
|
|
15
|
+
/**
|
|
16
|
+
* Apply LoRA transformation to input vector
|
|
17
|
+
*
|
|
18
|
+
* # Arguments
|
|
19
|
+
* * `input` - Input vector as Float32Array
|
|
20
|
+
*
|
|
21
|
+
* # Returns
|
|
22
|
+
* Transformed vector as Float32Array
|
|
23
|
+
*
|
|
24
|
+
* # Example
|
|
25
|
+
* ```javascript
|
|
26
|
+
* const input = new Float32Array(256).fill(1.0);
|
|
27
|
+
* const output = engine.apply_lora(input);
|
|
28
|
+
* ```
|
|
29
|
+
*/
|
|
30
|
+
applyLora(input: Float32Array): Float32Array;
|
|
31
|
+
/**
|
|
32
|
+
* Get configuration
|
|
33
|
+
*
|
|
34
|
+
* # Returns
|
|
35
|
+
* Configuration as JSON object
|
|
36
|
+
*/
|
|
37
|
+
getConfig(): any;
|
|
38
|
+
/**
|
|
39
|
+
* Check if engine is enabled
|
|
40
|
+
*
|
|
41
|
+
* # Returns
|
|
42
|
+
* true if enabled, false otherwise
|
|
43
|
+
*/
|
|
44
|
+
isEnabled(): boolean;
|
|
45
|
+
/**
|
|
46
|
+
* Force background learning cycle
|
|
47
|
+
*
|
|
48
|
+
* # Returns
|
|
49
|
+
* Learning statistics as JSON string
|
|
50
|
+
*
|
|
51
|
+
* # Example
|
|
52
|
+
* ```javascript
|
|
53
|
+
* const stats = engine.force_learn();
|
|
54
|
+
* console.log('Learning results:', stats);
|
|
55
|
+
* ```
|
|
56
|
+
*/
|
|
57
|
+
forceLearn(): string;
|
|
58
|
+
/**
|
|
59
|
+
* Record a step in the trajectory
|
|
60
|
+
*
|
|
61
|
+
* # Arguments
|
|
62
|
+
* * `trajectory_id` - ID returned from start_trajectory
|
|
63
|
+
* * `node_id` - Graph node visited
|
|
64
|
+
* * `score` - Step quality score [0.0, 1.0]
|
|
65
|
+
* * `latency_us` - Step latency in microseconds
|
|
66
|
+
*
|
|
67
|
+
* # Example
|
|
68
|
+
* ```javascript
|
|
69
|
+
* engine.record_step(trajectoryId, 42, 0.8, 1000);
|
|
70
|
+
* ```
|
|
71
|
+
*/
|
|
72
|
+
recordStep(trajectory_id: bigint, node_id: number, score: number, latency_us: bigint): void;
|
|
73
|
+
/**
|
|
74
|
+
* Enable or disable the engine
|
|
75
|
+
*
|
|
76
|
+
* # Arguments
|
|
77
|
+
* * `enabled` - Whether to enable the engine
|
|
78
|
+
*
|
|
79
|
+
* # Example
|
|
80
|
+
* ```javascript
|
|
81
|
+
* engine.set_enabled(false); // Pause learning
|
|
82
|
+
* ```
|
|
83
|
+
*/
|
|
84
|
+
setEnabled(enabled: boolean): void;
|
|
85
|
+
/**
|
|
86
|
+
* Create engine with custom configuration
|
|
87
|
+
*
|
|
88
|
+
* # Arguments
|
|
89
|
+
* * `config` - JSON configuration object
|
|
90
|
+
*
|
|
91
|
+
* # Example
|
|
92
|
+
* ```javascript
|
|
93
|
+
* const config = {
|
|
94
|
+
* hidden_dim: 256,
|
|
95
|
+
* embedding_dim: 256,
|
|
96
|
+
* micro_lora_rank: 2,
|
|
97
|
+
* base_lora_rank: 16,
|
|
98
|
+
* micro_lora_lr: 0.001,
|
|
99
|
+
* base_lora_lr: 0.0001,
|
|
100
|
+
* ewc_lambda: 1000.0,
|
|
101
|
+
* pattern_clusters: 128,
|
|
102
|
+
* trajectory_capacity: 10000,
|
|
103
|
+
* quality_threshold: 0.6
|
|
104
|
+
* };
|
|
105
|
+
* const engine = WasmSonaEngine.with_config(config);
|
|
106
|
+
* ```
|
|
107
|
+
*/
|
|
108
|
+
static withConfig(config: any): WasmSonaEngine;
|
|
109
|
+
/**
|
|
110
|
+
* Find similar patterns to query
|
|
111
|
+
*
|
|
112
|
+
* # Arguments
|
|
113
|
+
* * `query_embedding` - Query vector as Float32Array
|
|
114
|
+
* * `k` - Number of patterns to return
|
|
115
|
+
*
|
|
116
|
+
* # Returns
|
|
117
|
+
* Array of similar patterns as JSON
|
|
118
|
+
*
|
|
119
|
+
* # Example
|
|
120
|
+
* ```javascript
|
|
121
|
+
* const query = new Float32Array(256).fill(0.5);
|
|
122
|
+
* const patterns = engine.find_patterns(query, 5);
|
|
123
|
+
* console.log('Similar patterns:', patterns);
|
|
124
|
+
* ```
|
|
125
|
+
*/
|
|
126
|
+
findPatterns(query_embedding: Float32Array, k: number): any;
|
|
127
|
+
/**
|
|
128
|
+
* End the trajectory and submit for learning
|
|
129
|
+
*
|
|
130
|
+
* # Arguments
|
|
131
|
+
* * `trajectory_id` - ID returned from start_trajectory
|
|
132
|
+
* * `final_score` - Overall trajectory quality [0.0, 1.0]
|
|
133
|
+
*
|
|
134
|
+
* # Example
|
|
135
|
+
* ```javascript
|
|
136
|
+
* engine.end_trajectory(trajectoryId, 0.85);
|
|
137
|
+
* ```
|
|
138
|
+
*/
|
|
139
|
+
endTrajectory(trajectory_id: bigint, final_score: number): void;
|
|
140
|
+
/**
|
|
141
|
+
* Apply LoRA transformation to specific layer
|
|
142
|
+
*
|
|
143
|
+
* # Arguments
|
|
144
|
+
* * `layer_idx` - Layer index
|
|
145
|
+
* * `input` - Input vector as Float32Array
|
|
146
|
+
*
|
|
147
|
+
* # Returns
|
|
148
|
+
* Transformed vector as Float32Array
|
|
149
|
+
*/
|
|
150
|
+
applyLoraLayer(layer_idx: number, input: Float32Array): Float32Array;
|
|
151
|
+
/**
|
|
152
|
+
* Start recording a new trajectory
|
|
153
|
+
*
|
|
154
|
+
* # Arguments
|
|
155
|
+
* * `query_embedding` - Query vector as Float32Array
|
|
156
|
+
*
|
|
157
|
+
* # Returns
|
|
158
|
+
* Trajectory ID (u64)
|
|
159
|
+
*
|
|
160
|
+
* # Example
|
|
161
|
+
* ```javascript
|
|
162
|
+
* const embedding = new Float32Array(256).fill(0.1);
|
|
163
|
+
* const trajectoryId = engine.start_trajectory(embedding);
|
|
164
|
+
* ```
|
|
165
|
+
*/
|
|
166
|
+
startTrajectory(query_embedding: Float32Array): bigint;
|
|
167
|
+
/**
|
|
168
|
+
* Run instant learning cycle
|
|
169
|
+
*
|
|
170
|
+
* Flushes accumulated micro-LoRA updates
|
|
171
|
+
*
|
|
172
|
+
* # Example
|
|
173
|
+
* ```javascript
|
|
174
|
+
* engine.run_instant_cycle();
|
|
175
|
+
* ```
|
|
176
|
+
*/
|
|
177
|
+
runInstantCycle(): void;
|
|
178
|
+
/**
|
|
179
|
+
* Apply learning from user feedback
|
|
180
|
+
*
|
|
181
|
+
* # Arguments
|
|
182
|
+
* * `success` - Whether the operation succeeded
|
|
183
|
+
* * `latency_ms` - Operation latency in milliseconds
|
|
184
|
+
* * `quality` - User-perceived quality [0.0, 1.0]
|
|
185
|
+
*
|
|
186
|
+
* # Example
|
|
187
|
+
* ```javascript
|
|
188
|
+
* engine.learn_from_feedback(true, 50.0, 0.9);
|
|
189
|
+
* ```
|
|
190
|
+
*/
|
|
191
|
+
learnFromFeedback(success: boolean, latency_ms: number, quality: number): void;
|
|
192
|
+
/**
|
|
193
|
+
* Create a new SONA engine with specified hidden dimension
|
|
194
|
+
*
|
|
195
|
+
* # Arguments
|
|
196
|
+
* * `hidden_dim` - Size of hidden layer (typically 256, 512, or 1024)
|
|
197
|
+
*
|
|
198
|
+
* # Example
|
|
199
|
+
* ```javascript
|
|
200
|
+
* const engine = new WasmSonaEngine(256);
|
|
201
|
+
* ```
|
|
202
|
+
*/
|
|
203
|
+
constructor(hidden_dim: number);
|
|
204
|
+
/**
|
|
205
|
+
* Try to run background learning cycle
|
|
206
|
+
*
|
|
207
|
+
* Returns true if cycle was executed, false if not due yet
|
|
208
|
+
*
|
|
209
|
+
* # Example
|
|
210
|
+
* ```javascript
|
|
211
|
+
* if (engine.tick()) {
|
|
212
|
+
* console.log('Background learning completed');
|
|
213
|
+
* }
|
|
214
|
+
* ```
|
|
215
|
+
*/
|
|
216
|
+
tick(): boolean;
|
|
217
|
+
/**
|
|
218
|
+
* Get engine statistics
|
|
219
|
+
*
|
|
220
|
+
* # Returns
|
|
221
|
+
* Statistics as JSON object
|
|
222
|
+
*
|
|
223
|
+
* # Example
|
|
224
|
+
* ```javascript
|
|
225
|
+
* const stats = engine.get_stats();
|
|
226
|
+
* console.log('Trajectories buffered:', stats.trajectories_buffered);
|
|
227
|
+
* console.log('Patterns learned:', stats.patterns_learned);
|
|
228
|
+
* ```
|
|
229
|
+
*/
|
|
230
|
+
getStats(): any;
|
|
231
|
+
}
|
|
232
|
+
|
|
233
|
+
export type InitInput = RequestInfo | URL | Response | BufferSource | WebAssembly.Module;
|
|
234
|
+
|
|
235
|
+
export interface InitOutput {
|
|
236
|
+
readonly memory: WebAssembly.Memory;
|
|
237
|
+
readonly __wbg_wasmsonaengine_free: (a: number, b: number) => void;
|
|
238
|
+
readonly wasmsonaengine_applyLora: (a: number, b: number, c: number, d: number) => void;
|
|
239
|
+
readonly wasmsonaengine_applyLoraLayer: (a: number, b: number, c: number, d: number, e: number) => void;
|
|
240
|
+
readonly wasmsonaengine_endTrajectory: (a: number, b: bigint, c: number) => void;
|
|
241
|
+
readonly wasmsonaengine_findPatterns: (a: number, b: number, c: number, d: number) => number;
|
|
242
|
+
readonly wasmsonaengine_forceLearn: (a: number, b: number) => void;
|
|
243
|
+
readonly wasmsonaengine_getConfig: (a: number) => number;
|
|
244
|
+
readonly wasmsonaengine_getStats: (a: number) => number;
|
|
245
|
+
readonly wasmsonaengine_isEnabled: (a: number) => number;
|
|
246
|
+
readonly wasmsonaengine_learnFromFeedback: (a: number, b: number, c: number, d: number) => void;
|
|
247
|
+
readonly wasmsonaengine_new: (a: number, b: number) => void;
|
|
248
|
+
readonly wasmsonaengine_recordStep: (a: number, b: bigint, c: number, d: number, e: bigint) => void;
|
|
249
|
+
readonly wasmsonaengine_runInstantCycle: (a: number) => void;
|
|
250
|
+
readonly wasmsonaengine_setEnabled: (a: number, b: number) => void;
|
|
251
|
+
readonly wasmsonaengine_startTrajectory: (a: number, b: number, c: number) => bigint;
|
|
252
|
+
readonly wasmsonaengine_tick: (a: number) => number;
|
|
253
|
+
readonly wasmsonaengine_withConfig: (a: number, b: number) => void;
|
|
254
|
+
readonly wasm_init: () => void;
|
|
255
|
+
readonly __wbindgen_export: (a: number, b: number) => number;
|
|
256
|
+
readonly __wbindgen_export2: (a: number, b: number, c: number, d: number) => number;
|
|
257
|
+
readonly __wbindgen_export3: (a: number, b: number, c: number) => void;
|
|
258
|
+
readonly __wbindgen_add_to_stack_pointer: (a: number) => number;
|
|
259
|
+
readonly __wbindgen_start: () => void;
|
|
260
|
+
}
|
|
261
|
+
|
|
262
|
+
export type SyncInitInput = BufferSource | WebAssembly.Module;
|
|
263
|
+
/**
|
|
264
|
+
* Instantiates the given `module`, which can either be bytes or
|
|
265
|
+
* a precompiled `WebAssembly.Module`.
|
|
266
|
+
*
|
|
267
|
+
* @param {{ module: SyncInitInput }} module - Passing `SyncInitInput` directly is deprecated.
|
|
268
|
+
*
|
|
269
|
+
* @returns {InitOutput}
|
|
270
|
+
*/
|
|
271
|
+
export function initSync(module: { module: SyncInitInput } | SyncInitInput): InitOutput;
|
|
272
|
+
|
|
273
|
+
/**
|
|
274
|
+
* If `module_or_path` is {RequestInfo} or {URL}, makes a request and
|
|
275
|
+
* for everything else, calls `WebAssembly.instantiate` directly.
|
|
276
|
+
*
|
|
277
|
+
* @param {{ module_or_path: InitInput | Promise<InitInput> }} module_or_path - Passing `InitInput` directly is deprecated.
|
|
278
|
+
*
|
|
279
|
+
* @returns {Promise<InitOutput>}
|
|
280
|
+
*/
|
|
281
|
+
export default function __wbg_init (module_or_path?: { module_or_path: InitInput | Promise<InitInput> } | InitInput | Promise<InitInput>): Promise<InitOutput>;
|