ruvector 0.2.15 → 0.2.17
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/bin/mcp-server.js +1 -1
- package/dist/core/adaptive-embedder.d.ts.map +1 -1
- package/dist/core/adaptive-embedder.js +4 -3
- package/dist/core/onnx/pkg/ruvector_onnx_embeddings_wasm_cjs.js +127 -0
- package/dist/core/onnx-llm.d.ts +206 -0
- package/dist/core/onnx-llm.d.ts.map +1 -0
- package/dist/core/onnx-llm.js +430 -0
- package/dist/core/sona-wrapper.d.ts +5 -0
- package/dist/core/sona-wrapper.d.ts.map +1 -1
- package/dist/core/sona-wrapper.js +17 -1
- package/package.json +7 -3
- package/dist/core/core/agentdb-fast.d.ts +0 -148
- package/dist/core/core/agentdb-fast.js +0 -301
- package/dist/core/core/intelligence-engine.d.ts +0 -257
- package/dist/core/core/intelligence-engine.js +0 -1030
- package/dist/core/core/onnx-embedder.d.ts +0 -104
- package/dist/core/core/onnx-embedder.js +0 -410
- package/dist/core/core/parallel-intelligence.d.ts +0 -108
- package/dist/core/core/parallel-intelligence.js +0 -340
- package/dist/core/core/sona-wrapper.d.ts +0 -214
- package/dist/core/core/sona-wrapper.js +0 -258
- package/dist/core/types.d.ts +0 -144
- package/dist/core/types.js +0 -2
|
@@ -1,340 +0,0 @@
|
|
|
1
|
-
"use strict";
|
|
2
|
-
/**
|
|
3
|
-
* Parallel Intelligence - Worker-based acceleration for IntelligenceEngine
|
|
4
|
-
*
|
|
5
|
-
* Provides parallel processing for:
|
|
6
|
-
* - Q-learning batch updates (3-4x faster)
|
|
7
|
-
* - Multi-file pattern matching
|
|
8
|
-
* - Background memory indexing
|
|
9
|
-
* - Parallel similarity search
|
|
10
|
-
* - Multi-file code analysis
|
|
11
|
-
* - Parallel git commit analysis
|
|
12
|
-
*
|
|
13
|
-
* Uses worker_threads for CPU-bound operations, keeping hooks non-blocking.
|
|
14
|
-
*/
|
|
15
|
-
var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
|
|
16
|
-
if (k2 === undefined) k2 = k;
|
|
17
|
-
var desc = Object.getOwnPropertyDescriptor(m, k);
|
|
18
|
-
if (!desc || ("get" in desc ? !m.__esModule : desc.writable || desc.configurable)) {
|
|
19
|
-
desc = { enumerable: true, get: function() { return m[k]; } };
|
|
20
|
-
}
|
|
21
|
-
Object.defineProperty(o, k2, desc);
|
|
22
|
-
}) : (function(o, m, k, k2) {
|
|
23
|
-
if (k2 === undefined) k2 = k;
|
|
24
|
-
o[k2] = m[k];
|
|
25
|
-
}));
|
|
26
|
-
var __setModuleDefault = (this && this.__setModuleDefault) || (Object.create ? (function(o, v) {
|
|
27
|
-
Object.defineProperty(o, "default", { enumerable: true, value: v });
|
|
28
|
-
}) : function(o, v) {
|
|
29
|
-
o["default"] = v;
|
|
30
|
-
});
|
|
31
|
-
var __importStar = (this && this.__importStar) || (function () {
|
|
32
|
-
var ownKeys = function(o) {
|
|
33
|
-
ownKeys = Object.getOwnPropertyNames || function (o) {
|
|
34
|
-
var ar = [];
|
|
35
|
-
for (var k in o) if (Object.prototype.hasOwnProperty.call(o, k)) ar[ar.length] = k;
|
|
36
|
-
return ar;
|
|
37
|
-
};
|
|
38
|
-
return ownKeys(o);
|
|
39
|
-
};
|
|
40
|
-
return function (mod) {
|
|
41
|
-
if (mod && mod.__esModule) return mod;
|
|
42
|
-
var result = {};
|
|
43
|
-
if (mod != null) for (var k = ownKeys(mod), i = 0; i < k.length; i++) if (k[i] !== "default") __createBinding(result, mod, k[i]);
|
|
44
|
-
__setModuleDefault(result, mod);
|
|
45
|
-
return result;
|
|
46
|
-
};
|
|
47
|
-
})();
|
|
48
|
-
Object.defineProperty(exports, "__esModule", { value: true });
|
|
49
|
-
exports.ParallelIntelligence = void 0;
|
|
50
|
-
exports.getParallelIntelligence = getParallelIntelligence;
|
|
51
|
-
exports.initParallelIntelligence = initParallelIntelligence;
|
|
52
|
-
const worker_threads_1 = require("worker_threads");
|
|
53
|
-
const os = __importStar(require("os"));
|
|
54
|
-
// ============================================================================
|
|
55
|
-
// Worker Pool Manager
|
|
56
|
-
// ============================================================================
|
|
57
|
-
class ParallelIntelligence {
|
|
58
|
-
constructor(config = {}) {
|
|
59
|
-
this.workers = [];
|
|
60
|
-
this.taskQueue = [];
|
|
61
|
-
this.busyWorkers = new Set();
|
|
62
|
-
this.initialized = false;
|
|
63
|
-
const isCLI = process.env.RUVECTOR_CLI === '1';
|
|
64
|
-
const isMCP = process.env.MCP_SERVER === '1';
|
|
65
|
-
this.config = {
|
|
66
|
-
numWorkers: config.numWorkers ?? Math.max(1, os.cpus().length - 1),
|
|
67
|
-
enabled: config.enabled ?? (isMCP || (!isCLI && process.env.RUVECTOR_PARALLEL === '1')),
|
|
68
|
-
batchThreshold: config.batchThreshold ?? 4,
|
|
69
|
-
};
|
|
70
|
-
}
|
|
71
|
-
/**
|
|
72
|
-
* Initialize worker pool
|
|
73
|
-
*/
|
|
74
|
-
async init() {
|
|
75
|
-
if (this.initialized || !this.config.enabled)
|
|
76
|
-
return;
|
|
77
|
-
for (let i = 0; i < this.config.numWorkers; i++) {
|
|
78
|
-
const worker = new worker_threads_1.Worker(__filename, {
|
|
79
|
-
workerData: { workerId: i },
|
|
80
|
-
});
|
|
81
|
-
worker.on('message', (result) => {
|
|
82
|
-
this.busyWorkers.delete(worker);
|
|
83
|
-
this.processQueue();
|
|
84
|
-
});
|
|
85
|
-
worker.on('error', (err) => {
|
|
86
|
-
console.error(`Worker ${i} error:`, err);
|
|
87
|
-
this.busyWorkers.delete(worker);
|
|
88
|
-
});
|
|
89
|
-
this.workers.push(worker);
|
|
90
|
-
}
|
|
91
|
-
this.initialized = true;
|
|
92
|
-
console.error(`ParallelIntelligence: ${this.config.numWorkers} workers ready`);
|
|
93
|
-
}
|
|
94
|
-
processQueue() {
|
|
95
|
-
while (this.taskQueue.length > 0 && this.busyWorkers.size < this.workers.length) {
|
|
96
|
-
const availableWorker = this.workers.find(w => !this.busyWorkers.has(w));
|
|
97
|
-
if (!availableWorker)
|
|
98
|
-
break;
|
|
99
|
-
const task = this.taskQueue.shift();
|
|
100
|
-
this.busyWorkers.add(availableWorker);
|
|
101
|
-
availableWorker.postMessage(task.task);
|
|
102
|
-
}
|
|
103
|
-
}
|
|
104
|
-
/**
|
|
105
|
-
* Execute task in worker pool
|
|
106
|
-
*/
|
|
107
|
-
async executeInWorker(task) {
|
|
108
|
-
if (!this.initialized || !this.config.enabled) {
|
|
109
|
-
throw new Error('ParallelIntelligence not initialized');
|
|
110
|
-
}
|
|
111
|
-
return new Promise((resolve, reject) => {
|
|
112
|
-
const availableWorker = this.workers.find(w => !this.busyWorkers.has(w));
|
|
113
|
-
if (availableWorker) {
|
|
114
|
-
this.busyWorkers.add(availableWorker);
|
|
115
|
-
const handler = (result) => {
|
|
116
|
-
this.busyWorkers.delete(availableWorker);
|
|
117
|
-
availableWorker.off('message', handler);
|
|
118
|
-
if (result.error) {
|
|
119
|
-
reject(new Error(result.error));
|
|
120
|
-
}
|
|
121
|
-
else {
|
|
122
|
-
resolve(result.data);
|
|
123
|
-
}
|
|
124
|
-
};
|
|
125
|
-
availableWorker.on('message', handler);
|
|
126
|
-
availableWorker.postMessage(task);
|
|
127
|
-
}
|
|
128
|
-
else {
|
|
129
|
-
this.taskQueue.push({ task, resolve, reject });
|
|
130
|
-
}
|
|
131
|
-
});
|
|
132
|
-
}
|
|
133
|
-
// =========================================================================
|
|
134
|
-
// Parallel Operations
|
|
135
|
-
// =========================================================================
|
|
136
|
-
/**
|
|
137
|
-
* Batch Q-learning episode recording (3-4x faster)
|
|
138
|
-
*/
|
|
139
|
-
async recordEpisodesBatch(episodes) {
|
|
140
|
-
if (episodes.length < this.config.batchThreshold || !this.config.enabled) {
|
|
141
|
-
// Fall back to sequential
|
|
142
|
-
return;
|
|
143
|
-
}
|
|
144
|
-
// Split into chunks for workers
|
|
145
|
-
const chunkSize = Math.ceil(episodes.length / this.config.numWorkers);
|
|
146
|
-
const chunks = [];
|
|
147
|
-
for (let i = 0; i < episodes.length; i += chunkSize) {
|
|
148
|
-
chunks.push(episodes.slice(i, i + chunkSize));
|
|
149
|
-
}
|
|
150
|
-
await Promise.all(chunks.map(chunk => this.executeInWorker({ type: 'recordEpisodes', episodes: chunk })));
|
|
151
|
-
}
|
|
152
|
-
/**
|
|
153
|
-
* Multi-file pattern matching (parallel pretrain)
|
|
154
|
-
*/
|
|
155
|
-
async matchPatternsParallel(files) {
|
|
156
|
-
if (files.length < this.config.batchThreshold || !this.config.enabled) {
|
|
157
|
-
return [];
|
|
158
|
-
}
|
|
159
|
-
const chunkSize = Math.ceil(files.length / this.config.numWorkers);
|
|
160
|
-
const chunks = [];
|
|
161
|
-
for (let i = 0; i < files.length; i += chunkSize) {
|
|
162
|
-
chunks.push(files.slice(i, i + chunkSize));
|
|
163
|
-
}
|
|
164
|
-
const results = await Promise.all(chunks.map(chunk => this.executeInWorker({ type: 'matchPatterns', files: chunk })));
|
|
165
|
-
return results.flat();
|
|
166
|
-
}
|
|
167
|
-
/**
|
|
168
|
-
* Background memory indexing (non-blocking)
|
|
169
|
-
*/
|
|
170
|
-
async indexMemoriesBackground(memories) {
|
|
171
|
-
if (memories.length === 0 || !this.config.enabled)
|
|
172
|
-
return;
|
|
173
|
-
// Fire and forget - non-blocking
|
|
174
|
-
this.executeInWorker({ type: 'indexMemories', memories }).catch(() => { });
|
|
175
|
-
}
|
|
176
|
-
/**
|
|
177
|
-
* Parallel similarity search with sharding
|
|
178
|
-
*/
|
|
179
|
-
async searchParallel(query, topK = 5) {
|
|
180
|
-
if (!this.config.enabled)
|
|
181
|
-
return [];
|
|
182
|
-
// Each worker searches its shard
|
|
183
|
-
const shardResults = await Promise.all(this.workers.map((_, i) => this.executeInWorker({
|
|
184
|
-
type: 'search',
|
|
185
|
-
query,
|
|
186
|
-
topK,
|
|
187
|
-
shardId: i,
|
|
188
|
-
})));
|
|
189
|
-
// Merge and sort results
|
|
190
|
-
return shardResults
|
|
191
|
-
.flat()
|
|
192
|
-
.sort((a, b) => b.score - a.score)
|
|
193
|
-
.slice(0, topK);
|
|
194
|
-
}
|
|
195
|
-
/**
|
|
196
|
-
* Multi-file AST analysis for routing
|
|
197
|
-
*/
|
|
198
|
-
async analyzeFilesParallel(files) {
|
|
199
|
-
if (files.length < this.config.batchThreshold || !this.config.enabled) {
|
|
200
|
-
return new Map();
|
|
201
|
-
}
|
|
202
|
-
const chunkSize = Math.ceil(files.length / this.config.numWorkers);
|
|
203
|
-
const chunks = [];
|
|
204
|
-
for (let i = 0; i < files.length; i += chunkSize) {
|
|
205
|
-
chunks.push(files.slice(i, i + chunkSize));
|
|
206
|
-
}
|
|
207
|
-
const results = await Promise.all(chunks.map(chunk => this.executeInWorker({
|
|
208
|
-
type: 'analyzeFiles',
|
|
209
|
-
files: chunk,
|
|
210
|
-
})));
|
|
211
|
-
return new Map(results.flat());
|
|
212
|
-
}
|
|
213
|
-
/**
|
|
214
|
-
* Parallel git commit analysis for co-edit detection
|
|
215
|
-
*/
|
|
216
|
-
async analyzeCommitsParallel(commits) {
|
|
217
|
-
if (commits.length < this.config.batchThreshold || !this.config.enabled) {
|
|
218
|
-
return [];
|
|
219
|
-
}
|
|
220
|
-
const chunkSize = Math.ceil(commits.length / this.config.numWorkers);
|
|
221
|
-
const chunks = [];
|
|
222
|
-
for (let i = 0; i < commits.length; i += chunkSize) {
|
|
223
|
-
chunks.push(commits.slice(i, i + chunkSize));
|
|
224
|
-
}
|
|
225
|
-
const results = await Promise.all(chunks.map(chunk => this.executeInWorker({ type: 'analyzeCommits', commits: chunk })));
|
|
226
|
-
return results.flat();
|
|
227
|
-
}
|
|
228
|
-
/**
|
|
229
|
-
* Get worker pool stats
|
|
230
|
-
*/
|
|
231
|
-
getStats() {
|
|
232
|
-
return {
|
|
233
|
-
enabled: this.config.enabled,
|
|
234
|
-
workers: this.workers.length,
|
|
235
|
-
busy: this.busyWorkers.size,
|
|
236
|
-
queued: this.taskQueue.length,
|
|
237
|
-
};
|
|
238
|
-
}
|
|
239
|
-
/**
|
|
240
|
-
* Shutdown worker pool
|
|
241
|
-
*/
|
|
242
|
-
async shutdown() {
|
|
243
|
-
await Promise.all(this.workers.map(w => w.terminate()));
|
|
244
|
-
this.workers = [];
|
|
245
|
-
this.busyWorkers.clear();
|
|
246
|
-
this.taskQueue = [];
|
|
247
|
-
this.initialized = false;
|
|
248
|
-
}
|
|
249
|
-
}
|
|
250
|
-
exports.ParallelIntelligence = ParallelIntelligence;
|
|
251
|
-
// ============================================================================
|
|
252
|
-
// Worker Thread Code
|
|
253
|
-
// ============================================================================
|
|
254
|
-
if (!worker_threads_1.isMainThread && worker_threads_1.parentPort) {
|
|
255
|
-
// This code runs in worker threads
|
|
256
|
-
const { workerId } = worker_threads_1.workerData;
|
|
257
|
-
worker_threads_1.parentPort.on('message', async (task) => {
|
|
258
|
-
try {
|
|
259
|
-
let result;
|
|
260
|
-
switch (task.type) {
|
|
261
|
-
case 'recordEpisodes':
|
|
262
|
-
// Process episode batch
|
|
263
|
-
result = await processEpisodes(task.episodes);
|
|
264
|
-
break;
|
|
265
|
-
case 'matchPatterns':
|
|
266
|
-
// Match patterns in files
|
|
267
|
-
result = await matchPatterns(task.files);
|
|
268
|
-
break;
|
|
269
|
-
case 'indexMemories':
|
|
270
|
-
// Index memories
|
|
271
|
-
result = await indexMemories(task.memories);
|
|
272
|
-
break;
|
|
273
|
-
case 'search':
|
|
274
|
-
// Search shard
|
|
275
|
-
result = await searchShard(task.query, task.topK, task.shardId);
|
|
276
|
-
break;
|
|
277
|
-
case 'analyzeFiles':
|
|
278
|
-
// Analyze file ASTs
|
|
279
|
-
result = await analyzeFiles(task.files);
|
|
280
|
-
break;
|
|
281
|
-
case 'analyzeCommits':
|
|
282
|
-
// Analyze git commits
|
|
283
|
-
result = await analyzeCommits(task.commits);
|
|
284
|
-
break;
|
|
285
|
-
default:
|
|
286
|
-
throw new Error(`Unknown task type: ${task.type}`);
|
|
287
|
-
}
|
|
288
|
-
worker_threads_1.parentPort.postMessage({ data: result });
|
|
289
|
-
}
|
|
290
|
-
catch (error) {
|
|
291
|
-
worker_threads_1.parentPort.postMessage({ error: error.message });
|
|
292
|
-
}
|
|
293
|
-
});
|
|
294
|
-
// Worker task implementations
|
|
295
|
-
async function processEpisodes(episodes) {
|
|
296
|
-
// Embed and process episodes
|
|
297
|
-
// In a real implementation, this would use the embedder and update Q-values
|
|
298
|
-
return episodes.length;
|
|
299
|
-
}
|
|
300
|
-
async function matchPatterns(files) {
|
|
301
|
-
// Match patterns in files
|
|
302
|
-
// Would read files and extract patterns
|
|
303
|
-
return files.map(file => ({
|
|
304
|
-
file,
|
|
305
|
-
patterns: [],
|
|
306
|
-
}));
|
|
307
|
-
}
|
|
308
|
-
async function indexMemories(memories) {
|
|
309
|
-
// Index memories in background
|
|
310
|
-
return memories.length;
|
|
311
|
-
}
|
|
312
|
-
async function searchShard(query, topK, shardId) {
|
|
313
|
-
// Search this worker's shard
|
|
314
|
-
return [];
|
|
315
|
-
}
|
|
316
|
-
async function analyzeFiles(files) {
|
|
317
|
-
// Analyze file ASTs
|
|
318
|
-
return files.map(f => [f, { agent: 'coder', confidence: 0.5 }]);
|
|
319
|
-
}
|
|
320
|
-
async function analyzeCommits(commits) {
|
|
321
|
-
// Analyze git commits for co-edit patterns
|
|
322
|
-
return [];
|
|
323
|
-
}
|
|
324
|
-
}
|
|
325
|
-
// ============================================================================
|
|
326
|
-
// Singleton for easy access
|
|
327
|
-
// ============================================================================
|
|
328
|
-
let instance = null;
|
|
329
|
-
function getParallelIntelligence(config) {
|
|
330
|
-
if (!instance) {
|
|
331
|
-
instance = new ParallelIntelligence(config);
|
|
332
|
-
}
|
|
333
|
-
return instance;
|
|
334
|
-
}
|
|
335
|
-
async function initParallelIntelligence(config) {
|
|
336
|
-
const pi = getParallelIntelligence(config);
|
|
337
|
-
await pi.init();
|
|
338
|
-
return pi;
|
|
339
|
-
}
|
|
340
|
-
exports.default = ParallelIntelligence;
|
|
@@ -1,214 +0,0 @@
|
|
|
1
|
-
/**
|
|
2
|
-
* SONA Wrapper - Self-Optimizing Neural Architecture
|
|
3
|
-
*
|
|
4
|
-
* Provides a safe, flexible interface to @ruvector/sona with:
|
|
5
|
-
* - Automatic array type conversion (Array <-> Float64Array)
|
|
6
|
-
* - Graceful handling when sona is not installed
|
|
7
|
-
* - TypeScript types for all APIs
|
|
8
|
-
*
|
|
9
|
-
* SONA Features:
|
|
10
|
-
* - Micro-LoRA: Ultra-fast rank-1/2 adaptations (~0.1ms)
|
|
11
|
-
* - Base-LoRA: Deeper adaptations for complex patterns
|
|
12
|
-
* - EWC++: Elastic Weight Consolidation to prevent catastrophic forgetting
|
|
13
|
-
* - ReasoningBank: Pattern storage and retrieval
|
|
14
|
-
* - Trajectory tracking: Record and learn from execution paths
|
|
15
|
-
*/
|
|
16
|
-
/** Array input type - accepts both regular arrays and typed arrays */
|
|
17
|
-
export type ArrayInput = number[] | Float32Array | Float64Array;
|
|
18
|
-
/** SONA configuration options */
|
|
19
|
-
export interface SonaConfig {
|
|
20
|
-
/** Hidden dimension size (required) */
|
|
21
|
-
hiddenDim: number;
|
|
22
|
-
/** Embedding dimension (defaults to hiddenDim) */
|
|
23
|
-
embeddingDim?: number;
|
|
24
|
-
/** Micro-LoRA rank (1-2, default: 1) */
|
|
25
|
-
microLoraRank?: number;
|
|
26
|
-
/** Base LoRA rank (default: 8) */
|
|
27
|
-
baseLoraRank?: number;
|
|
28
|
-
/** Micro-LoRA learning rate (default: 0.001) */
|
|
29
|
-
microLoraLr?: number;
|
|
30
|
-
/** Base LoRA learning rate (default: 0.0001) */
|
|
31
|
-
baseLoraLr?: number;
|
|
32
|
-
/** EWC lambda regularization (default: 1000.0) */
|
|
33
|
-
ewcLambda?: number;
|
|
34
|
-
/** Number of pattern clusters (default: 50) */
|
|
35
|
-
patternClusters?: number;
|
|
36
|
-
/** Trajectory buffer capacity (default: 10000) */
|
|
37
|
-
trajectoryCapacity?: number;
|
|
38
|
-
/** Background learning interval in ms (default: 3600000 = 1 hour) */
|
|
39
|
-
backgroundIntervalMs?: number;
|
|
40
|
-
/** Quality threshold for learning (default: 0.5) */
|
|
41
|
-
qualityThreshold?: number;
|
|
42
|
-
/** Enable SIMD optimizations (default: true) */
|
|
43
|
-
enableSimd?: boolean;
|
|
44
|
-
}
|
|
45
|
-
/** Learned pattern from ReasoningBank */
|
|
46
|
-
export interface LearnedPattern {
|
|
47
|
-
/** Pattern identifier */
|
|
48
|
-
id: string;
|
|
49
|
-
/** Cluster centroid embedding */
|
|
50
|
-
centroid: number[];
|
|
51
|
-
/** Number of trajectories in cluster */
|
|
52
|
-
clusterSize: number;
|
|
53
|
-
/** Total weight of trajectories */
|
|
54
|
-
totalWeight: number;
|
|
55
|
-
/** Average quality of member trajectories */
|
|
56
|
-
avgQuality: number;
|
|
57
|
-
/** Creation timestamp */
|
|
58
|
-
createdAt: string;
|
|
59
|
-
/** Last access timestamp */
|
|
60
|
-
lastAccessed: string;
|
|
61
|
-
/** Total access count */
|
|
62
|
-
accessCount: number;
|
|
63
|
-
/** Pattern type */
|
|
64
|
-
patternType: string;
|
|
65
|
-
}
|
|
66
|
-
/** SONA engine statistics */
|
|
67
|
-
export interface SonaStats {
|
|
68
|
-
trajectoriesRecorded: number;
|
|
69
|
-
patternsLearned: number;
|
|
70
|
-
microLoraUpdates: number;
|
|
71
|
-
baseLoraUpdates: number;
|
|
72
|
-
ewcConsolidations: number;
|
|
73
|
-
avgLearningTimeMs: number;
|
|
74
|
-
}
|
|
75
|
-
/** Check if sona is available */
|
|
76
|
-
export declare function isSonaAvailable(): boolean;
|
|
77
|
-
/**
|
|
78
|
-
* SONA Engine - Self-Optimizing Neural Architecture
|
|
79
|
-
*
|
|
80
|
-
* Provides runtime-adaptive learning with:
|
|
81
|
-
* - Micro-LoRA for instant adaptations
|
|
82
|
-
* - Base-LoRA for deeper learning
|
|
83
|
-
* - EWC++ for preventing forgetting
|
|
84
|
-
* - ReasoningBank for pattern storage
|
|
85
|
-
*
|
|
86
|
-
* @example
|
|
87
|
-
* ```typescript
|
|
88
|
-
* import { Sona } from 'ruvector';
|
|
89
|
-
*
|
|
90
|
-
* // Create engine with hidden dimension
|
|
91
|
-
* const engine = new Sona.Engine(256);
|
|
92
|
-
*
|
|
93
|
-
* // Or with custom config
|
|
94
|
-
* const engine = Sona.Engine.withConfig({
|
|
95
|
-
* hiddenDim: 256,
|
|
96
|
-
* microLoraRank: 2,
|
|
97
|
-
* patternClusters: 100
|
|
98
|
-
* });
|
|
99
|
-
*
|
|
100
|
-
* // Record a trajectory
|
|
101
|
-
* const trajId = engine.beginTrajectory([0.1, 0.2, ...]);
|
|
102
|
-
* engine.addStep(trajId, activations, attentionWeights, 0.8);
|
|
103
|
-
* engine.endTrajectory(trajId, 0.9);
|
|
104
|
-
*
|
|
105
|
-
* // Apply learned adaptations
|
|
106
|
-
* const adapted = engine.applyMicroLora(input);
|
|
107
|
-
* ```
|
|
108
|
-
*/
|
|
109
|
-
export declare class SonaEngine {
|
|
110
|
-
private _native;
|
|
111
|
-
/**
|
|
112
|
-
* Create a new SONA engine
|
|
113
|
-
* @param hiddenDim Hidden dimension size (e.g., 256, 512, 768)
|
|
114
|
-
*/
|
|
115
|
-
constructor(hiddenDim: number);
|
|
116
|
-
/**
|
|
117
|
-
* Create engine with custom configuration
|
|
118
|
-
* @param config SONA configuration options
|
|
119
|
-
*/
|
|
120
|
-
static withConfig(config: SonaConfig): SonaEngine;
|
|
121
|
-
/**
|
|
122
|
-
* Begin recording a new trajectory
|
|
123
|
-
* @param queryEmbedding Initial query embedding
|
|
124
|
-
* @returns Trajectory ID for subsequent operations
|
|
125
|
-
*/
|
|
126
|
-
beginTrajectory(queryEmbedding: ArrayInput): number;
|
|
127
|
-
/**
|
|
128
|
-
* Add a step to an active trajectory
|
|
129
|
-
* @param trajectoryId Trajectory ID from beginTrajectory
|
|
130
|
-
* @param activations Layer activations
|
|
131
|
-
* @param attentionWeights Attention weights
|
|
132
|
-
* @param reward Reward signal for this step (0.0 - 1.0)
|
|
133
|
-
*/
|
|
134
|
-
addStep(trajectoryId: number, activations: ArrayInput, attentionWeights: ArrayInput, reward: number): void;
|
|
135
|
-
/**
|
|
136
|
-
* Alias for addStep for API compatibility
|
|
137
|
-
*/
|
|
138
|
-
addTrajectoryStep(trajectoryId: number, activations: ArrayInput, attentionWeights: ArrayInput, reward: number): void;
|
|
139
|
-
/**
|
|
140
|
-
* Set the model route for a trajectory
|
|
141
|
-
* @param trajectoryId Trajectory ID
|
|
142
|
-
* @param route Model route identifier (e.g., "gpt-4", "claude-3")
|
|
143
|
-
*/
|
|
144
|
-
setRoute(trajectoryId: number, route: string): void;
|
|
145
|
-
/**
|
|
146
|
-
* Add context to a trajectory
|
|
147
|
-
* @param trajectoryId Trajectory ID
|
|
148
|
-
* @param contextId Context identifier
|
|
149
|
-
*/
|
|
150
|
-
addContext(trajectoryId: number, contextId: string): void;
|
|
151
|
-
/**
|
|
152
|
-
* Complete a trajectory and submit for learning
|
|
153
|
-
* @param trajectoryId Trajectory ID
|
|
154
|
-
* @param quality Final quality score (0.0 - 1.0)
|
|
155
|
-
*/
|
|
156
|
-
endTrajectory(trajectoryId: number, quality: number): void;
|
|
157
|
-
/**
|
|
158
|
-
* Apply micro-LoRA transformation (ultra-fast, ~0.1ms)
|
|
159
|
-
* @param input Input vector
|
|
160
|
-
* @returns Transformed output vector
|
|
161
|
-
*/
|
|
162
|
-
applyMicroLora(input: ArrayInput): number[];
|
|
163
|
-
/**
|
|
164
|
-
* Apply base-LoRA transformation to a specific layer
|
|
165
|
-
* @param layerIdx Layer index
|
|
166
|
-
* @param input Input vector
|
|
167
|
-
* @returns Transformed output vector
|
|
168
|
-
*/
|
|
169
|
-
applyBaseLora(layerIdx: number, input: ArrayInput): number[];
|
|
170
|
-
/**
|
|
171
|
-
* Run background learning cycle if due
|
|
172
|
-
* Call this periodically (e.g., every few seconds)
|
|
173
|
-
* @returns Status message if learning occurred, null otherwise
|
|
174
|
-
*/
|
|
175
|
-
tick(): string | null;
|
|
176
|
-
/**
|
|
177
|
-
* Force immediate background learning cycle
|
|
178
|
-
* @returns Status message with learning results
|
|
179
|
-
*/
|
|
180
|
-
forceLearn(): string;
|
|
181
|
-
/**
|
|
182
|
-
* Flush pending instant loop updates
|
|
183
|
-
*/
|
|
184
|
-
flush(): void;
|
|
185
|
-
/**
|
|
186
|
-
* Find similar learned patterns to a query
|
|
187
|
-
* @param queryEmbedding Query embedding
|
|
188
|
-
* @param k Number of patterns to return
|
|
189
|
-
* @returns Array of similar patterns
|
|
190
|
-
*/
|
|
191
|
-
findPatterns(queryEmbedding: ArrayInput, k: number): LearnedPattern[];
|
|
192
|
-
/**
|
|
193
|
-
* Get engine statistics
|
|
194
|
-
* @returns Statistics object
|
|
195
|
-
*/
|
|
196
|
-
getStats(): SonaStats;
|
|
197
|
-
/**
|
|
198
|
-
* Enable or disable the engine
|
|
199
|
-
* @param enabled Whether to enable
|
|
200
|
-
*/
|
|
201
|
-
setEnabled(enabled: boolean): void;
|
|
202
|
-
/**
|
|
203
|
-
* Check if engine is enabled
|
|
204
|
-
*/
|
|
205
|
-
isEnabled(): boolean;
|
|
206
|
-
}
|
|
207
|
-
/**
|
|
208
|
-
* SONA namespace with all exports
|
|
209
|
-
*/
|
|
210
|
-
export declare const Sona: {
|
|
211
|
-
Engine: typeof SonaEngine;
|
|
212
|
-
isAvailable: typeof isSonaAvailable;
|
|
213
|
-
};
|
|
214
|
-
export default Sona;
|