@intelliweave/embedded 2.0.72-beta.8 → 2.1.73

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -60,6 +60,8 @@ declare class TokenWindowGroup<DataType> {
60
60
  recalculateTokens(): void;
61
61
  /** Add an item to the group */
62
62
  add(item: string | TokenWindowGroupItemParams<DataType>): TokenWindowGroupItem<DataType>;
63
+ /** Manually remove an item */
64
+ remove(itemId: string): boolean;
63
65
  /** Get all items as a string */
64
66
  getAllAsString(): string;
65
67
  /** Get all items. Doesn't return disabled items. */
@@ -176,6 +178,17 @@ declare class MCPKnowledgeClient {
176
178
  searchToolName?: string;
177
179
  /** Keep search function available for the AI to use. */
178
180
  searchToolVisible?: boolean;
181
+ /** Use the IntelliWeave proxy */
182
+ proxy?: {
183
+ /** If true, will send requests via the IntelliWeave MCP proxy */
184
+ enabled?: boolean;
185
+ /** The URL of the proxy server, defaults to the standard IntelliWeave proxy */
186
+ url?: string;
187
+ /** IntelliWeave API key */
188
+ apiKey?: string;
189
+ };
190
+ /** Pass extra headers to the MCP server */
191
+ headers?: Record<string, string>;
179
192
  };
180
193
  /** Constructor */
181
194
  constructor(config: MCPKnowledgeClient['config']);
@@ -186,16 +199,10 @@ declare class MCPKnowledgeClient {
186
199
  method: string;
187
200
  params?: {
188
201
  [x: string]: unknown;
189
- task?: {
190
- [x: string]: unknown;
191
- ttl?: number | null | undefined;
192
- pollInterval?: number | undefined;
193
- } | undefined;
194
202
  _meta?: {
195
203
  [x: string]: unknown;
196
204
  progressToken?: string | number | undefined;
197
205
  "io.modelcontextprotocol/related-task"?: {
198
- [x: string]: unknown;
199
206
  taskId: string;
200
207
  } | undefined;
201
208
  } | undefined;
@@ -206,8 +213,8 @@ declare class MCPKnowledgeClient {
206
213
  [x: string]: unknown;
207
214
  _meta?: {
208
215
  [x: string]: unknown;
216
+ progressToken?: string | number | undefined;
209
217
  "io.modelcontextprotocol/related-task"?: {
210
- [x: string]: unknown;
211
218
  taskId: string;
212
219
  } | undefined;
213
220
  } | undefined;
@@ -216,8 +223,8 @@ declare class MCPKnowledgeClient {
216
223
  [x: string]: unknown;
217
224
  _meta?: {
218
225
  [x: string]: unknown;
226
+ progressToken?: string | number | undefined;
219
227
  "io.modelcontextprotocol/related-task"?: {
220
- [x: string]: unknown;
221
228
  taskId: string;
222
229
  } | undefined;
223
230
  } | undefined;
@@ -226,16 +233,10 @@ declare class MCPKnowledgeClient {
226
233
  method: string;
227
234
  params?: {
228
235
  [x: string]: unknown;
229
- task?: {
230
- [x: string]: unknown;
231
- ttl?: number | null | undefined;
232
- pollInterval?: number | undefined;
233
- } | undefined;
234
236
  _meta?: {
235
237
  [x: string]: unknown;
236
238
  progressToken?: string | number | undefined;
237
239
  "io.modelcontextprotocol/related-task"?: {
238
- [x: string]: unknown;
239
240
  taskId: string;
240
241
  } | undefined;
241
242
  } | undefined;
@@ -246,8 +247,8 @@ declare class MCPKnowledgeClient {
246
247
  [x: string]: unknown;
247
248
  _meta?: {
248
249
  [x: string]: unknown;
250
+ progressToken?: string | number | undefined;
249
251
  "io.modelcontextprotocol/related-task"?: {
250
- [x: string]: unknown;
251
252
  taskId: string;
252
253
  } | undefined;
253
254
  } | undefined;
@@ -256,8 +257,8 @@ declare class MCPKnowledgeClient {
256
257
  [x: string]: unknown;
257
258
  _meta?: {
258
259
  [x: string]: unknown;
260
+ progressToken?: string | number | undefined;
259
261
  "io.modelcontextprotocol/related-task"?: {
260
- [x: string]: unknown;
261
262
  taskId: string;
262
263
  } | undefined;
263
264
  } | undefined;
@@ -462,6 +463,8 @@ declare class KnowledgeBase {
462
463
  allowWindowSources: boolean;
463
464
  /** If true, allows using knowledge specified in the global configuration object */
464
465
  allowGlobalConfigSources: boolean;
466
+ /** If true, allows the AI to search the knowledge base. If false, essentially disables RAG lookup. */
467
+ allowRagSearch: boolean;
465
468
  /** Constructor */
466
469
  constructor(ai: IntelliWeave);
467
470
  /** Ensures the internal knowledge is set correctly */
@@ -490,7 +493,7 @@ declare class KnowledgeBase {
490
493
  /** Create and register an external knowledge base source from a URL */
491
494
  registerSourceFromURL(url: string, id?: string): void;
492
495
  /** Clone this instance */
493
- clone(): KnowledgeBase;
496
+ clone(newIW: IntelliWeave): KnowledgeBase;
494
497
  /** Registers an MCP server as a knowledge base source */
495
498
  registerMCPSource(config: MCPKnowledgeClient['config']): MCPKnowledgeClient;
496
499
  }
@@ -758,7 +761,7 @@ declare class IntelliWeaveTranscriptionNode extends VoiceChunkOutputNode {
758
761
  static debugExportWav: boolean;
759
762
  /** Server address for transcription */
760
763
  apiAddress: string;
761
- /** OpenAI API key */
764
+ /** IntelliWeave API key */
762
765
  apiKey: string;
763
766
  /** WebSocket connection */
764
767
  private ws?;
@@ -778,6 +781,32 @@ declare class IntelliWeaveTranscriptionNode extends VoiceChunkOutputNode {
778
781
  onSocketClose(): void;
779
782
  }
780
783
 
784
+ /**
785
+ * This AudioNode uses ElevenLabs to transcribe spoken speech to text.
786
+ *
787
+ * - event `transcription` - Fired when a transcription is ready. `text` contains the transcribed text.
788
+ */
789
+ declare class ElevenLabsTranscriptionNode extends VoiceChunkOutputNode {
790
+ /** ElevenLabs API key */
791
+ apiKey: string;
792
+ /** ElevenLabs stream connection */
793
+ private connection?;
794
+ /** True if currently transcribing */
795
+ isTranscribing: boolean;
796
+ /** WebSocket shutdown timer */
797
+ private shutdownTimer?;
798
+ /** Constructor */
799
+ constructor(audioContext: AudioContext, apiKey: string);
800
+ /** Called when a voice chunk is received */
801
+ onVoiceChunk(buffer: Float32Array): Promise<void>;
802
+ /** Start reading the stream */
803
+ private startReading;
804
+ /** Called when the voice recording ends */
805
+ onVoiceEnd(buffers: Float32Array[]): Promise<void>;
806
+ /** Called when a transcription is ready */
807
+ onVoiceTranscription(text: string): void;
808
+ }
809
+
781
810
  /**
782
811
  * Handles speech recognition from the microphone
783
812
  *
@@ -803,7 +832,7 @@ declare class WebWeaverSpeechRecognition extends EventTarget {
803
832
  /** Returns true if speech recognition is supported by this persona and browser */
804
833
  get isSupported(): boolean;
805
834
  /** Currently active voice detection node */
806
- voiceDetection?: IntelliWeaveTranscriptionNode | OpenAITranscriptionNode;
835
+ voiceDetection?: IntelliWeaveTranscriptionNode | OpenAITranscriptionNode | ElevenLabsTranscriptionNode;
807
836
  /** Constructor */
808
837
  constructor(ai: IntelliWeave);
809
838
  private _skipEvents;
@@ -1054,6 +1083,8 @@ interface SubAgentConfig {
1054
1083
  usageInstructions?: string;
1055
1084
  /** If true, will remove all Persona knowledge entries */
1056
1085
  clearExistingKnowledge?: boolean;
1086
+ /** Disable RAG search for subagents. If true, only KB entries with isContext=true will be used. */
1087
+ disableRagSearch?: boolean;
1057
1088
  /** Extra knowledge base sources for the sub-agent */
1058
1089
  knowledge?: KnowledgeFetcher;
1059
1090
  /** Optional extra configuration for the subagent instance */
@@ -1219,6 +1250,8 @@ declare class IntelliWeave extends EventTarget {
1219
1250
  private _lastSystemMsg;
1220
1251
  /** Get the system message prefix, before the KB entries are added */
1221
1252
  getContextPrefix(): Promise<string>;
1253
+ /** KB items added in the last run */
1254
+ private lastKBItems;
1222
1255
  /** Get system message to send to the AI */
1223
1256
  onBeforeMessageProcessing(): Promise<void>;
1224
1257
  /** @private Process incoming message(s) from the AI. Can be used to respond to encoded actions in the text response. */