mem0ai 2.1.13 → 2.1.15

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -363,7 +363,11 @@ declare class Memory {
363
363
  private apiVersion;
364
364
  private graphMemory?;
365
365
  private enableGraph;
366
+ telemetryId: string;
366
367
  constructor(config?: Partial<MemoryConfig>);
368
+ private _initializeTelemetry;
369
+ private _getTelemetryId;
370
+ private _captureEvent;
367
371
  static fromConfig(configDict: Record<string, any>): Memory;
368
372
  add(messages: string | Message[], config: AddMemoryOptions): Promise<SearchResult>;
369
373
  private addToVectorStore;
@@ -409,6 +413,14 @@ declare class OllamaEmbedder implements Embedder {
409
413
  private ensureModelExists;
410
414
  }
411
415
 
416
+ declare class GoogleEmbedder implements Embedder {
417
+ private google;
418
+ private model;
419
+ constructor(config: EmbeddingConfig);
420
+ embed(text: string): Promise<number[]>;
421
+ embedBatch(texts: string[]): Promise<number[][]>;
422
+ }
423
+
412
424
  interface LLMResponse {
413
425
  content: string;
414
426
  role: string;
@@ -437,6 +449,16 @@ declare class OpenAILLM implements LLM {
437
449
  generateChat(messages: Message[]): Promise<LLMResponse>;
438
450
  }
439
451
 
452
+ declare class GoogleLLM implements LLM {
453
+ private google;
454
+ private model;
455
+ constructor(config: LLMConfig);
456
+ generateResponse(messages: Message[], responseFormat?: {
457
+ type: string;
458
+ }, tools?: any[]): Promise<string | LLMResponse>;
459
+ generateChat(messages: Message[]): Promise<LLMResponse>;
460
+ }
461
+
440
462
  declare class OpenAIStructuredLLM implements LLM {
441
463
  private openai;
442
464
  private model;
@@ -487,6 +509,9 @@ interface VectorStore {
487
509
  delete(vectorId: string): Promise<void>;
488
510
  deleteCol(): Promise<void>;
489
511
  list(filters?: SearchFilters, limit?: number): Promise<[VectorStoreResult[], number]>;
512
+ getUserId(): Promise<string>;
513
+ setUserId(userId: string): Promise<void>;
514
+ initialize(): Promise<void>;
490
515
  }
491
516
 
492
517
  declare class MemoryVectorStore implements VectorStore {
@@ -507,6 +532,9 @@ declare class MemoryVectorStore implements VectorStore {
507
532
  delete(vectorId: string): Promise<void>;
508
533
  deleteCol(): Promise<void>;
509
534
  list(filters?: SearchFilters, limit?: number): Promise<[VectorStoreResult[], number]>;
535
+ getUserId(): Promise<string>;
536
+ setUserId(userId: string): Promise<void>;
537
+ initialize(): Promise<void>;
510
538
  }
511
539
 
512
540
  interface QdrantConfig extends VectorStoreConfig {
@@ -519,12 +547,13 @@ interface QdrantConfig extends VectorStoreConfig {
519
547
  onDisk?: boolean;
520
548
  collectionName: string;
521
549
  embeddingModelDims: number;
550
+ dimension?: number;
522
551
  }
523
552
  declare class Qdrant implements VectorStore {
524
553
  private client;
525
554
  private readonly collectionName;
555
+ private dimension;
526
556
  constructor(config: QdrantConfig);
527
- private createCol;
528
557
  private createFilter;
529
558
  insert(vectors: number[][], ids: string[], payloads: Record<string, any>[]): Promise<void>;
530
559
  search(query: number[], limit?: number, filters?: SearchFilters): Promise<VectorStoreResult[]>;
@@ -533,6 +562,10 @@ declare class Qdrant implements VectorStore {
533
562
  delete(vectorId: string): Promise<void>;
534
563
  deleteCol(): Promise<void>;
535
564
  list(filters?: SearchFilters, limit?: number): Promise<[VectorStoreResult[], number]>;
565
+ private generateUUID;
566
+ getUserId(): Promise<string>;
567
+ setUserId(userId: string): Promise<void>;
568
+ initialize(): Promise<void>;
536
569
  }
537
570
 
538
571
  interface RedisConfig extends VectorStoreConfig {
@@ -548,8 +581,8 @@ declare class RedisDB implements VectorStore {
548
581
  private readonly indexPrefix;
549
582
  private readonly schema;
550
583
  constructor(config: RedisConfig);
551
- private initialize;
552
584
  private createIndex;
585
+ initialize(): Promise<void>;
553
586
  insert(vectors: number[][], ids: string[], payloads: Record<string, any>[]): Promise<void>;
554
587
  search(query: number[], limit?: number, filters?: SearchFilters): Promise<VectorStoreResult[]>;
555
588
  get(vectorId: string): Promise<VectorStoreResult | null>;
@@ -558,6 +591,8 @@ declare class RedisDB implements VectorStore {
558
591
  deleteCol(): Promise<void>;
559
592
  list(filters?: SearchFilters, limit?: number): Promise<[VectorStoreResult[], number]>;
560
593
  close(): Promise<void>;
594
+ getUserId(): Promise<string>;
595
+ setUserId(userId: string): Promise<void>;
561
596
  }
562
597
 
563
598
  interface HistoryManager {
@@ -580,4 +615,4 @@ declare class HistoryManagerFactory {
580
615
  static create(provider: string, config: HistoryStoreConfig): HistoryManager;
581
616
  }
582
617
 
583
- export { type AddMemoryOptions, AnthropicLLM, type DeleteAllMemoryOptions, type Embedder, EmbedderFactory, type EmbeddingConfig, type Entity, type GetAllMemoryOptions, type GraphStoreConfig, GroqLLM, HistoryManagerFactory, type HistoryStoreConfig, type LLM, type LLMConfig, LLMFactory, type LLMResponse, Memory, type MemoryConfig, MemoryConfigSchema, type MemoryItem, MemoryVectorStore, type Message, type MultiModalMessages, type Neo4jConfig, OllamaEmbedder, OllamaLLM, OpenAIEmbedder, OpenAILLM, OpenAIStructuredLLM, Qdrant, RedisDB, type SearchFilters, type SearchMemoryOptions, type SearchResult, type VectorStore, type VectorStoreConfig, VectorStoreFactory, type VectorStoreResult };
618
+ export { type AddMemoryOptions, AnthropicLLM, type DeleteAllMemoryOptions, type Embedder, EmbedderFactory, type EmbeddingConfig, type Entity, type GetAllMemoryOptions, GoogleEmbedder, GoogleLLM, type GraphStoreConfig, GroqLLM, HistoryManagerFactory, type HistoryStoreConfig, type LLM, type LLMConfig, LLMFactory, type LLMResponse, Memory, type MemoryConfig, MemoryConfigSchema, type MemoryItem, MemoryVectorStore, type Message, type MultiModalMessages, type Neo4jConfig, OllamaEmbedder, OllamaLLM, OpenAIEmbedder, OpenAILLM, OpenAIStructuredLLM, Qdrant, RedisDB, type SearchFilters, type SearchMemoryOptions, type SearchResult, type VectorStore, type VectorStoreConfig, VectorStoreFactory, type VectorStoreResult };