@elizaos/plugin-knowledge 1.2.3 → 1.5.10

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.js CHANGED
@@ -810,11 +810,11 @@ async function generateAnthropicText(config, prompt, system, modelName, maxToken
810
810
  prompt,
811
811
  system,
812
812
  temperature: 0.3,
813
- maxTokens
813
+ maxOutputTokens: maxTokens
814
814
  });
815
- const totalTokens = result.usage.promptTokens + result.usage.completionTokens;
815
+ const totalTokens = (result.usage.inputTokens || 0) + (result.usage.outputTokens || 0);
816
816
  logger2.debug(
817
- `[Document Processor] ${modelName}: ${totalTokens} tokens (${result.usage.promptTokens}\u2192${result.usage.completionTokens})`
817
+ `[Document Processor] ${modelName}: ${totalTokens} tokens (${result.usage.inputTokens || 0}\u2192${result.usage.outputTokens || 0})`
818
818
  );
819
819
  return result;
820
820
  } catch (error) {
@@ -843,11 +843,11 @@ async function generateOpenAIText(config, prompt, system, modelName, maxTokens)
843
843
  prompt,
844
844
  system,
845
845
  temperature: 0.3,
846
- maxTokens
846
+ maxOutputTokens: maxTokens
847
847
  });
848
- const totalTokens = result.usage.promptTokens + result.usage.completionTokens;
848
+ const totalTokens = (result.usage.inputTokens || 0) + (result.usage.outputTokens || 0);
849
849
  logger2.debug(
850
- `[Document Processor] OpenAI ${modelName}: ${totalTokens} tokens (${result.usage.promptTokens}\u2192${result.usage.completionTokens})`
850
+ `[Document Processor] OpenAI ${modelName}: ${totalTokens} tokens (${result.usage.inputTokens || 0}\u2192${result.usage.outputTokens || 0})`
851
851
  );
852
852
  return result;
853
853
  }
@@ -862,15 +862,15 @@ async function generateGoogleText(prompt, system, modelName, maxTokens, config)
862
862
  prompt,
863
863
  system,
864
864
  temperature: 0.3,
865
- maxTokens
865
+ maxOutputTokens: maxTokens
866
866
  });
867
- const totalTokens = result.usage.promptTokens + result.usage.completionTokens;
867
+ const totalTokens = (result.usage.inputTokens || 0) + (result.usage.outputTokens || 0);
868
868
  logger2.debug(
869
- `[Document Processor] Google ${modelName}: ${totalTokens} tokens (${result.usage.promptTokens}\u2192${result.usage.completionTokens})`
869
+ `[Document Processor] Google ${modelName}: ${totalTokens} tokens (${result.usage.inputTokens || 0}\u2192${result.usage.outputTokens || 0})`
870
870
  );
871
871
  return result;
872
872
  }
873
- async function generateOpenRouterText(config, prompt, system, modelName, maxTokens, cacheDocument, cacheOptions, autoCacheContextualRetrieval = true) {
873
+ async function generateOpenRouterText(config, prompt, system, modelName, maxTokens, cacheDocument, _cacheOptions, autoCacheContextualRetrieval = true) {
874
874
  const openrouter = createOpenRouter({
875
875
  apiKey: config.OPENROUTER_API_KEY,
876
876
  baseURL: config.OPENROUTER_BASE_URL
@@ -891,7 +891,6 @@ async function generateOpenRouterText(config, prompt, system, modelName, maxToke
891
891
  }
892
892
  }
893
893
  if (documentForCaching && supportsCaching) {
894
- const effectiveCacheOptions = cacheOptions || { type: "ephemeral" };
895
894
  let promptText = prompt;
896
895
  if (promptText.includes("<document>")) {
897
896
  promptText = promptText.replace(/<document>[\s\S]*?<\/document>/, "").trim();
@@ -978,7 +977,7 @@ async function generateClaudeWithCaching(promptText, system, modelInstance, mode
978
977
  model: modelInstance,
979
978
  messages,
980
979
  temperature: 0.3,
981
- maxTokens,
980
+ maxOutputTokens: maxTokens,
982
981
  providerOptions: {
983
982
  openrouter: {
984
983
  usage: {
@@ -988,9 +987,9 @@ async function generateClaudeWithCaching(promptText, system, modelInstance, mode
988
987
  }
989
988
  });
990
989
  logCacheMetrics(result);
991
- const totalTokens = result.usage.promptTokens + result.usage.completionTokens;
990
+ const totalTokens = (result.usage.inputTokens || 0) + (result.usage.outputTokens || 0);
992
991
  logger2.debug(
993
- `[Document Processor] OpenRouter ${modelName}: ${totalTokens} tokens (${result.usage.promptTokens}\u2192${result.usage.completionTokens})`
992
+ `[Document Processor] OpenRouter ${modelName}: ${totalTokens} tokens (${result.usage.inputTokens || 0}\u2192${result.usage.outputTokens || 0})`
994
993
  );
995
994
  return result;
996
995
  }
@@ -1029,7 +1028,7 @@ ${promptText}`;
1029
1028
  model: modelInstance,
1030
1029
  prompt: geminiPrompt,
1031
1030
  temperature: 0.3,
1032
- maxTokens,
1031
+ maxOutputTokens: maxTokens,
1033
1032
  providerOptions: {
1034
1033
  openrouter: {
1035
1034
  usage: {
@@ -1040,10 +1039,10 @@ ${promptText}`;
1040
1039
  }
1041
1040
  });
1042
1041
  logCacheMetrics(result);
1043
- const totalTokens = result.usage.promptTokens + result.usage.completionTokens;
1042
+ const totalTokens = (result.usage.inputTokens || 0) + (result.usage.outputTokens || 0);
1044
1043
  const cachingType = usingImplicitCaching ? "implicit" : "standard";
1045
1044
  logger2.debug(
1046
- `[Document Processor] OpenRouter ${modelName} (${cachingType} caching): ${totalTokens} tokens (${result.usage.promptTokens}\u2192${result.usage.completionTokens})`
1045
+ `[Document Processor] OpenRouter ${modelName} (${cachingType} caching): ${totalTokens} tokens (${result.usage.inputTokens || 0}\u2192${result.usage.outputTokens || 0})`
1047
1046
  );
1048
1047
  return result;
1049
1048
  }
@@ -1053,7 +1052,7 @@ async function generateStandardOpenRouterText(prompt, system, modelInstance, mod
1053
1052
  prompt,
1054
1053
  system,
1055
1054
  temperature: 0.3,
1056
- maxTokens,
1055
+ maxOutputTokens: maxTokens,
1057
1056
  providerOptions: {
1058
1057
  openrouter: {
1059
1058
  usage: {
@@ -1063,9 +1062,9 @@ async function generateStandardOpenRouterText(prompt, system, modelInstance, mod
1063
1062
  }
1064
1063
  }
1065
1064
  });
1066
- const totalTokens = result.usage.promptTokens + result.usage.completionTokens;
1065
+ const totalTokens = (result.usage.inputTokens || 0) + (result.usage.outputTokens || 0);
1067
1066
  logger2.debug(
1068
- `[Document Processor] OpenRouter ${modelName}: ${totalTokens} tokens (${result.usage.promptTokens}\u2192${result.usage.completionTokens})`
1067
+ `[Document Processor] OpenRouter ${modelName}: ${totalTokens} tokens (${result.usage.inputTokens || 0}\u2192${result.usage.outputTokens || 0})`
1069
1068
  );
1070
1069
  return result;
1071
1070
  }
@@ -2609,7 +2608,7 @@ var KnowledgeService = class _KnowledgeService extends Service {
2609
2608
  return !!existingDocument;
2610
2609
  }
2611
2610
  async getKnowledge(message, scope) {
2612
- logger6.debug("KnowledgeService: getKnowledge called for message id: " + message.id);
2611
+ logger6.debug(`KnowledgeService: getKnowledge called for message id: ${message.id}`);
2613
2612
  if (!message?.content?.text || message?.content?.text.trim().length === 0) {
2614
2613
  logger6.warn("KnowledgeService: Invalid or empty message content for knowledge query.");
2615
2614
  return [];
@@ -2900,7 +2899,7 @@ var KnowledgeService = class _KnowledgeService extends Service {
2900
2899
  return chunks.map((chunk, index) => {
2901
2900
  const fragmentIdContent = `${document.id}-fragment-${index}-${Date.now()}`;
2902
2901
  const fragmentId = createUniqueUuid(
2903
- this.runtime.agentId + fragmentIdContent,
2902
+ this.runtime,
2904
2903
  fragmentIdContent
2905
2904
  );
2906
2905
  return {
@@ -4016,7 +4015,7 @@ var processKnowledgeAction = {
4016
4015
  };
4017
4016
  const result = await service.addKnowledge(knowledgeOptions);
4018
4017
  response = {
4019
- text: `I've successfully processed the document "${fileName}". It has been split into ${result.fragmentCount} searchable fragments and added to my knowledge base.`
4018
+ text: `I've successfully processed the document "${fileName}". It has been split into ${result?.fragmentCount || 0} searchable fragments and added to my knowledge base.`
4020
4019
  };
4021
4020
  } else {
4022
4021
  const knowledgeContent = text.replace(/^(add|store|remember|process|learn)\s+(this|that|the following)?:?\s*/i, "").trim();
@@ -4038,7 +4037,7 @@ var processKnowledgeAction = {
4038
4037
  roomId: message.roomId,
4039
4038
  entityId: message.entityId
4040
4039
  };
4041
- const result = await service.addKnowledge(knowledgeOptions);
4040
+ await service.addKnowledge(knowledgeOptions);
4042
4041
  response = {
4043
4042
  text: `I've added that information to my knowledge base. It has been stored and indexed for future reference.`
4044
4043
  };