@n8n/n8n-nodes-langchain 1.59.0 → 1.60.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (29) hide show
  1. package/dist/build.tsbuildinfo +1 -1
  2. package/dist/nodes/chains/ChainLLM/ChainLlm.node.js +10 -0
  3. package/dist/nodes/chains/ChainLLM/ChainLlm.node.js.map +1 -1
  4. package/dist/nodes/llms/LMChatOpenAi/LmChatOpenAi.node.js +18 -0
  5. package/dist/nodes/llms/LMChatOpenAi/LmChatOpenAi.node.js.map +1 -1
  6. package/dist/nodes/llms/LmChatAzureOpenAi/LmChatAzureOpenAi.node.js +34 -2
  7. package/dist/nodes/llms/LmChatAzureOpenAi/LmChatAzureOpenAi.node.js.map +1 -1
  8. package/dist/nodes/llms/LmChatGoogleVertex/LmChatGoogleVertex.node.js +9 -4
  9. package/dist/nodes/llms/LmChatGoogleVertex/LmChatGoogleVertex.node.js.map +1 -1
  10. package/dist/nodes/llms/N8nLlmTracing.js +2 -2
  11. package/dist/nodes/llms/N8nLlmTracing.js.map +1 -1
  12. package/dist/nodes/tools/ToolHttpRequest/utils.d.ts +1 -1
  13. package/dist/nodes/tools/ToolHttpRequest/utils.js +5 -5
  14. package/dist/nodes/tools/ToolHttpRequest/utils.js.map +1 -1
  15. package/dist/nodes/vector_store/shared/createVectorStoreNode.js +3 -3
  16. package/dist/nodes/vector_store/shared/createVectorStoreNode.js.map +1 -1
  17. package/dist/nodes/vendors/OpenAi/actions/router.js +8 -0
  18. package/dist/nodes/vendors/OpenAi/actions/router.js.map +1 -1
  19. package/dist/nodes/vendors/OpenAi/helpers/error-handling.d.ts +3 -0
  20. package/dist/nodes/vendors/OpenAi/helpers/error-handling.js +16 -0
  21. package/dist/nodes/vendors/OpenAi/helpers/error-handling.js.map +1 -0
  22. package/dist/nodes/vendors/OpenAi/methods/listSearch.js +1 -1
  23. package/dist/nodes/vendors/OpenAi/methods/listSearch.js.map +1 -1
  24. package/dist/types/nodes.json +2 -2
  25. package/dist/utils/helpers.d.ts +2 -2
  26. package/dist/utils/helpers.js.map +1 -1
  27. package/dist/utils/logWrapper.js +11 -11
  28. package/dist/utils/logWrapper.js.map +1 -1
  29. package/package.json +8 -7
@@ -0,0 +1,16 @@
1
+ "use strict";
2
+ Object.defineProperty(exports, "__esModule", { value: true });
3
+ exports.getCustomErrorMessage = getCustomErrorMessage;
4
+ exports.isOpenAiError = isOpenAiError;
5
+ const error_1 = require("openai/error");
6
+ const errorMap = {
7
+ insufficient_quota: 'OpenAI: Insufficient quota',
8
+ rate_limit_exceeded: 'OpenAI: Rate limit reached',
9
+ };
10
+ function getCustomErrorMessage(errorCode) {
11
+ return errorMap[errorCode];
12
+ }
13
+ function isOpenAiError(error) {
14
+ return error instanceof error_1.OpenAIError;
15
+ }
16
+ //# sourceMappingURL=error-handling.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"error-handling.js","sourceRoot":"","sources":["../../../../../nodes/vendors/OpenAi/helpers/error-handling.ts"],"names":[],"mappings":";;AAOA,sDAEC;AAED,sCAEC;AAbD,wCAA2C;AAE3C,MAAM,QAAQ,GAA2B;IACxC,kBAAkB,EAAE,4BAA4B;IAChD,mBAAmB,EAAE,4BAA4B;CACjD,CAAC;AAEF,SAAgB,qBAAqB,CAAC,SAAiB;IACtD,OAAO,QAAQ,CAAC,SAAS,CAAC,CAAC;AAC5B,CAAC;AAED,SAAgB,aAAa,CAAC,KAAU;IACvC,OAAO,KAAK,YAAY,mBAAW,CAAC;AACrC,CAAC"}
@@ -56,7 +56,7 @@ const getModelSearch = (filterCondition) => async (ctx, filter) => {
56
56
  };
57
57
  };
58
58
  async function modelSearch(filter) {
59
- return await getModelSearch((model) => model.id.startsWith('gpt-') || model.id.startsWith('ft:'))(this, filter);
59
+ return await getModelSearch((model) => model.id.startsWith('gpt-') || model.id.startsWith('ft:') || model.id.startsWith('o1'))(this, filter);
60
60
  }
61
61
  async function imageModelSearch(filter) {
62
62
  return await getModelSearch((model) => model.id.includes('vision') || model.id.includes('gpt-4o'))(this, filter);
@@ -1 +1 @@
1
- {"version":3,"file":"listSearch.js","sourceRoot":"","sources":["../../../../../nodes/vendors/OpenAi/methods/listSearch.ts"],"names":[],"mappings":";;AAWA,gCA6BC;AAkCD,kCAQC;AAED,4CAOC;AAED,0CAkDC;AAtID,4CAA0C;AAEnC,KAAK,UAAU,UAAU,CAE/B,MAAe;IAEf,MAAM,EAAE,IAAI,EAAE,GAAG,MAAM,sBAAU,CAAC,IAAI,CAAC,IAAI,EAAE,KAAK,EAAE,QAAQ,CAAC,CAAC;IAE9D,IAAI,MAAM,EAAE,CAAC;QACZ,MAAM,OAAO,GAA2B,EAAE,CAAC;QAE3C,KAAK,MAAM,IAAI,IAAI,IAAI,IAAI,EAAE,EAAE,CAAC;YAC/B,IAAK,IAAI,CAAC,QAAmB,EAAE,WAAW,EAAE,CAAC,QAAQ,CAAC,MAAM,CAAC,WAAW,EAAE,CAAC,EAAE,CAAC;gBAC7E,OAAO,CAAC,IAAI,CAAC;oBACZ,IAAI,EAAE,IAAI,CAAC,QAAkB;oBAC7B,KAAK,EAAE,IAAI,CAAC,EAAY;iBACxB,CAAC,CAAC;YACJ,CAAC;QACF,CAAC;QAED,OAAO;YACN,OAAO;SACP,CAAC;IACH,CAAC;SAAM,CAAC;QACP,OAAO;YACN,OAAO,EAAE,CAAC,IAAI,IAAI,EAAE,CAAC,CAAC,GAAG,CAAC,CAAC,IAAiB,EAAE,EAAE,CAAC,CAAC;gBACjD,IAAI,EAAE,IAAI,CAAC,QAAkB;gBAC7B,KAAK,EAAE,IAAI,CAAC,EAAY;aACxB,CAAC,CAAC;SACH,CAAC;IACH,CAAC;AACF,CAAC;AAED,MAAM,cAAc,GACnB,CAAC,eAA0C,EAAE,EAAE,CAC/C,KAAK,EAAE,GAA0B,EAAE,MAAe,EAAkC,EAAE;IACrF,IAAI,EAAE,IAAI,EAAE,GAAG,CAAC,MAAM,sBAAU,CAAC,IAAI,CAAC,GAAG,EAAE,KAAK,EAAE,SAAS,CAAC,CAAsB,CAAC;IAEnF,IAAI,GAAG,IAAI,EAAE,MAAM,CAAC,CAAC,KAAK,EAAE,EAAE,CAAC,eAAe,CAAC,KAAK,CAAC,CAAC,CAAC;IAEvD,IAAI,OAAO,GAA2B,EAAE,CAAC;IAEzC,IAAI,MAAM,EAAE,CAAC;QACZ,KAAK,MAAM,KAAK,IAAI,IAAI,IAAI,EAAE,EAAE,CAAC;YAChC,IAAI,KAAK,CAAC,EAAE,EAAE,WAAW,EAAE,CAAC,QAAQ,CAAC,MAAM,CAAC,WAAW,EAAE,CAAC,EAAE,CAAC;gBAC5D,OAAO,CAAC,IAAI,CAAC;oBACZ,IAAI,EAAE,KAAK,CAAC,EAAE,CAAC,WAAW,EAAE;oBAC5B,KAAK,EAAE,KAAK,CAAC,EAAE;iBACf,CAAC,CAAC;YACJ,CAAC;QACF,CAAC;IACF,CAAC;SAAM,CAAC;QACP,OAAO,GAAG,CAAC,IAAI,IAAI,EAAE,CAAC,CAAC,GAAG,CAAC,CAAC,KAAK,EAAE,EAAE,CAAC,CAAC;YACtC,IAAI,EAAE,KAAK,CAAC,EAAE,CAAC,WAAW,EAAE;YAC5B,KAAK,EAAE,KAAK,CAAC,EAAE;SACf,CAAC,CAAC,CAAC;IACL,CAAC;IAED,OAAO,GAAG,OAAO,CAAC,IAAI,CAAC,CAAC,CAAC,EAAE,CAAC,EAAE,EAAE,CAAC,CAAC,CAAC,IAAI,CAAC,aAAa,CAAC,CAAC,CAAC,IAAI,CAAC,CAAC,CAAC;IAE/D,OAAO;QACN,OAAO;KACP,CAAC;AACH,CAAC,CAAC;AAEI,KAAK,UAAU,WAAW,CAEhC,MAAe;IAEf,OAAO,MAAM,cAAc,CAAC,CAAC,KAAK,EAAE,EAAE,CAAC,KAAK,CAAC,EAAE,CAAC,UAAU,CAAC,MAAM,CAAC,IAAI,KAAK,CAAC,EAAE,CAAC,UAAU,CAAC,KAAK,CAAC,CAAC,CAChG,IAAI,EACJ,MAAM,CACN,CAAC;AACH,CAAC;AAEM,KAAK,UAAU,gBAAgB,CAErC,MAAe;IAEf,OAAO,MAAM,cAAc,CAC1B,CAAC,KAAK,EAAE,EAAE,CAAC,KAAK,CAAC,EAAE,CAAC,QAAQ,CAAC,QAAQ,CAAC,IAAI,KAAK,CAAC,EAAE,CAAC,QAAQ,CAAC,QAAQ,CAAC,CACrE,CAAC,IAAI,EAAE,MAAM,CAAC,CAAC;AACjB,CAAC;AAEM,KAAK,UAAU,eAAe,CAEpC,MAAe,EACf,eAAwB;IAExB,MAAM,EAAE,IAAI,EAAE,QAAQ,EAAE,OAAO,EAAE,GAAG,CAAC,MAAM,sBAAU,CAAC,IAAI,CAAC,IAAI,EAAE,KAAK,EAAE,aAAa,EAAE;QACtF,OAAO,EAAE;YACR,aAAa,EAAE,eAAe;SAC9B;QACD,EAAE,EAAE;YACH,KAAK,EAAE,GAAG;YACV,KAAK,EAAE,eAAe;SACtB;KACD,CAAC,CAKD,CAAC;IAEF,IAAI,QAAQ,EAAE,CAAC;QACd,eAAe,GAAG,OAAO,CAAC;IAC3B,CAAC;SAAM,CAAC;QACP,eAAe,GAAG,SAAS,CAAC;IAC7B,CAAC;IAED,IAAI,MAAM,EAAE,CAAC;QACZ,MAAM,OAAO,GAA2B,EAAE,CAAC;QAE3C,KAAK,MAAM,SAAS,IAAI,IAAI,IAAI,EAAE,EAAE,CAAC;YACpC,IAAI,SAAS,CAAC,IAAI,EAAE,WAAW,EAAE,CAAC,QAAQ,CAAC,MAAM,CAAC,WAAW,EAAE,CAAC,EAAE,CAAC;gBAClE,OAAO,CAAC,IAAI,CAAC;oBACZ,IAAI,EAAE,SAAS,CAAC,IAAI;oBACpB,KAAK,EAAE,SAAS,CAAC,EAAE;iBACnB,CAAC,CAAC;YACJ,CAAC;QACF,CAAC;QAED,OAAO;YACN,OAAO;SACP,CAAC;IACH,CAAC;SAAM,CAAC;QACP,OAAO;YACN,OAAO,EAAE,CAAC,IAAI,IAAI,EAAE,CAAC,CAAC,GAAG,CAAC,CAAC,SAAS,EAAE,EAAE,CAAC,CAAC;gBACzC,IAAI,EAAE,SAAS,CAAC,IAAI,IAAI,SAAS,CAAC,EAAE;gBACpC,KAAK,EAAE,SAAS,CAAC,EAAE;aACnB,CAAC,CAAC;YACH,eAAe;SACf,CAAC;IACH,CAAC;AACF,CAAC"}
1
+ {"version":3,"file":"listSearch.js","sourceRoot":"","sources":["../../../../../nodes/vendors/OpenAi/methods/listSearch.ts"],"names":[],"mappings":";;AAWA,gCA6BC;AAkCD,kCAQC;AAED,4CAOC;AAED,0CAkDC;AAtID,4CAA0C;AAEnC,KAAK,UAAU,UAAU,CAE/B,MAAe;IAEf,MAAM,EAAE,IAAI,EAAE,GAAG,MAAM,sBAAU,CAAC,IAAI,CAAC,IAAI,EAAE,KAAK,EAAE,QAAQ,CAAC,CAAC;IAE9D,IAAI,MAAM,EAAE,CAAC;QACZ,MAAM,OAAO,GAA2B,EAAE,CAAC;QAE3C,KAAK,MAAM,IAAI,IAAI,IAAI,IAAI,EAAE,EAAE,CAAC;YAC/B,IAAK,IAAI,CAAC,QAAmB,EAAE,WAAW,EAAE,CAAC,QAAQ,CAAC,MAAM,CAAC,WAAW,EAAE,CAAC,EAAE,CAAC;gBAC7E,OAAO,CAAC,IAAI,CAAC;oBACZ,IAAI,EAAE,IAAI,CAAC,QAAkB;oBAC7B,KAAK,EAAE,IAAI,CAAC,EAAY;iBACxB,CAAC,CAAC;YACJ,CAAC;QACF,CAAC;QAED,OAAO;YACN,OAAO;SACP,CAAC;IACH,CAAC;SAAM,CAAC;QACP,OAAO;YACN,OAAO,EAAE,CAAC,IAAI,IAAI,EAAE,CAAC,CAAC,GAAG,CAAC,CAAC,IAAiB,EAAE,EAAE,CAAC,CAAC;gBACjD,IAAI,EAAE,IAAI,CAAC,QAAkB;gBAC7B,KAAK,EAAE,IAAI,CAAC,EAAY;aACxB,CAAC,CAAC;SACH,CAAC;IACH,CAAC;AACF,CAAC;AAED,MAAM,cAAc,GACnB,CAAC,eAA0C,EAAE,EAAE,CAC/C,KAAK,EAAE,GAA0B,EAAE,MAAe,EAAkC,EAAE;IACrF,IAAI,EAAE,IAAI,EAAE,GAAG,CAAC,MAAM,sBAAU,CAAC,IAAI,CAAC,GAAG,EAAE,KAAK,EAAE,SAAS,CAAC,CAAsB,CAAC;IAEnF,IAAI,GAAG,IAAI,EAAE,MAAM,CAAC,CAAC,KAAK,EAAE,EAAE,CAAC,eAAe,CAAC,KAAK,CAAC,CAAC,CAAC;IAEvD,IAAI,OAAO,GAA2B,EAAE,CAAC;IAEzC,IAAI,MAAM,EAAE,CAAC;QACZ,KAAK,MAAM,KAAK,IAAI,IAAI,IAAI,EAAE,EAAE,CAAC;YAChC,IAAI,KAAK,CAAC,EAAE,EAAE,WAAW,EAAE,CAAC,QAAQ,CAAC,MAAM,CAAC,WAAW,EAAE,CAAC,EAAE,CAAC;gBAC5D,OAAO,CAAC,IAAI,CAAC;oBACZ,IAAI,EAAE,KAAK,CAAC,EAAE,CAAC,WAAW,EAAE;oBAC5B,KAAK,EAAE,KAAK,CAAC,EAAE;iBACf,CAAC,CAAC;YACJ,CAAC;QACF,CAAC;IACF,CAAC;SAAM,CAAC;QACP,OAAO,GAAG,CAAC,IAAI,IAAI,EAAE,CAAC,CAAC,GAAG,CAAC,CAAC,KAAK,EAAE,EAAE,CAAC,CAAC;YACtC,IAAI,EAAE,KAAK,CAAC,EAAE,CAAC,WAAW,EAAE;YAC5B,KAAK,EAAE,KAAK,CAAC,EAAE;SACf,CAAC,CAAC,CAAC;IACL,CAAC;IAED,OAAO,GAAG,OAAO,CAAC,IAAI,CAAC,CAAC,CAAC,EAAE,CAAC,EAAE,EAAE,CAAC,CAAC,CAAC,IAAI,CAAC,aAAa,CAAC,CAAC,CAAC,IAAI,CAAC,CAAC,CAAC;IAE/D,OAAO;QACN,OAAO;KACP,CAAC;AACH,CAAC,CAAC;AAEI,KAAK,UAAU,WAAW,CAEhC,MAAe;IAEf,OAAO,MAAM,cAAc,CAC1B,CAAC,KAAK,EAAE,EAAE,CACT,KAAK,CAAC,EAAE,CAAC,UAAU,CAAC,MAAM,CAAC,IAAI,KAAK,CAAC,EAAE,CAAC,UAAU,CAAC,KAAK,CAAC,IAAI,KAAK,CAAC,EAAE,CAAC,UAAU,CAAC,IAAI,CAAC,CACvF,CAAC,IAAI,EAAE,MAAM,CAAC,CAAC;AACjB,CAAC;AAEM,KAAK,UAAU,gBAAgB,CAErC,MAAe;IAEf,OAAO,MAAM,cAAc,CAC1B,CAAC,KAAK,EAAE,EAAE,CAAC,KAAK,CAAC,EAAE,CAAC,QAAQ,CAAC,QAAQ,CAAC,IAAI,KAAK,CAAC,EAAE,CAAC,QAAQ,CAAC,QAAQ,CAAC,CACrE,CAAC,IAAI,EAAE,MAAM,CAAC,CAAC;AACjB,CAAC;AAEM,KAAK,UAAU,eAAe,CAEpC,MAAe,EACf,eAAwB;IAExB,MAAM,EAAE,IAAI,EAAE,QAAQ,EAAE,OAAO,EAAE,GAAG,CAAC,MAAM,sBAAU,CAAC,IAAI,CAAC,IAAI,EAAE,KAAK,EAAE,aAAa,EAAE;QACtF,OAAO,EAAE;YACR,aAAa,EAAE,eAAe;SAC9B;QACD,EAAE,EAAE;YACH,KAAK,EAAE,GAAG;YACV,KAAK,EAAE,eAAe;SACtB;KACD,CAAC,CAKD,CAAC;IAEF,IAAI,QAAQ,EAAE,CAAC;QACd,eAAe,GAAG,OAAO,CAAC;IAC3B,CAAC;SAAM,CAAC;QACP,eAAe,GAAG,SAAS,CAAC;IAC7B,CAAC;IAED,IAAI,MAAM,EAAE,CAAC;QACZ,MAAM,OAAO,GAA2B,EAAE,CAAC;QAE3C,KAAK,MAAM,SAAS,IAAI,IAAI,IAAI,EAAE,EAAE,CAAC;YACpC,IAAI,SAAS,CAAC,IAAI,EAAE,WAAW,EAAE,CAAC,QAAQ,CAAC,MAAM,CAAC,WAAW,EAAE,CAAC,EAAE,CAAC;gBAClE,OAAO,CAAC,IAAI,CAAC;oBACZ,IAAI,EAAE,SAAS,CAAC,IAAI;oBACpB,KAAK,EAAE,SAAS,CAAC,EAAE;iBACnB,CAAC,CAAC;YACJ,CAAC;QACF,CAAC;QAED,OAAO;YACN,OAAO;SACP,CAAC;IACH,CAAC;SAAM,CAAC;QACP,OAAO;YACN,OAAO,EAAE,CAAC,IAAI,IAAI,EAAE,CAAC,CAAC,GAAG,CAAC,CAAC,SAAS,EAAE,EAAE,CAAC,CAAC;gBACzC,IAAI,EAAE,SAAS,CAAC,IAAI,IAAI,SAAS,CAAC,EAAE;gBACpC,KAAK,EAAE,SAAS,CAAC,EAAE;aACnB,CAAC,CAAC;YACH,eAAe;SACf,CAAC;IACH,CAAC;AACF,CAAC"}
@@ -24,7 +24,7 @@
24
24
  {"displayName":"Embeddings OpenAI","name":"@n8n/n8n-nodes-langchain.embeddingsOpenAi","credentials":[{"name":"openAiApi","required":true}],"group":["transform"],"version":1,"description":"Use Embeddings OpenAI","defaults":{"name":"Embeddings OpenAI"},"codex":{"categories":["AI"],"subcategories":{"AI":["Embeddings"]},"resources":{"primaryDocumentation":[{"url":"https://docs.n8n.io/integrations/builtin/cluster-nodes/sub-nodes/n8n-nodes-langchain.embeddingsopenai/"}]}},"inputs":[],"outputs":["ai_embedding"],"outputNames":["Embeddings"],"requestDefaults":{"ignoreHttpStatusErrors":true,"baseURL":"={{ $parameter.options?.baseURL?.split(\"/\").slice(0,-1).join(\"/\") || \"https://api.openai.com\" }}"},"properties":[{"displayName":"This node must be connected to a vector store. <a data-action='openSelectiveNodeCreator' data-action-parameter-connectiontype='ai_vectorStore'>Insert one</a>","name":"notice","type":"notice","default":"","typeOptions":{"containerClass":"ndv-connection-hint-notice"}},{"displayName":"Model","name":"model","type":"options","description":"The model which will generate the embeddings. <a href=\"https://platform.openai.com/docs/models/overview\">Learn more</a>.","typeOptions":{"loadOptions":{"routing":{"request":{"method":"GET","url":"={{ $parameter.options?.baseURL?.split(\"/\").slice(-1).pop() || \"v1\" }}/models"},"output":{"postReceive":[{"type":"rootProperty","properties":{"property":"data"}},{"type":"filter","properties":{"pass":"={{ $responseItem.id.includes('embed') }}"}},{"type":"setKeyValue","properties":{"name":"={{$responseItem.id}}","value":"={{$responseItem.id}}"}},{"type":"sort","properties":{"key":"name"}}]}}}},"routing":{"send":{"type":"body","property":"model"}},"default":"text-embedding-ada-002","displayOptions":{"show":{"@version":[1]}}},{"displayName":"Model","name":"model","type":"options","description":"The model which will generate the embeddings. <a href=\"https://platform.openai.com/docs/models/overview\">Learn more</a>.","typeOptions":{"loadOptions":{"routing":{"request":{"method":"GET","url":"={{ $parameter.options?.baseURL?.split(\"/\").slice(-1).pop() || \"v1\" }}/models"},"output":{"postReceive":[{"type":"rootProperty","properties":{"property":"data"}},{"type":"filter","properties":{"pass":"={{ $responseItem.id.includes('embed') }}"}},{"type":"setKeyValue","properties":{"name":"={{$responseItem.id}}","value":"={{$responseItem.id}}"}},{"type":"sort","properties":{"key":"name"}}]}}}},"routing":{"send":{"type":"body","property":"model"}},"default":"text-embedding-3-small","displayOptions":{"hide":{"@version":[1]}}},{"displayName":"Options","name":"options","placeholder":"Add Option","description":"Additional options to add","type":"collection","default":{},"options":[{"displayName":"Base URL","name":"baseURL","default":"https://api.openai.com/v1","description":"Override the default base URL for the API","type":"string"},{"displayName":"Batch Size","name":"batchSize","default":512,"typeOptions":{"maxValue":2048},"description":"Maximum number of documents to send in each request","type":"number"},{"displayName":"Strip New Lines","name":"stripNewLines","default":true,"description":"Whether to strip new lines from the input text","type":"boolean"},{"displayName":"Timeout","name":"timeout","default":-1,"description":"Maximum amount of time a request is allowed to take in seconds. Set to -1 for no timeout.","type":"number"}]}],"iconUrl":{"light":"icons/@n8n/n8n-nodes-langchain/dist/nodes/embeddings/EmbeddingsOpenAI/openAiLight.svg","dark":"icons/@n8n/n8n-nodes-langchain/dist/nodes/embeddings/EmbeddingsOpenAI/openAiLight.dark.svg"}},
25
25
  {"displayName":"Embeddings Ollama","name":"@n8n/n8n-nodes-langchain.embeddingsOllama","group":["transform"],"version":1,"description":"Use Ollama Embeddings","defaults":{"name":"Embeddings Ollama"},"credentials":[{"name":"ollamaApi","required":true}],"requestDefaults":{"ignoreHttpStatusErrors":true,"baseURL":"={{ $credentials.baseUrl.replace(new RegExp(\"/$\"), \"\") }}"},"codex":{"categories":["AI"],"subcategories":{"AI":["Embeddings"]},"resources":{"primaryDocumentation":[{"url":"https://docs.n8n.io/integrations/builtin/cluster-nodes/sub-nodes/n8n-nodes-langchain.embeddingsollama/"}]}},"inputs":[],"outputs":["ai_embedding"],"outputNames":["Embeddings"],"properties":[{"displayName":"This node must be connected to a vector store. <a data-action='openSelectiveNodeCreator' data-action-parameter-connectiontype='ai_vectorStore'>Insert one</a>","name":"notice","type":"notice","default":"","typeOptions":{"containerClass":"ndv-connection-hint-notice"}},{"displayName":"Model","name":"model","type":"options","default":"llama2","description":"The model which will generate the completion. To download models, visit <a href=\"https://ollama.ai/library\">Ollama Models Library</a>.","typeOptions":{"loadOptions":{"routing":{"request":{"method":"GET","url":"/api/tags"},"output":{"postReceive":[{"type":"rootProperty","properties":{"property":"models"}},{"type":"setKeyValue","properties":{"name":"={{$responseItem.name}}","value":"={{$responseItem.name}}"}},{"type":"sort","properties":{"key":"name"}}]}}}},"routing":{"send":{"type":"body","property":"model"}},"required":true}],"iconUrl":"icons/@n8n/n8n-nodes-langchain/dist/nodes/embeddings/EmbeddingsOllama/ollama.svg"},
26
26
  {"displayName":"Anthropic Chat Model","name":"@n8n/n8n-nodes-langchain.lmChatAnthropic","group":["transform"],"version":[1,1.1,1.2],"defaultVersion":1.2,"description":"Language Model Anthropic","defaults":{"name":"Anthropic Chat Model"},"codex":{"categories":["AI"],"subcategories":{"AI":["Language Models","Root Nodes"],"Language Models":["Chat Models (Recommended)"]},"resources":{"primaryDocumentation":[{"url":"https://docs.n8n.io/integrations/builtin/cluster-nodes/sub-nodes/n8n-nodes-langchain.lmchatanthropic/"}]},"alias":["claude","sonnet","opus"]},"inputs":[],"outputs":["ai_languageModel"],"outputNames":["Model"],"credentials":[{"name":"anthropicApi","required":true}],"properties":[{"displayName":"This node must be connected to an AI chain. <a data-action='openSelectiveNodeCreator' data-action-parameter-creatorview='AI'>Insert one</a>","name":"notice","type":"notice","default":"","typeOptions":{"containerClass":"ndv-connection-hint-notice"}},{"displayName":"Model","name":"model","type":"options","options":[{"name":"Claude 3 Opus(20240229)","value":"claude-3-opus-20240229"},{"name":"Claude 3.5 Sonnet(20240620)","value":"claude-3-5-sonnet-20240620"},{"name":"Claude 3 Sonnet(20240229)","value":"claude-3-sonnet-20240229"},{"name":"Claude 3 Haiku(20240307)","value":"claude-3-haiku-20240307"},{"name":"LEGACY: Claude 2","value":"claude-2"},{"name":"LEGACY: Claude 2.1","value":"claude-2.1"},{"name":"LEGACY: Claude Instant 1.2","value":"claude-instant-1.2"},{"name":"LEGACY: Claude Instant 1","value":"claude-instant-1"}],"description":"The model which will generate the completion. <a href=\"https://docs.anthropic.com/claude/docs/models-overview\">Learn more</a>.","default":"claude-2","displayOptions":{"show":{"@version":[1]}}},{"displayName":"Model","name":"model","type":"options","options":[{"name":"Claude 3 Opus(20240229)","value":"claude-3-opus-20240229"},{"name":"Claude 3.5 Sonnet(20240620)","value":"claude-3-5-sonnet-20240620"},{"name":"Claude 3 Sonnet(20240229)","value":"claude-3-sonnet-20240229"},{"name":"Claude 3 Haiku(20240307)","value":"claude-3-haiku-20240307"},{"name":"LEGACY: Claude 2","value":"claude-2"},{"name":"LEGACY: Claude 2.1","value":"claude-2.1"},{"name":"LEGACY: Claude Instant 1.2","value":"claude-instant-1.2"},{"name":"LEGACY: Claude Instant 1","value":"claude-instant-1"}],"description":"The model which will generate the completion. <a href=\"https://docs.anthropic.com/claude/docs/models-overview\">Learn more</a>.","default":"claude-3-sonnet-20240229","displayOptions":{"show":{"@version":[1.1]}}},{"displayName":"Model","name":"model","type":"options","options":[{"name":"Claude 3 Opus(20240229)","value":"claude-3-opus-20240229"},{"name":"Claude 3.5 Sonnet(20240620)","value":"claude-3-5-sonnet-20240620"},{"name":"Claude 3 Sonnet(20240229)","value":"claude-3-sonnet-20240229"},{"name":"Claude 3 Haiku(20240307)","value":"claude-3-haiku-20240307"}],"description":"The model which will generate the completion. <a href=\"https://docs.anthropic.com/claude/docs/models-overview\">Learn more</a>.","default":"claude-3-5-sonnet-20240620","displayOptions":{"show":{"@version":[{"_cnd":{"gte":1.2}}]}}},{"displayName":"Options","name":"options","placeholder":"Add Option","description":"Additional options to add","type":"collection","default":{},"options":[{"displayName":"Maximum Number of Tokens","name":"maxTokensToSample","default":4096,"description":"The maximum number of tokens to generate in the completion","type":"number"},{"displayName":"Sampling Temperature","name":"temperature","default":0.7,"typeOptions":{"maxValue":1,"minValue":0,"numberPrecision":1},"description":"Controls randomness: Lowering results in less random completions. As the temperature approaches zero, the model will become deterministic and repetitive.","type":"number"},{"displayName":"Top K","name":"topK","default":-1,"typeOptions":{"maxValue":1,"minValue":-1,"numberPrecision":1},"description":"Used to remove \"long tail\" low probability responses. Defaults to -1, which disables it.","type":"number"},{"displayName":"Top P","name":"topP","default":1,"typeOptions":{"maxValue":1,"minValue":0,"numberPrecision":1},"description":"Controls diversity via nucleus sampling: 0.5 means half of all likelihood-weighted options are considered. We generally recommend altering this or temperature but not both.","type":"number"}]}],"iconUrl":"icons/@n8n/n8n-nodes-langchain/dist/nodes/llms/LMChatAnthropic/anthropic.svg"},
27
- {"displayName":"Azure OpenAI Chat Model","name":"@n8n/n8n-nodes-langchain.lmChatAzureOpenAi","group":["transform"],"version":1,"description":"For advanced usage with an AI chain","defaults":{"name":"Azure OpenAI Chat Model"},"codex":{"categories":["AI"],"subcategories":{"AI":["Language Models","Root Nodes"],"Language Models":["Chat Models (Recommended)"]},"resources":{"primaryDocumentation":[{"url":"https://docs.n8n.io/integrations/builtin/cluster-nodes/sub-nodes/n8n-nodes-langchain.lmchatazureopenai/"}]}},"inputs":[],"outputs":["ai_languageModel"],"outputNames":["Model"],"credentials":[{"name":"azureOpenAiApi","required":true}],"properties":[{"displayName":"This node must be connected to an AI chain. <a data-action='openSelectiveNodeCreator' data-action-parameter-creatorview='AI'>Insert one</a>","name":"notice","type":"notice","default":"","typeOptions":{"containerClass":"ndv-connection-hint-notice"}},{"displayName":"Model (Deployment) Name","name":"model","type":"string","description":"The name of the model(deployment) to use","default":""},{"displayName":"Options","name":"options","placeholder":"Add Option","description":"Additional options to add","type":"collection","default":{},"options":[{"displayName":"Frequency Penalty","name":"frequencyPenalty","default":0,"typeOptions":{"maxValue":2,"minValue":-2,"numberPrecision":1},"description":"Positive values penalize new tokens based on their existing frequency in the text so far, decreasing the model's likelihood to repeat the same line verbatim","type":"number"},{"displayName":"Maximum Number of Tokens","name":"maxTokens","default":-1,"description":"The maximum number of tokens to generate in the completion. Most models have a context length of 2048 tokens (except for the newest models, which support 32,768).","type":"number","typeOptions":{"maxValue":32768}},{"displayName":"Presence Penalty","name":"presencePenalty","default":0,"typeOptions":{"maxValue":2,"minValue":-2,"numberPrecision":1},"description":"Positive values penalize new tokens based on whether they appear in the text so far, increasing the model's likelihood to talk about new topics","type":"number"},{"displayName":"Sampling Temperature","name":"temperature","default":0.7,"typeOptions":{"maxValue":1,"minValue":0,"numberPrecision":1},"description":"Controls randomness: Lowering results in less random completions. As the temperature approaches zero, the model will become deterministic and repetitive.","type":"number"},{"displayName":"Timeout","name":"timeout","default":60000,"description":"Maximum amount of time a request is allowed to take in milliseconds","type":"number"},{"displayName":"Max Retries","name":"maxRetries","default":2,"description":"Maximum number of retries to attempt","type":"number"},{"displayName":"Top P","name":"topP","default":1,"typeOptions":{"maxValue":1,"minValue":0,"numberPrecision":1},"description":"Controls diversity via nucleus sampling: 0.5 means half of all likelihood-weighted options are considered. We generally recommend altering this or temperature but not both.","type":"number"}]}],"iconUrl":"icons/@n8n/n8n-nodes-langchain/dist/nodes/llms/LmChatAzureOpenAi/azure.svg"},
27
+ {"displayName":"Azure OpenAI Chat Model","name":"@n8n/n8n-nodes-langchain.lmChatAzureOpenAi","group":["transform"],"version":1,"description":"For advanced usage with an AI chain","defaults":{"name":"Azure OpenAI Chat Model"},"codex":{"categories":["AI"],"subcategories":{"AI":["Language Models","Root Nodes"],"Language Models":["Chat Models (Recommended)"]},"resources":{"primaryDocumentation":[{"url":"https://docs.n8n.io/integrations/builtin/cluster-nodes/sub-nodes/n8n-nodes-langchain.lmchatazureopenai/"}]}},"inputs":[],"outputs":["ai_languageModel"],"outputNames":["Model"],"credentials":[{"name":"azureOpenAiApi","required":true}],"properties":[{"displayName":"This node must be connected to an AI chain. <a data-action='openSelectiveNodeCreator' data-action-parameter-creatorview='AI'>Insert one</a>","name":"notice","type":"notice","default":"","typeOptions":{"containerClass":"ndv-connection-hint-notice"}},{"displayName":"If using JSON response format, you must include word \"json\" in the prompt in your chain or agent. Also, make sure to select latest models released post November 2023.","name":"notice","type":"notice","default":"","displayOptions":{"show":{"/options.responseFormat":["json_object"]}}},{"displayName":"Model (Deployment) Name","name":"model","type":"string","description":"The name of the model(deployment) to use","default":""},{"displayName":"Options","name":"options","placeholder":"Add Option","description":"Additional options to add","type":"collection","default":{},"options":[{"displayName":"Frequency Penalty","name":"frequencyPenalty","default":0,"typeOptions":{"maxValue":2,"minValue":-2,"numberPrecision":1},"description":"Positive values penalize new tokens based on their existing frequency in the text so far, decreasing the model's likelihood to repeat the same line verbatim","type":"number"},{"displayName":"Maximum Number of Tokens","name":"maxTokens","default":-1,"description":"The maximum number of tokens to generate in the completion. Most models have a context length of 2048 tokens (except for the newest models, which support 32,768).","type":"number","typeOptions":{"maxValue":32768}},{"displayName":"Response Format","name":"responseFormat","default":"text","type":"options","options":[{"name":"Text","value":"text","description":"Regular text response"},{"name":"JSON","value":"json_object","description":"Enables JSON mode, which should guarantee the message the model generates is valid JSON"}]},{"displayName":"Presence Penalty","name":"presencePenalty","default":0,"typeOptions":{"maxValue":2,"minValue":-2,"numberPrecision":1},"description":"Positive values penalize new tokens based on whether they appear in the text so far, increasing the model's likelihood to talk about new topics","type":"number"},{"displayName":"Sampling Temperature","name":"temperature","default":0.7,"typeOptions":{"maxValue":1,"minValue":0,"numberPrecision":1},"description":"Controls randomness: Lowering results in less random completions. As the temperature approaches zero, the model will become deterministic and repetitive.","type":"number"},{"displayName":"Timeout","name":"timeout","default":60000,"description":"Maximum amount of time a request is allowed to take in milliseconds","type":"number"},{"displayName":"Max Retries","name":"maxRetries","default":2,"description":"Maximum number of retries to attempt","type":"number"},{"displayName":"Top P","name":"topP","default":1,"typeOptions":{"maxValue":1,"minValue":0,"numberPrecision":1},"description":"Controls diversity via nucleus sampling: 0.5 means half of all likelihood-weighted options are considered. We generally recommend altering this or temperature but not both.","type":"number"}]}],"iconUrl":"icons/@n8n/n8n-nodes-langchain/dist/nodes/llms/LmChatAzureOpenAi/azure.svg"},
28
28
  {"displayName":"Google PaLM Language Model","name":"@n8n/n8n-nodes-langchain.lmGooglePalm","hidden":true,"group":["transform"],"version":1,"description":"Language Model Google PaLM","defaults":{"name":"Google PaLM Language Model"},"codex":{"categories":["AI"],"subcategories":{"AI":["Language Models","Root Nodes"],"Language Models":["Text Completion Models"]},"resources":{"primaryDocumentation":[{"url":"https://docs.n8n.io/integrations/builtin/cluster-nodes/sub-nodes/n8n-nodes-langchain.lmgooglepalm/"}]}},"inputs":[],"outputs":["ai_languageModel"],"outputNames":["Model"],"credentials":[{"name":"googlePalmApi","required":true}],"requestDefaults":{"ignoreHttpStatusErrors":true,"baseURL":"={{ $credentials.host }}"},"properties":[{"displayName":"Google PaLM API is <a href='https://ai.google.dev/palm_docs/deprecation' target='_blank'>deprecated</a>. Please use Google Vertex or Google Gemini nodes instead.","name":"deprecated","type":"notice","default":""},{"displayName":"Model","name":"modelName","type":"options","description":"The model which will generate the completion. <a href=\"https://developers.generativeai.google/api/rest/generativelanguage/models/list\">Learn more</a>.","typeOptions":{"loadOptions":{"routing":{"request":{"method":"GET","url":"/v1beta3/models"},"output":{"postReceive":[{"type":"rootProperty","properties":{"property":"models"}},{"type":"filter","properties":{"pass":"={{ $responseItem.name.startsWith('models/text') }}"}},{"type":"setKeyValue","properties":{"name":"={{$responseItem.name}}","value":"={{$responseItem.name}}","description":"={{$responseItem.description}}"}},{"type":"sort","properties":{"key":"name"}}]}}}},"routing":{"send":{"type":"body","property":"model"}},"default":"models/text-bison-001"},{"displayName":"Options","name":"options","placeholder":"Add Option","description":"Additional options to add","type":"collection","default":{},"options":[{"displayName":"Maximum Number of Tokens","name":"maxOutputTokens","default":1024,"description":"The maximum number of tokens to generate in the completion","type":"number"},{"displayName":"Sampling Temperature","name":"temperature","default":0.7,"typeOptions":{"maxValue":1,"minValue":0,"numberPrecision":1},"description":"Controls randomness: Lowering results in less random completions. As the temperature approaches zero, the model will become deterministic and repetitive.","type":"number"},{"displayName":"Top K","name":"topK","default":40,"typeOptions":{"maxValue":1,"minValue":-1,"numberPrecision":1},"description":"Used to remove \"long tail\" low probability responses. Defaults to -1, which disables it.","type":"number"},{"displayName":"Top P","name":"topP","default":0.9,"typeOptions":{"maxValue":1,"minValue":0,"numberPrecision":1},"description":"Controls diversity via nucleus sampling: 0.5 means half of all likelihood-weighted options are considered. We generally recommend altering this or temperature but not both.","type":"number"}]}],"iconUrl":"icons/@n8n/n8n-nodes-langchain/dist/nodes/llms/LmGooglePalm/google.svg"},
29
29
  {"displayName":"AWS Bedrock Chat Model","name":"@n8n/n8n-nodes-langchain.lmChatAwsBedrock","group":["transform"],"version":1,"description":"Language Model AWS Bedrock","defaults":{"name":"AWS Bedrock Chat Model"},"codex":{"categories":["AI"],"subcategories":{"AI":["Language Models","Root Nodes"],"Language Models":["Chat Models (Recommended)"]},"resources":{"primaryDocumentation":[{"url":"https://docs.n8n.io/integrations/builtin/cluster-nodes/sub-nodes/n8n-nodes-langchain.lmchatawsbedrock/"}]}},"inputs":[],"outputs":["ai_languageModel"],"outputNames":["Model"],"credentials":[{"name":"aws","required":true}],"requestDefaults":{"ignoreHttpStatusErrors":true,"baseURL":"=https://bedrock.{{$credentials?.region ?? \"eu-central-1\"}}.amazonaws.com"},"properties":[{"displayName":"This node must be connected to an AI chain. <a data-action='openSelectiveNodeCreator' data-action-parameter-creatorview='AI'>Insert one</a>","name":"notice","type":"notice","default":"","typeOptions":{"containerClass":"ndv-connection-hint-notice"}},{"displayName":"Model","name":"model","type":"options","description":"The model which will generate the completion. <a href=\"https://docs.aws.amazon.com/bedrock/latest/userguide/foundation-models.html\">Learn more</a>.","typeOptions":{"loadOptions":{"routing":{"request":{"method":"GET","url":"/foundation-models?&byOutputModality=TEXT&byInferenceType=ON_DEMAND"},"output":{"postReceive":[{"type":"rootProperty","properties":{"property":"modelSummaries"}},{"type":"setKeyValue","properties":{"name":"={{$responseItem.modelName}}","description":"={{$responseItem.modelArn}}","value":"={{$responseItem.modelId}}"}},{"type":"sort","properties":{"key":"name"}}]}}}},"routing":{"send":{"type":"body","property":"model"}},"default":""},{"displayName":"Options","name":"options","placeholder":"Add Option","description":"Additional options to add","type":"collection","default":{},"options":[{"displayName":"Maximum Number of Tokens","name":"maxTokensToSample","default":2000,"description":"The maximum number of tokens to generate in the completion","type":"number"},{"displayName":"Sampling Temperature","name":"temperature","default":0.7,"typeOptions":{"maxValue":1,"minValue":0,"numberPrecision":1},"description":"Controls randomness: Lowering results in less random completions. As the temperature approaches zero, the model will become deterministic and repetitive.","type":"number"}]}],"iconUrl":"icons/@n8n/n8n-nodes-langchain/dist/nodes/llms/LmChatAwsBedrock/bedrock.svg"},
30
30
  {"displayName":"Google PaLM Chat Model","name":"@n8n/n8n-nodes-langchain.lmChatGooglePalm","hidden":true,"group":["transform"],"version":1,"description":"Chat Model Google PaLM","defaults":{"name":"Google PaLM Chat Model"},"codex":{"categories":["AI"],"subcategories":{"AI":["Language Models","Root Nodes"],"Language Models":["Chat Models (Recommended)"]},"resources":{"primaryDocumentation":[{"url":"https://docs.n8n.io/integrations/builtin/cluster-nodes/sub-nodes/n8n-nodes-langchain.lmchatgooglepalm/"}]}},"inputs":[],"outputs":["ai_languageModel"],"outputNames":["Model"],"credentials":[{"name":"googlePalmApi","required":true}],"requestDefaults":{"ignoreHttpStatusErrors":true,"baseURL":"={{ $credentials.host }}"},"properties":[{"displayName":"Google PaLM API is <a href='https://ai.google.dev/palm_docs/deprecation' target='_blank'>deprecated</a>. Please use Google Vertex or Google Gemini nodes instead.","name":"deprecated","type":"notice","default":""},{"displayName":"Model","name":"modelName","type":"options","description":"The model which will generate the completion. <a href=\"https://developers.generativeai.google/api/rest/generativelanguage/models/list\">Learn more</a>.","typeOptions":{"loadOptions":{"routing":{"request":{"method":"GET","url":"/v1beta3/models"},"output":{"postReceive":[{"type":"rootProperty","properties":{"property":"models"}},{"type":"filter","properties":{"pass":"={{ $responseItem.name.startsWith('models/chat') }}"}},{"type":"setKeyValue","properties":{"name":"={{$responseItem.name}}","value":"={{$responseItem.name}}","description":"={{$responseItem.description}}"}},{"type":"sort","properties":{"key":"name"}}]}}}},"routing":{"send":{"type":"body","property":"model"}},"default":"models/chat-bison-001"},{"displayName":"Options","name":"options","placeholder":"Add Option","description":"Additional options to add","type":"collection","default":{},"options":[{"displayName":"Sampling Temperature","name":"temperature","default":0.7,"typeOptions":{"maxValue":1,"minValue":0,"numberPrecision":1},"description":"Controls randomness: Lowering results in less random completions. As the temperature approaches zero, the model will become deterministic and repetitive.","type":"number"},{"displayName":"Top K","name":"topK","default":40,"typeOptions":{"maxValue":1,"minValue":-1,"numberPrecision":1},"description":"Used to remove \"long tail\" low probability responses. Defaults to -1, which disables it.","type":"number"},{"displayName":"Top P","name":"topP","default":0.9,"typeOptions":{"maxValue":1,"minValue":0,"numberPrecision":1},"description":"Controls diversity via nucleus sampling: 0.5 means half of all likelihood-weighted options are considered. We generally recommend altering this or temperature but not both.","type":"number"}]}],"iconUrl":"icons/@n8n/n8n-nodes-langchain/dist/nodes/llms/LmChatGooglePalm/google.svg"},
@@ -33,7 +33,7 @@
33
33
  {"displayName":"Groq Chat Model","name":"@n8n/n8n-nodes-langchain.lmChatGroq","group":["transform"],"version":1,"description":"Language Model Groq","defaults":{"name":"Groq Chat Model"},"codex":{"categories":["AI"],"subcategories":{"AI":["Language Models","Root Nodes"],"Language Models":["Chat Models (Recommended)"]},"resources":{"primaryDocumentation":[{"url":"https://docs.n8n.io/integrations/builtin/cluster-nodes/sub-nodes/n8n-nodes-langchain.lmchatgroq/"}]}},"inputs":[],"outputs":["ai_languageModel"],"outputNames":["Model"],"credentials":[{"name":"groqApi","required":true}],"requestDefaults":{"baseURL":"https://api.groq.com/openai/v1"},"properties":[{"displayName":"This node must be connected to an AI chain. <a data-action='openSelectiveNodeCreator' data-action-parameter-creatorview='AI'>Insert one</a>","name":"notice","type":"notice","default":"","typeOptions":{"containerClass":"ndv-connection-hint-notice"}},{"displayName":"Model","name":"model","type":"options","typeOptions":{"loadOptions":{"routing":{"request":{"method":"GET","url":"/models"},"output":{"postReceive":[{"type":"rootProperty","properties":{"property":"data"}},{"type":"filter","properties":{"pass":"={{ $responseItem.active === true && $responseItem.object === \"model\" }}"}},{"type":"setKeyValue","properties":{"name":"={{$responseItem.id}}","value":"={{$responseItem.id}}"}}]}}}},"routing":{"send":{"type":"body","property":"model"}},"description":"The model which will generate the completion. <a href=\"https://console.groq.com/docs/models\">Learn more</a>.","default":"llama3-8b-8192"},{"displayName":"Options","name":"options","placeholder":"Add Option","description":"Additional options to add","type":"collection","default":{},"options":[{"displayName":"Maximum Number of Tokens","name":"maxTokensToSample","default":4096,"description":"The maximum number of tokens to generate in the completion","type":"number"},{"displayName":"Sampling Temperature","name":"temperature","default":0.7,"typeOptions":{"maxValue":1,"minValue":0,"numberPrecision":1},"description":"Controls randomness: Lowering results in less random completions. As the temperature approaches zero, the model will become deterministic and repetitive.","type":"number"}]}],"iconUrl":"icons/@n8n/n8n-nodes-langchain/dist/nodes/llms/LmChatGroq/groq.svg"},
34
34
  {"displayName":"Mistral Cloud Chat Model","name":"@n8n/n8n-nodes-langchain.lmChatMistralCloud","group":["transform"],"version":1,"description":"For advanced usage with an AI chain","defaults":{"name":"Mistral Cloud Chat Model"},"codex":{"categories":["AI"],"subcategories":{"AI":["Language Models","Root Nodes"],"Language Models":["Chat Models (Recommended)"]},"resources":{"primaryDocumentation":[{"url":"https://docs.n8n.io/integrations/builtin/cluster-nodes/sub-nodes/n8n-nodes-langchain.lmchatmistralcloud/"}]}},"inputs":[],"outputs":["ai_languageModel"],"outputNames":["Model"],"credentials":[{"name":"mistralCloudApi","required":true}],"requestDefaults":{"ignoreHttpStatusErrors":true,"baseURL":"https://api.mistral.ai/v1"},"properties":[{"displayName":"This node must be connected to an AI chain. <a data-action='openSelectiveNodeCreator' data-action-parameter-creatorview='AI'>Insert one</a>","name":"notice","type":"notice","default":"","typeOptions":{"containerClass":"ndv-connection-hint-notice"}},{"displayName":"Model","name":"model","type":"options","description":"The model which will generate the completion. <a href=\"https://docs.mistral.ai/platform/endpoints/\">Learn more</a>.","typeOptions":{"loadOptions":{"routing":{"request":{"method":"GET","url":"/models"},"output":{"postReceive":[{"type":"rootProperty","properties":{"property":"data"}},{"type":"filter","properties":{"pass":"={{ !$responseItem.id.includes('embed') }}"}},{"type":"setKeyValue","properties":{"name":"={{ $responseItem.id }}","value":"={{ $responseItem.id }}"}},{"type":"sort","properties":{"key":"name"}}]}}}},"routing":{"send":{"type":"body","property":"model"}},"default":"mistral-small"},{"displayName":"Options","name":"options","placeholder":"Add Option","description":"Additional options to add","type":"collection","default":{},"options":[{"displayName":"Maximum Number of Tokens","name":"maxTokens","default":-1,"description":"The maximum number of tokens to generate in the completion. Most models have a context length of 2048 tokens (except for the newest models, which support 32,768).","type":"number","typeOptions":{"maxValue":32768}},{"displayName":"Sampling Temperature","name":"temperature","default":0.7,"typeOptions":{"maxValue":1,"minValue":0,"numberPrecision":1},"description":"Controls randomness: Lowering results in less random completions. As the temperature approaches zero, the model will become deterministic and repetitive.","type":"number"},{"displayName":"Max Retries","name":"maxRetries","default":2,"description":"Maximum number of retries to attempt","type":"number"},{"displayName":"Top P","name":"topP","default":1,"typeOptions":{"maxValue":1,"minValue":0,"numberPrecision":1},"description":"Controls diversity via nucleus sampling: 0.5 means half of all likelihood-weighted options are considered. We generally recommend altering this or temperature but not both.","type":"number"},{"displayName":"Enable Safe Mode","name":"safeMode","default":false,"type":"boolean","description":"Whether to inject a safety prompt before all conversations"},{"displayName":"Random Seed","name":"randomSeed","type":"number","description":"The seed to use for random sampling. If set, different calls will generate deterministic results."}]}],"iconUrl":"icons/@n8n/n8n-nodes-langchain/dist/nodes/llms/LmChatMistralCloud/mistral.svg"},
35
35
  {"displayName":"Ollama Chat Model","name":"@n8n/n8n-nodes-langchain.lmChatOllama","group":["transform"],"version":1,"description":"Language Model Ollama","defaults":{"name":"Ollama Chat Model"},"codex":{"categories":["AI"],"subcategories":{"AI":["Language Models","Root Nodes"],"Language Models":["Chat Models (Recommended)"]},"resources":{"primaryDocumentation":[{"url":"https://docs.n8n.io/integrations/builtin/cluster-nodes/sub-nodes/n8n-nodes-langchain.lmchatollama/"}]}},"inputs":[],"outputs":["ai_languageModel"],"outputNames":["Model"],"credentials":[{"name":"ollamaApi","required":true}],"requestDefaults":{"ignoreHttpStatusErrors":true,"baseURL":"={{ $credentials.baseUrl.replace(new RegExp(\"/$\"), \"\") }}"},"properties":[{"displayName":"This node must be connected to an AI chain. <a data-action='openSelectiveNodeCreator' data-action-parameter-creatorview='AI'>Insert one</a>","name":"notice","type":"notice","default":"","typeOptions":{"containerClass":"ndv-connection-hint-notice"}},{"displayName":"Model","name":"model","type":"options","default":"llama2","description":"The model which will generate the completion. To download models, visit <a href=\"https://ollama.ai/library\">Ollama Models Library</a>.","typeOptions":{"loadOptions":{"routing":{"request":{"method":"GET","url":"/api/tags"},"output":{"postReceive":[{"type":"rootProperty","properties":{"property":"models"}},{"type":"setKeyValue","properties":{"name":"={{$responseItem.name}}","value":"={{$responseItem.name}}"}},{"type":"sort","properties":{"key":"name"}}]}}}},"routing":{"send":{"type":"body","property":"model"}},"required":true},{"displayName":"Options","name":"options","placeholder":"Add Option","description":"Additional options to add","type":"collection","default":{},"options":[{"displayName":"Sampling Temperature","name":"temperature","default":0.7,"typeOptions":{"maxValue":1,"minValue":0,"numberPrecision":1},"description":"Controls the randomness of the generated text. Lower values make the output more focused and deterministic, while higher values make it more diverse and random.","type":"number"},{"displayName":"Top K","name":"topK","default":-1,"typeOptions":{"maxValue":100,"minValue":-1,"numberPrecision":1},"description":"Limits the number of highest probability vocabulary tokens to consider at each step. A higher value increases diversity but may reduce coherence. Set to -1 to disable.","type":"number"},{"displayName":"Top P","name":"topP","default":1,"typeOptions":{"maxValue":1,"minValue":0,"numberPrecision":1},"description":"Chooses from the smallest possible set of tokens whose cumulative probability exceeds the probability top_p. Helps generate more human-like text by reducing repetitions.","type":"number"},{"displayName":"Frequency Penalty","name":"frequencyPenalty","type":"number","default":0,"typeOptions":{"minValue":0},"description":"Adjusts the penalty for tokens that have already appeared in the generated text. Higher values discourage repetition."},{"displayName":"Keep Alive","name":"keepAlive","type":"string","default":"5m","description":"Specifies the duration to keep the loaded model in memory after use. Useful for frequently used models. Format: 1h30m (1 hour 30 minutes)."},{"displayName":"Low VRAM Mode","name":"lowVram","type":"boolean","default":false,"description":"Whether to Activate low VRAM mode, which reduces memory usage at the cost of slower generation speed. Useful for GPUs with limited memory."},{"displayName":"Main GPU ID","name":"mainGpu","type":"number","default":0,"description":"Specifies the ID of the GPU to use for the main computation. Only change this if you have multiple GPUs."},{"displayName":"Context Batch Size","name":"numBatch","type":"number","default":512,"description":"Sets the batch size for prompt processing. Larger batch sizes may improve generation speed but increase memory usage."},{"displayName":"Context Length","name":"numCtx","type":"number","default":2048,"description":"The maximum number of tokens to use as context for generating the next token. Smaller values reduce memory usage, while larger values provide more context to the model."},{"displayName":"Number of GPUs","name":"numGpu","type":"number","default":-1,"description":"Specifies the number of GPUs to use for parallel processing. Set to -1 for auto-detection."},{"displayName":"Max Tokens to Generate","name":"numPredict","type":"number","default":-1,"description":"The maximum number of tokens to generate. Set to -1 for no limit. Be cautious when setting this to a large value, as it can lead to very long outputs."},{"displayName":"Number of CPU Threads","name":"numThread","type":"number","default":0,"description":"Specifies the number of CPU threads to use for processing. Set to 0 for auto-detection."},{"displayName":"Penalize Newlines","name":"penalizeNewline","type":"boolean","default":true,"description":"Whether the model will be less likely to generate newline characters, encouraging longer continuous sequences of text"},{"displayName":"Presence Penalty","name":"presencePenalty","type":"number","default":0,"description":"Adjusts the penalty for tokens based on their presence in the generated text so far. Positive values penalize tokens that have already appeared, encouraging diversity."},{"displayName":"Repetition Penalty","name":"repeatPenalty","type":"number","default":1,"description":"Adjusts the penalty factor for repeated tokens. Higher values more strongly discourage repetition. Set to 1.0 to disable repetition penalty."},{"displayName":"Use Memory Locking","name":"useMLock","type":"boolean","default":false,"description":"Whether to lock the model in memory to prevent swapping. This can improve performance but requires sufficient available memory."},{"displayName":"Use Memory Mapping","name":"useMMap","type":"boolean","default":true,"description":"Whether to use memory mapping for loading the model. This can reduce memory usage but may impact performance. Recommended to keep enabled."},{"displayName":"Load Vocabulary Only","name":"vocabOnly","type":"boolean","default":false,"description":"Whether to only load the model vocabulary without the weights. Useful for quickly testing tokenization."},{"displayName":"Output Format","name":"format","type":"options","options":[{"name":"Default","value":"default"},{"name":"JSON","value":"json"}],"default":"default","description":"Specifies the format of the API response"}]}],"iconUrl":"icons/@n8n/n8n-nodes-langchain/dist/nodes/llms/LMChatOllama/ollama.svg"},
36
- {"displayName":"OpenAI Chat Model","name":"@n8n/n8n-nodes-langchain.lmChatOpenAi","group":["transform"],"version":1,"description":"For advanced usage with an AI chain","defaults":{"name":"OpenAI Chat Model"},"codex":{"categories":["AI"],"subcategories":{"AI":["Language Models","Root Nodes"],"Language Models":["Chat Models (Recommended)"]},"resources":{"primaryDocumentation":[{"url":"https://docs.n8n.io/integrations/builtin/cluster-nodes/sub-nodes/n8n-nodes-langchain.lmchatopenai/"}]}},"inputs":[],"outputs":["ai_languageModel"],"outputNames":["Model"],"credentials":[{"name":"openAiApi","required":true}],"requestDefaults":{"ignoreHttpStatusErrors":true,"baseURL":"={{ $parameter.options?.baseURL?.split(\"/\").slice(0,-1).join(\"/\") || \"https://api.openai.com\" }}"},"properties":[{"displayName":"This node must be connected to an AI chain. <a data-action='openSelectiveNodeCreator' data-action-parameter-creatorview='AI'>Insert one</a>","name":"notice","type":"notice","default":"","typeOptions":{"containerClass":"ndv-connection-hint-notice"}},{"displayName":"If using JSON response format, you must include word \"json\" in the prompt in your chain or agent. Also, make sure to select latest models released post November 2023.","name":"notice","type":"notice","default":"","displayOptions":{"show":{"/options.responseFormat":["json_object"]}}},{"displayName":"Model","name":"model","type":"options","description":"The model which will generate the completion. <a href=\"https://beta.openai.com/docs/models/overview\">Learn more</a>.","typeOptions":{"loadOptions":{"routing":{"request":{"method":"GET","url":"={{ $parameter.options?.baseURL?.split(\"/\").slice(-1).pop() || \"v1\" }}/models"},"output":{"postReceive":[{"type":"rootProperty","properties":{"property":"data"}},{"type":"filter","properties":{"pass":"={{\n\t\t\t\t\t\t\t\t\t\t\t\t($parameter.options?.baseURL && !$parameter.options?.baseURL?.includes('api.openai.com')) ||\n\t\t\t\t\t\t\t\t\t\t\t\t$responseItem.id.startsWith('ft:') ||\n\t\t\t\t\t\t\t\t\t\t\t\t($responseItem.id.startsWith('gpt-') && !$responseItem.id.includes('instruct'))\n\t\t\t\t\t\t\t\t\t\t\t}}"}},{"type":"setKeyValue","properties":{"name":"={{$responseItem.id}}","value":"={{$responseItem.id}}"}},{"type":"sort","properties":{"key":"name"}}]}}}},"routing":{"send":{"type":"body","property":"model"}},"default":"gpt-3.5-turbo"},{"displayName":"When using non-OpenAI models via \"Base URL\" override, not all models might be chat-compatible or support other features, like tools calling or JSON response format","name":"notice","type":"notice","default":"","displayOptions":{"show":{"/options.baseURL":[{"_cnd":{"exists":true}}]}}},{"displayName":"Options","name":"options","placeholder":"Add Option","description":"Additional options to add","type":"collection","default":{},"options":[{"displayName":"Base URL","name":"baseURL","default":"https://api.openai.com/v1","description":"Override the default base URL for the API","type":"string"},{"displayName":"Frequency Penalty","name":"frequencyPenalty","default":0,"typeOptions":{"maxValue":2,"minValue":-2,"numberPrecision":1},"description":"Positive values penalize new tokens based on their existing frequency in the text so far, decreasing the model's likelihood to repeat the same line verbatim","type":"number"},{"displayName":"Maximum Number of Tokens","name":"maxTokens","default":-1,"description":"The maximum number of tokens to generate in the completion. Most models have a context length of 2048 tokens (except for the newest models, which support 32,768).","type":"number","typeOptions":{"maxValue":32768}},{"displayName":"Response Format","name":"responseFormat","default":"text","type":"options","options":[{"name":"Text","value":"text","description":"Regular text response"},{"name":"JSON","value":"json_object","description":"Enables JSON mode, which should guarantee the message the model generates is valid JSON"}]},{"displayName":"Presence Penalty","name":"presencePenalty","default":0,"typeOptions":{"maxValue":2,"minValue":-2,"numberPrecision":1},"description":"Positive values penalize new tokens based on whether they appear in the text so far, increasing the model's likelihood to talk about new topics","type":"number"},{"displayName":"Sampling Temperature","name":"temperature","default":0.7,"typeOptions":{"maxValue":1,"minValue":0,"numberPrecision":1},"description":"Controls randomness: Lowering results in less random completions. As the temperature approaches zero, the model will become deterministic and repetitive.","type":"number"},{"displayName":"Timeout","name":"timeout","default":60000,"description":"Maximum amount of time a request is allowed to take in milliseconds","type":"number"},{"displayName":"Max Retries","name":"maxRetries","default":2,"description":"Maximum number of retries to attempt","type":"number"},{"displayName":"Top P","name":"topP","default":1,"typeOptions":{"maxValue":1,"minValue":0,"numberPrecision":1},"description":"Controls diversity via nucleus sampling: 0.5 means half of all likelihood-weighted options are considered. We generally recommend altering this or temperature but not both.","type":"number"}]}],"iconUrl":{"light":"icons/@n8n/n8n-nodes-langchain/dist/nodes/llms/LMChatOpenAi/openAiLight.svg","dark":"icons/@n8n/n8n-nodes-langchain/dist/nodes/llms/LMChatOpenAi/openAiLight.dark.svg"}},
36
+ {"displayName":"OpenAI Chat Model","name":"@n8n/n8n-nodes-langchain.lmChatOpenAi","group":["transform"],"version":1,"description":"For advanced usage with an AI chain","defaults":{"name":"OpenAI Chat Model"},"codex":{"categories":["AI"],"subcategories":{"AI":["Language Models","Root Nodes"],"Language Models":["Chat Models (Recommended)"]},"resources":{"primaryDocumentation":[{"url":"https://docs.n8n.io/integrations/builtin/cluster-nodes/sub-nodes/n8n-nodes-langchain.lmchatopenai/"}]}},"inputs":[],"outputs":["ai_languageModel"],"outputNames":["Model"],"credentials":[{"name":"openAiApi","required":true}],"requestDefaults":{"ignoreHttpStatusErrors":true,"baseURL":"={{ $parameter.options?.baseURL?.split(\"/\").slice(0,-1).join(\"/\") || \"https://api.openai.com\" }}"},"properties":[{"displayName":"This node must be connected to an AI chain. <a data-action='openSelectiveNodeCreator' data-action-parameter-creatorview='AI'>Insert one</a>","name":"notice","type":"notice","default":"","typeOptions":{"containerClass":"ndv-connection-hint-notice"}},{"displayName":"If using JSON response format, you must include word \"json\" in the prompt in your chain or agent. Also, make sure to select latest models released post November 2023.","name":"notice","type":"notice","default":"","displayOptions":{"show":{"/options.responseFormat":["json_object"]}}},{"displayName":"Model","name":"model","type":"options","description":"The model which will generate the completion. <a href=\"https://beta.openai.com/docs/models/overview\">Learn more</a>.","typeOptions":{"loadOptions":{"routing":{"request":{"method":"GET","url":"={{ $parameter.options?.baseURL?.split(\"/\").slice(-1).pop() || \"v1\" }}/models"},"output":{"postReceive":[{"type":"rootProperty","properties":{"property":"data"}},{"type":"filter","properties":{"pass":"={{\n\t\t\t\t\t\t\t\t\t\t\t\t($parameter.options?.baseURL && !$parameter.options?.baseURL?.includes('api.openai.com')) ||\n\t\t\t\t\t\t\t\t\t\t\t\t$responseItem.id.startsWith('ft:') ||\n\t\t\t\t\t\t\t\t\t\t\t\t$responseItem.id.startsWith('o1') ||\n\t\t\t\t\t\t\t\t\t\t\t\t($responseItem.id.startsWith('gpt-') && !$responseItem.id.includes('instruct'))\n\t\t\t\t\t\t\t\t\t\t\t}}"}},{"type":"setKeyValue","properties":{"name":"={{$responseItem.id}}","value":"={{$responseItem.id}}"}},{"type":"sort","properties":{"key":"name"}}]}}}},"routing":{"send":{"type":"body","property":"model"}},"default":"gpt-3.5-turbo"},{"displayName":"When using non-OpenAI models via \"Base URL\" override, not all models might be chat-compatible or support other features, like tools calling or JSON response format","name":"notice","type":"notice","default":"","displayOptions":{"show":{"/options.baseURL":[{"_cnd":{"exists":true}}]}}},{"displayName":"Options","name":"options","placeholder":"Add Option","description":"Additional options to add","type":"collection","default":{},"options":[{"displayName":"Base URL","name":"baseURL","default":"https://api.openai.com/v1","description":"Override the default base URL for the API","type":"string"},{"displayName":"Frequency Penalty","name":"frequencyPenalty","default":0,"typeOptions":{"maxValue":2,"minValue":-2,"numberPrecision":1},"description":"Positive values penalize new tokens based on their existing frequency in the text so far, decreasing the model's likelihood to repeat the same line verbatim","type":"number"},{"displayName":"Maximum Number of Tokens","name":"maxTokens","default":-1,"description":"The maximum number of tokens to generate in the completion. Most models have a context length of 2048 tokens (except for the newest models, which support 32,768).","type":"number","typeOptions":{"maxValue":32768}},{"displayName":"Response Format","name":"responseFormat","default":"text","type":"options","options":[{"name":"Text","value":"text","description":"Regular text response"},{"name":"JSON","value":"json_object","description":"Enables JSON mode, which should guarantee the message the model generates is valid JSON"}]},{"displayName":"Presence Penalty","name":"presencePenalty","default":0,"typeOptions":{"maxValue":2,"minValue":-2,"numberPrecision":1},"description":"Positive values penalize new tokens based on whether they appear in the text so far, increasing the model's likelihood to talk about new topics","type":"number"},{"displayName":"Sampling Temperature","name":"temperature","default":0.7,"typeOptions":{"maxValue":1,"minValue":0,"numberPrecision":1},"description":"Controls randomness: Lowering results in less random completions. As the temperature approaches zero, the model will become deterministic and repetitive.","type":"number"},{"displayName":"Timeout","name":"timeout","default":60000,"description":"Maximum amount of time a request is allowed to take in milliseconds","type":"number"},{"displayName":"Max Retries","name":"maxRetries","default":2,"description":"Maximum number of retries to attempt","type":"number"},{"displayName":"Top P","name":"topP","default":1,"typeOptions":{"maxValue":1,"minValue":0,"numberPrecision":1},"description":"Controls diversity via nucleus sampling: 0.5 means half of all likelihood-weighted options are considered. We generally recommend altering this or temperature but not both.","type":"number"}]}],"iconUrl":{"light":"icons/@n8n/n8n-nodes-langchain/dist/nodes/llms/LMChatOpenAi/openAiLight.svg","dark":"icons/@n8n/n8n-nodes-langchain/dist/nodes/llms/LMChatOpenAi/openAiLight.dark.svg"}},
37
37
  {"displayName":"OpenAI Model","name":"@n8n/n8n-nodes-langchain.lmOpenAi","hidden":true,"group":["transform"],"version":1,"description":"For advanced usage with an AI chain","defaults":{"name":"OpenAI Model"},"codex":{"categories":["AI"],"subcategories":{"AI":["Language Models","Root Nodes"],"Language Models":["Text Completion Models"]},"resources":{"primaryDocumentation":[{"url":"https://docs.n8n.io/integrations/builtin/cluster-nodes/sub-nodes/n8n-nodes-langchain.lmopenai/"}]}},"inputs":[],"outputs":["ai_languageModel"],"outputNames":["Model"],"credentials":[{"name":"openAiApi","required":true}],"requestDefaults":{"ignoreHttpStatusErrors":true,"baseURL":"={{ $parameter.options?.baseURL?.split(\"/\").slice(0,-1).join(\"/\") || \"https://api.openai.com\" }}"},"properties":[{"displayName":"This node is using OpenAI completions which are now deprecated. Please use the OpenAI Chat Model node instead.","name":"deprecated","type":"notice","default":""},{"displayName":"Model","name":"model","type":"resourceLocator","default":{"mode":"list","value":"gpt-3.5-turbo-instruct"},"required":true,"description":"The model which will generate the completion. <a href=\"https://beta.openai.com/docs/models/overview\">Learn more</a>.","modes":[{"displayName":"From List","name":"list","type":"list","typeOptions":{"searchListMethod":"openAiModelSearch"}},{"displayName":"ID","name":"id","type":"string"}],"routing":{"send":{"type":"body","property":"model","value":"={{$parameter.model.value}}"}}},{"displayName":"When using non OpenAI models via Base URL override, not all models might be chat-compatible or support other features, like tools calling or JSON response format.","name":"notice","type":"notice","default":"","displayOptions":{"show":{"/options.baseURL":[{"_cnd":{"exists":true}}]}}},{"displayName":"Options","name":"options","placeholder":"Add Option","description":"Additional options to add","type":"collection","default":{},"options":[{"displayName":"Base URL","name":"baseURL","default":"https://api.openai.com/v1","description":"Override the default base URL for the API","type":"string"},{"displayName":"Frequency Penalty","name":"frequencyPenalty","default":0,"typeOptions":{"maxValue":2,"minValue":-2,"numberPrecision":1},"description":"Positive values penalize new tokens based on their existing frequency in the text so far, decreasing the model's likelihood to repeat the same line verbatim","type":"number"},{"displayName":"Maximum Number of Tokens","name":"maxTokens","default":-1,"description":"The maximum number of tokens to generate in the completion. Most models have a context length of 2048 tokens (except for the newest models, which support 32,768).","type":"number","typeOptions":{"maxValue":32768}},{"displayName":"Presence Penalty","name":"presencePenalty","default":0,"typeOptions":{"maxValue":2,"minValue":-2,"numberPrecision":1},"description":"Positive values penalize new tokens based on whether they appear in the text so far, increasing the model's likelihood to talk about new topics","type":"number"},{"displayName":"Sampling Temperature","name":"temperature","default":0.7,"typeOptions":{"maxValue":1,"minValue":0,"numberPrecision":1},"description":"Controls randomness: Lowering results in less random completions. As the temperature approaches zero, the model will become deterministic and repetitive.","type":"number"},{"displayName":"Timeout","name":"timeout","default":60000,"description":"Maximum amount of time a request is allowed to take in milliseconds","type":"number"},{"displayName":"Max Retries","name":"maxRetries","default":2,"description":"Maximum number of retries to attempt","type":"number"},{"displayName":"Top P","name":"topP","default":1,"typeOptions":{"maxValue":1,"minValue":0,"numberPrecision":1},"description":"Controls diversity via nucleus sampling: 0.5 means half of all likelihood-weighted options are considered. We generally recommend altering this or temperature but not both.","type":"number"}]}],"iconUrl":{"light":"icons/@n8n/n8n-nodes-langchain/dist/nodes/llms/LMOpenAi/openAiLight.svg","dark":"icons/@n8n/n8n-nodes-langchain/dist/nodes/llms/LMOpenAi/openAiLight.dark.svg"}},
38
38
  {"displayName":"Cohere Model","name":"@n8n/n8n-nodes-langchain.lmCohere","group":["transform"],"version":1,"description":"Language Model Cohere","defaults":{"name":"Cohere Model"},"codex":{"categories":["AI"],"subcategories":{"AI":["Language Models","Root Nodes"],"Language Models":["Text Completion Models"]},"resources":{"primaryDocumentation":[{"url":"https://docs.n8n.io/integrations/builtin/cluster-nodes/sub-nodes/n8n-nodes-langchain.lmcohere/"}]}},"inputs":[],"outputs":["ai_languageModel"],"outputNames":["Model"],"credentials":[{"name":"cohereApi","required":true}],"properties":[{"displayName":"This node must be connected to an AI chain. <a data-action='openSelectiveNodeCreator' data-action-parameter-creatorview='AI'>Insert one</a>","name":"notice","type":"notice","default":"","typeOptions":{"containerClass":"ndv-connection-hint-notice"}},{"displayName":"Options","name":"options","placeholder":"Add Option","description":"Additional options to add","type":"collection","default":{},"options":[{"displayName":"Maximum Number of Tokens","name":"maxTokens","default":250,"description":"The maximum number of tokens to generate in the completion. Most models have a context length of 2048 tokens (except for the newest models, which support 32,768).","type":"number","typeOptions":{"maxValue":32768}},{"displayName":"Model","name":"model","type":"string","description":"The name of the model to use","default":""},{"displayName":"Sampling Temperature","name":"temperature","default":0,"typeOptions":{"maxValue":1,"minValue":0,"numberPrecision":1},"description":"Controls randomness: Lowering results in less random completions. As the temperature approaches zero, the model will become deterministic and repetitive.","type":"number"}]}],"iconUrl":{"light":"icons/@n8n/n8n-nodes-langchain/dist/nodes/llms/LMCohere/cohere.svg","dark":"icons/@n8n/n8n-nodes-langchain/dist/nodes/llms/LMCohere/cohere.dark.svg"}},
39
39
  {"displayName":"Ollama Model","name":"@n8n/n8n-nodes-langchain.lmOllama","group":["transform"],"version":1,"description":"Language Model Ollama","defaults":{"name":"Ollama Model"},"codex":{"categories":["AI"],"subcategories":{"AI":["Language Models","Root Nodes"],"Language Models":["Text Completion Models"]},"resources":{"primaryDocumentation":[{"url":"https://docs.n8n.io/integrations/builtin/cluster-nodes/sub-nodes/n8n-nodes-langchain.lmollama/"}]}},"inputs":[],"outputs":["ai_languageModel"],"outputNames":["Model"],"credentials":[{"name":"ollamaApi","required":true}],"requestDefaults":{"ignoreHttpStatusErrors":true,"baseURL":"={{ $credentials.baseUrl.replace(new RegExp(\"/$\"), \"\") }}"},"properties":[{"displayName":"This node must be connected to an AI chain. <a data-action='openSelectiveNodeCreator' data-action-parameter-creatorview='AI'>Insert one</a>","name":"notice","type":"notice","default":"","typeOptions":{"containerClass":"ndv-connection-hint-notice"}},{"displayName":"Model","name":"model","type":"options","default":"llama2","description":"The model which will generate the completion. To download models, visit <a href=\"https://ollama.ai/library\">Ollama Models Library</a>.","typeOptions":{"loadOptions":{"routing":{"request":{"method":"GET","url":"/api/tags"},"output":{"postReceive":[{"type":"rootProperty","properties":{"property":"models"}},{"type":"setKeyValue","properties":{"name":"={{$responseItem.name}}","value":"={{$responseItem.name}}"}},{"type":"sort","properties":{"key":"name"}}]}}}},"routing":{"send":{"type":"body","property":"model"}},"required":true},{"displayName":"Options","name":"options","placeholder":"Add Option","description":"Additional options to add","type":"collection","default":{},"options":[{"displayName":"Sampling Temperature","name":"temperature","default":0.7,"typeOptions":{"maxValue":1,"minValue":0,"numberPrecision":1},"description":"Controls the randomness of the generated text. Lower values make the output more focused and deterministic, while higher values make it more diverse and random.","type":"number"},{"displayName":"Top K","name":"topK","default":-1,"typeOptions":{"maxValue":100,"minValue":-1,"numberPrecision":1},"description":"Limits the number of highest probability vocabulary tokens to consider at each step. A higher value increases diversity but may reduce coherence. Set to -1 to disable.","type":"number"},{"displayName":"Top P","name":"topP","default":1,"typeOptions":{"maxValue":1,"minValue":0,"numberPrecision":1},"description":"Chooses from the smallest possible set of tokens whose cumulative probability exceeds the probability top_p. Helps generate more human-like text by reducing repetitions.","type":"number"},{"displayName":"Frequency Penalty","name":"frequencyPenalty","type":"number","default":0,"typeOptions":{"minValue":0},"description":"Adjusts the penalty for tokens that have already appeared in the generated text. Higher values discourage repetition."},{"displayName":"Keep Alive","name":"keepAlive","type":"string","default":"5m","description":"Specifies the duration to keep the loaded model in memory after use. Useful for frequently used models. Format: 1h30m (1 hour 30 minutes)."},{"displayName":"Low VRAM Mode","name":"lowVram","type":"boolean","default":false,"description":"Whether to Activate low VRAM mode, which reduces memory usage at the cost of slower generation speed. Useful for GPUs with limited memory."},{"displayName":"Main GPU ID","name":"mainGpu","type":"number","default":0,"description":"Specifies the ID of the GPU to use for the main computation. Only change this if you have multiple GPUs."},{"displayName":"Context Batch Size","name":"numBatch","type":"number","default":512,"description":"Sets the batch size for prompt processing. Larger batch sizes may improve generation speed but increase memory usage."},{"displayName":"Context Length","name":"numCtx","type":"number","default":2048,"description":"The maximum number of tokens to use as context for generating the next token. Smaller values reduce memory usage, while larger values provide more context to the model."},{"displayName":"Number of GPUs","name":"numGpu","type":"number","default":-1,"description":"Specifies the number of GPUs to use for parallel processing. Set to -1 for auto-detection."},{"displayName":"Max Tokens to Generate","name":"numPredict","type":"number","default":-1,"description":"The maximum number of tokens to generate. Set to -1 for no limit. Be cautious when setting this to a large value, as it can lead to very long outputs."},{"displayName":"Number of CPU Threads","name":"numThread","type":"number","default":0,"description":"Specifies the number of CPU threads to use for processing. Set to 0 for auto-detection."},{"displayName":"Penalize Newlines","name":"penalizeNewline","type":"boolean","default":true,"description":"Whether the model will be less likely to generate newline characters, encouraging longer continuous sequences of text"},{"displayName":"Presence Penalty","name":"presencePenalty","type":"number","default":0,"description":"Adjusts the penalty for tokens based on their presence in the generated text so far. Positive values penalize tokens that have already appeared, encouraging diversity."},{"displayName":"Repetition Penalty","name":"repeatPenalty","type":"number","default":1,"description":"Adjusts the penalty factor for repeated tokens. Higher values more strongly discourage repetition. Set to 1.0 to disable repetition penalty."},{"displayName":"Use Memory Locking","name":"useMLock","type":"boolean","default":false,"description":"Whether to lock the model in memory to prevent swapping. This can improve performance but requires sufficient available memory."},{"displayName":"Use Memory Mapping","name":"useMMap","type":"boolean","default":true,"description":"Whether to use memory mapping for loading the model. This can reduce memory usage but may impact performance. Recommended to keep enabled."},{"displayName":"Load Vocabulary Only","name":"vocabOnly","type":"boolean","default":false,"description":"Whether to only load the model vocabulary without the weights. Useful for quickly testing tokenization."},{"displayName":"Output Format","name":"format","type":"options","options":[{"name":"Default","value":"default"},{"name":"JSON","value":"json"}],"default":"default","description":"Specifies the format of the API response"}]}],"iconUrl":"icons/@n8n/n8n-nodes-langchain/dist/nodes/llms/LMOllama/ollama.svg"},
@@ -1,4 +1,4 @@
1
- import type { EventNamesAiNodesType, IDataObject, IExecuteFunctions, IWebhookFunctions } from 'n8n-workflow';
1
+ import type { AiEvent, IDataObject, IExecuteFunctions, IWebhookFunctions } from 'n8n-workflow';
2
2
  import type { BaseChatModel } from '@langchain/core/language_models/chat_models';
3
3
  import type { BaseOutputParser } from '@langchain/core/output_parsers';
4
4
  import type { BaseMessage } from '@langchain/core/messages';
@@ -18,6 +18,6 @@ export declare function getPromptInputByType(options: {
18
18
  inputKey: string;
19
19
  }): string;
20
20
  export declare function getSessionId(ctx: IExecuteFunctions | IWebhookFunctions, itemIndex: number, selectorKey?: string, autoSelect?: string, customKey?: string): string;
21
- export declare function logAiEvent(executeFunctions: IExecuteFunctions, event: EventNamesAiNodesType, data?: IDataObject): Promise<void>;
21
+ export declare function logAiEvent(executeFunctions: IExecuteFunctions, event: AiEvent, data?: IDataObject): Promise<void>;
22
22
  export declare function serializeChatHistory(chatHistory: BaseMessage[]): string;
23
23
  export declare const getConnectedTools: (ctx: IExecuteFunctions, enforceUniqueNames: boolean, convertStructuredTool?: boolean) => Promise<Tool[]>;
@@ -1 +1 @@
1
- {"version":3,"file":"helpers.js","sourceRoot":"","sources":["../../utils/helpers.ts"],"names":[],"mappings":";;;AA0BA,4DAyBC;AAED,4CAEC;AAED,4DAEC;AAED,wCAIC;AAED,0CAIC;AAED,4DAaC;AAED,oDAwBC;AAED,oCAuCC;AAED,gCAUC;AAED,oDAYC;AAnLD,+CAAqF;AAcrF,uCAAoC;AAEpC,SAAS,UAAU,CAAI,GAAY,EAAE,GAAG,WAAmC;IAC1E,OAAO,WAAW,CAAC,KAAK,CACvB,CAAC,UAAU,EAAE,EAAE,CACd,OAAO,GAAG,KAAK,QAAQ;QACvB,GAAG,KAAK,IAAI;QACZ,UAAU,IAAI,GAAG;QACjB,OAAQ,GAAwC,CAAC,UAAU,CAAC,KAAK,UAAU,CAC5E,CAAC;AACH,CAAC;AAED,SAAgB,wBAAwB,CACvC,GAAsB,EACtB,SAAiB;IAEjB,MAAM,OAAO,GAAG,GAAG,CAAC,gBAAgB,CAAC,SAAS,EAAE,SAAS,EAAE,EAAE,CAAC,CAAC;IAE/D,IAAI,OAAO,CAAC,QAAQ,EAAE,CAAC;QACtB,MAAM,EAAE,cAAc,EAAE,QAAQ,EAAE,GAAG,OAAO,CAAC,QAK5C,CAAC;QACF,IAAI,QAAQ,CAAC,MAAM,GAAG,CAAC,EAAE,CAAC;YACzB,OAAO,QAAQ,CAAC,MAAM,CAAC,CAAC,GAAG,EAAE,EAAE,IAAI,EAAE,KAAK,EAAE,EAAE,EAAE,CAAC,CAAC,EAAE,GAAG,GAAG,EAAE,CAAC,IAAI,CAAC,EAAE,KAAK,EAAE,CAAC,EAAE,EAAE,CAAC,CAAC;QACnF,CAAC;IACF,CAAC;IAED,IAAI,OAAO,CAAC,gBAAgB,EAAE,CAAC;QAC9B,OAAO,GAAG,CAAC,gBAAgB,CAAC,0BAA0B,EAAE,SAAS,EAAE,EAAE,EAAE;YACtE,UAAU,EAAE,QAAQ;SACpB,CAA0B,CAAC;IAC7B,CAAC;IAED,OAAO,SAAS,CAAC;AAClB,CAAC;AAED,SAAgB,gBAAgB,CAAC,GAAY;IAC5C,OAAO,UAAU,CAAiB,GAAG,EAAE,qBAAqB,EAAE,aAAa,CAAC,CAAC;AAC9E,CAAC;AAED,SAAgB,wBAAwB,CAAC,GAAY;IACpD,OAAO,UAAU,CAAyB,GAAG,EAAE,aAAa,EAAE,YAAY,CAAC,CAAC;AAC7E,CAAC;AAED,SAAgB,cAAc,CAAC,KAAc;IAC5C,MAAM,SAAS,GAAI,KAAiB,EAAE,YAAY,IAAI,EAAE,CAAC;IAEzD,OAAO,SAAS,CAAC,QAAQ,CAAC,aAAa,CAAC,CAAC;AAC1C,CAAC;AAED,SAAgB,eAAe,CAAC,KAAc;IAC7C,MAAM,SAAS,GAAI,KAAc,EAAE,YAAY,IAAI,EAAE,CAAC;IAEtD,OAAO,SAAS,CAAC,QAAQ,CAAC,OAAO,CAAC,CAAC;AACpC,CAAC;AAEM,KAAK,UAAU,wBAAwB,CAC7C,GAAsB;IAEtB,IAAI,aAAa,GAAuB,EAAE,CAAC;IAE3C,IAAI,GAAG,CAAC,gBAAgB,CAAC,iBAAiB,EAAE,CAAC,EAAE,IAAI,CAAC,KAAK,IAAI,EAAE,CAAC;QAC/D,aAAa,GAAG,CAAC,MAAM,GAAG,CAAC,sBAAsB,oBAEhD,CAAC,CACD,CAAuB,CAAC;IAC1B,CAAC;IAED,OAAO,aAAa,CAAC;AACtB,CAAC;AAED,SAAgB,oBAAoB,CAAC,OAKpC;IACA,MAAM,EAAE,GAAG,EAAE,CAAC,EAAE,aAAa,EAAE,QAAQ,EAAE,GAAG,OAAO,CAAC;IACpD,MAAM,MAAM,GAAG,GAAG,CAAC,gBAAgB,CAAC,aAAa,EAAE,CAAC,CAAW,CAAC;IAEhE,IAAI,KAAK,CAAC;IACV,IAAI,MAAM,KAAK,MAAM,EAAE,CAAC;QACvB,KAAK,GAAG,GAAG,CAAC,kBAAkB,CAAC,0BAA0B,EAAE,CAAC,CAAW,CAAC;IACzE,CAAC;SAAM,CAAC;QACP,KAAK,GAAG,GAAG,CAAC,gBAAgB,CAAC,QAAQ,EAAE,CAAC,CAAW,CAAC;IACrD,CAAC;IAED,IAAI,KAAK,KAAK,SAAS,EAAE,CAAC;QACzB,MAAM,IAAI,iCAAkB,CAAC,GAAG,CAAC,OAAO,EAAE,EAAE,qBAAqB,EAAE;YAClE,WAAW,EACV,qKAAqK;SACtK,CAAC,CAAC;IACJ,CAAC;IAED,OAAO,KAAK,CAAC;AACd,CAAC;AAED,SAAgB,YAAY,CAC3B,GAA0C,EAC1C,SAAiB,EACjB,WAAW,GAAG,eAAe,EAC7B,UAAU,GAAG,WAAW,EACxB,SAAS,GAAG,YAAY;IAExB,IAAI,SAAS,GAAG,EAAE,CAAC;IACnB,MAAM,YAAY,GAAG,GAAG,CAAC,gBAAgB,CAAC,WAAW,EAAE,SAAS,CAAW,CAAC;IAE5E,IAAI,YAAY,KAAK,UAAU,EAAE,CAAC;QAGjC,IAAI,aAAa,IAAI,GAAG,EAAE,CAAC;YAC1B,MAAM,QAAQ,GAAG,GAAG,CAAC,WAAW,EAAE,IAAI,EAAE,CAAC;YACzC,SAAS,GAAG,QAAQ,CAAC,SAAmB,CAAC;QAC1C,CAAC;aAAM,CAAC;YACP,SAAS,GAAG,GAAG,CAAC,kBAAkB,CAAC,uBAAuB,EAAE,SAAS,CAAW,CAAC;QAClF,CAAC;QAED,IAAI,SAAS,KAAK,EAAE,IAAI,SAAS,KAAK,SAAS,EAAE,CAAC;YACjD,MAAM,IAAI,iCAAkB,CAAC,GAAG,CAAC,OAAO,EAAE,EAAE,qBAAqB,EAAE;gBAClE,WAAW,EACV,6KAA6K;gBAC9K,SAAS;aACT,CAAC,CAAC;QACJ,CAAC;IACF,CAAC;SAAM,CAAC;QACP,SAAS,GAAG,GAAG,CAAC,gBAAgB,CAAC,SAAS,EAAE,SAAS,EAAE,EAAE,CAAW,CAAC;QACrE,IAAI,SAAS,KAAK,EAAE,IAAI,SAAS,KAAK,SAAS,EAAE,CAAC;YACjD,MAAM,IAAI,iCAAkB,CAAC,GAAG,CAAC,OAAO,EAAE,EAAE,wBAAwB,EAAE;gBACrE,WAAW,EACV,0LAA0L;gBAC3L,SAAS;aACT,CAAC,CAAC;QACJ,CAAC;IACF,CAAC;IAED,OAAO,SAAS,CAAC;AAClB,CAAC;AAEM,KAAK,UAAU,UAAU,CAC/B,gBAAmC,EACnC,KAA4B,EAC5B,IAAkB;IAElB,IAAI,CAAC;QACJ,MAAM,gBAAgB,CAAC,UAAU,CAAC,KAAK,EAAE,IAAI,CAAC,CAAC,CAAC,IAAA,4BAAa,EAAC,IAAI,CAAC,CAAC,CAAC,CAAC,SAAS,CAAC,CAAC;IAClF,CAAC;IAAC,OAAO,KAAK,EAAE,CAAC;QAChB,gBAAgB,CAAC,MAAM,CAAC,KAAK,CAAC,2BAA2B,KAAK,EAAE,CAAC,CAAC;IACnE,CAAC;AACF,CAAC;AAED,SAAgB,oBAAoB,CAAC,WAA0B;IAC9D,OAAO,WAAW;SAChB,GAAG,CAAC,CAAC,WAAW,EAAE,EAAE;QACpB,IAAI,WAAW,CAAC,QAAQ,EAAE,KAAK,OAAO,EAAE,CAAC;YACxC,OAAO,UAAU,WAAW,CAAC,OAAO,EAAE,CAAC;QACxC,CAAC;aAAM,IAAI,WAAW,CAAC,QAAQ,EAAE,KAAK,IAAI,EAAE,CAAC;YAC5C,OAAO,cAAc,WAAW,CAAC,OAAO,EAAE,CAAC;QAC5C,CAAC;aAAM,CAAC;YACP,OAAO,GAAG,WAAW,CAAC,OAAO,EAAE,CAAC;QACjC,CAAC;IACF,CAAC,CAAC;SACD,IAAI,CAAC,IAAI,CAAC,CAAC;AACd,CAAC;AAEM,MAAM,iBAAiB,GAAG,KAAK,EACrC,GAAsB,EACtB,kBAA2B,EAC3B,wBAAiC,IAAI,EACpC,EAAE;IACH,MAAM,cAAc,GAClB,CAAC,MAAM,GAAG,CAAC,sBAAsB,YAA4B,CAAC,CAAC,CAAY,IAAI,EAAE,CAAC;IAEpF,IAAI,CAAC,kBAAkB;QAAE,OAAO,cAAc,CAAC;IAE/C,MAAM,SAAS,GAAG,IAAI,GAAG,EAAU,CAAC;IAEpC,MAAM,UAAU,GAAG,EAAE,CAAC;IAEtB,KAAK,MAAM,IAAI,IAAI,cAAc,EAAE,CAAC;QACnC,MAAM,EAAE,IAAI,EAAE,GAAG,IAAI,CAAC;QACtB,IAAI,SAAS,CAAC,GAAG,CAAC,IAAI,CAAC,EAAE,CAAC;YACzB,MAAM,IAAI,iCAAkB,CAC3B,GAAG,CAAC,OAAO,EAAE,EACb,gDAAgD,IAAI,0CAA0C,CAC9F,CAAC;QACH,CAAC;QACD,SAAS,CAAC,GAAG,CAAC,IAAI,CAAC,CAAC;QAEpB,IAAI,qBAAqB,IAAI,IAAI,YAAY,iBAAO,EAAE,CAAC;YACtD,UAAU,CAAC,IAAI,CAAC,IAAI,CAAC,aAAa,EAAE,CAAC,CAAC;QACvC,CAAC;aAAM,CAAC;YACP,UAAU,CAAC,IAAI,CAAC,IAAI,CAAC,CAAC;QACvB,CAAC;IACF,CAAC;IAED,OAAO,UAAU,CAAC;AACnB,CAAC,CAAC;AAhCW,QAAA,iBAAiB,qBAgC5B"}
1
+ {"version":3,"file":"helpers.js","sourceRoot":"","sources":["../../utils/helpers.ts"],"names":[],"mappings":";;;AAqBA,4DAyBC;AAED,4CAEC;AAED,4DAEC;AAED,wCAIC;AAED,0CAIC;AAED,4DAaC;AAED,oDAwBC;AAED,oCAuCC;AAED,gCAUC;AAED,oDAYC;AA9KD,+CAAqF;AASrF,uCAAoC;AAEpC,SAAS,UAAU,CAAI,GAAY,EAAE,GAAG,WAAmC;IAC1E,OAAO,WAAW,CAAC,KAAK,CACvB,CAAC,UAAU,EAAE,EAAE,CACd,OAAO,GAAG,KAAK,QAAQ;QACvB,GAAG,KAAK,IAAI;QACZ,UAAU,IAAI,GAAG;QACjB,OAAQ,GAAwC,CAAC,UAAU,CAAC,KAAK,UAAU,CAC5E,CAAC;AACH,CAAC;AAED,SAAgB,wBAAwB,CACvC,GAAsB,EACtB,SAAiB;IAEjB,MAAM,OAAO,GAAG,GAAG,CAAC,gBAAgB,CAAC,SAAS,EAAE,SAAS,EAAE,EAAE,CAAC,CAAC;IAE/D,IAAI,OAAO,CAAC,QAAQ,EAAE,CAAC;QACtB,MAAM,EAAE,cAAc,EAAE,QAAQ,EAAE,GAAG,OAAO,CAAC,QAK5C,CAAC;QACF,IAAI,QAAQ,CAAC,MAAM,GAAG,CAAC,EAAE,CAAC;YACzB,OAAO,QAAQ,CAAC,MAAM,CAAC,CAAC,GAAG,EAAE,EAAE,IAAI,EAAE,KAAK,EAAE,EAAE,EAAE,CAAC,CAAC,EAAE,GAAG,GAAG,EAAE,CAAC,IAAI,CAAC,EAAE,KAAK,EAAE,CAAC,EAAE,EAAE,CAAC,CAAC;QACnF,CAAC;IACF,CAAC;IAED,IAAI,OAAO,CAAC,gBAAgB,EAAE,CAAC;QAC9B,OAAO,GAAG,CAAC,gBAAgB,CAAC,0BAA0B,EAAE,SAAS,EAAE,EAAE,EAAE;YACtE,UAAU,EAAE,QAAQ;SACpB,CAA0B,CAAC;IAC7B,CAAC;IAED,OAAO,SAAS,CAAC;AAClB,CAAC;AAED,SAAgB,gBAAgB,CAAC,GAAY;IAC5C,OAAO,UAAU,CAAiB,GAAG,EAAE,qBAAqB,EAAE,aAAa,CAAC,CAAC;AAC9E,CAAC;AAED,SAAgB,wBAAwB,CAAC,GAAY;IACpD,OAAO,UAAU,CAAyB,GAAG,EAAE,aAAa,EAAE,YAAY,CAAC,CAAC;AAC7E,CAAC;AAED,SAAgB,cAAc,CAAC,KAAc;IAC5C,MAAM,SAAS,GAAI,KAAiB,EAAE,YAAY,IAAI,EAAE,CAAC;IAEzD,OAAO,SAAS,CAAC,QAAQ,CAAC,aAAa,CAAC,CAAC;AAC1C,CAAC;AAED,SAAgB,eAAe,CAAC,KAAc;IAC7C,MAAM,SAAS,GAAI,KAAc,EAAE,YAAY,IAAI,EAAE,CAAC;IAEtD,OAAO,SAAS,CAAC,QAAQ,CAAC,OAAO,CAAC,CAAC;AACpC,CAAC;AAEM,KAAK,UAAU,wBAAwB,CAC7C,GAAsB;IAEtB,IAAI,aAAa,GAAuB,EAAE,CAAC;IAE3C,IAAI,GAAG,CAAC,gBAAgB,CAAC,iBAAiB,EAAE,CAAC,EAAE,IAAI,CAAC,KAAK,IAAI,EAAE,CAAC;QAC/D,aAAa,GAAG,CAAC,MAAM,GAAG,CAAC,sBAAsB,oBAEhD,CAAC,CACD,CAAuB,CAAC;IAC1B,CAAC;IAED,OAAO,aAAa,CAAC;AACtB,CAAC;AAED,SAAgB,oBAAoB,CAAC,OAKpC;IACA,MAAM,EAAE,GAAG,EAAE,CAAC,EAAE,aAAa,EAAE,QAAQ,EAAE,GAAG,OAAO,CAAC;IACpD,MAAM,MAAM,GAAG,GAAG,CAAC,gBAAgB,CAAC,aAAa,EAAE,CAAC,CAAW,CAAC;IAEhE,IAAI,KAAK,CAAC;IACV,IAAI,MAAM,KAAK,MAAM,EAAE,CAAC;QACvB,KAAK,GAAG,GAAG,CAAC,kBAAkB,CAAC,0BAA0B,EAAE,CAAC,CAAW,CAAC;IACzE,CAAC;SAAM,CAAC;QACP,KAAK,GAAG,GAAG,CAAC,gBAAgB,CAAC,QAAQ,EAAE,CAAC,CAAW,CAAC;IACrD,CAAC;IAED,IAAI,KAAK,KAAK,SAAS,EAAE,CAAC;QACzB,MAAM,IAAI,iCAAkB,CAAC,GAAG,CAAC,OAAO,EAAE,EAAE,qBAAqB,EAAE;YAClE,WAAW,EACV,qKAAqK;SACtK,CAAC,CAAC;IACJ,CAAC;IAED,OAAO,KAAK,CAAC;AACd,CAAC;AAED,SAAgB,YAAY,CAC3B,GAA0C,EAC1C,SAAiB,EACjB,WAAW,GAAG,eAAe,EAC7B,UAAU,GAAG,WAAW,EACxB,SAAS,GAAG,YAAY;IAExB,IAAI,SAAS,GAAG,EAAE,CAAC;IACnB,MAAM,YAAY,GAAG,GAAG,CAAC,gBAAgB,CAAC,WAAW,EAAE,SAAS,CAAW,CAAC;IAE5E,IAAI,YAAY,KAAK,UAAU,EAAE,CAAC;QAGjC,IAAI,aAAa,IAAI,GAAG,EAAE,CAAC;YAC1B,MAAM,QAAQ,GAAG,GAAG,CAAC,WAAW,EAAE,IAAI,EAAE,CAAC;YACzC,SAAS,GAAG,QAAQ,CAAC,SAAmB,CAAC;QAC1C,CAAC;aAAM,CAAC;YACP,SAAS,GAAG,GAAG,CAAC,kBAAkB,CAAC,uBAAuB,EAAE,SAAS,CAAW,CAAC;QAClF,CAAC;QAED,IAAI,SAAS,KAAK,EAAE,IAAI,SAAS,KAAK,SAAS,EAAE,CAAC;YACjD,MAAM,IAAI,iCAAkB,CAAC,GAAG,CAAC,OAAO,EAAE,EAAE,qBAAqB,EAAE;gBAClE,WAAW,EACV,6KAA6K;gBAC9K,SAAS;aACT,CAAC,CAAC;QACJ,CAAC;IACF,CAAC;SAAM,CAAC;QACP,SAAS,GAAG,GAAG,CAAC,gBAAgB,CAAC,SAAS,EAAE,SAAS,EAAE,EAAE,CAAW,CAAC;QACrE,IAAI,SAAS,KAAK,EAAE,IAAI,SAAS,KAAK,SAAS,EAAE,CAAC;YACjD,MAAM,IAAI,iCAAkB,CAAC,GAAG,CAAC,OAAO,EAAE,EAAE,wBAAwB,EAAE;gBACrE,WAAW,EACV,0LAA0L;gBAC3L,SAAS;aACT,CAAC,CAAC;QACJ,CAAC;IACF,CAAC;IAED,OAAO,SAAS,CAAC;AAClB,CAAC;AAEM,KAAK,UAAU,UAAU,CAC/B,gBAAmC,EACnC,KAAc,EACd,IAAkB;IAElB,IAAI,CAAC;QACJ,MAAM,gBAAgB,CAAC,UAAU,CAAC,KAAK,EAAE,IAAI,CAAC,CAAC,CAAC,IAAA,4BAAa,EAAC,IAAI,CAAC,CAAC,CAAC,CAAC,SAAS,CAAC,CAAC;IAClF,CAAC;IAAC,OAAO,KAAK,EAAE,CAAC;QAChB,gBAAgB,CAAC,MAAM,CAAC,KAAK,CAAC,2BAA2B,KAAK,EAAE,CAAC,CAAC;IACnE,CAAC;AACF,CAAC;AAED,SAAgB,oBAAoB,CAAC,WAA0B;IAC9D,OAAO,WAAW;SAChB,GAAG,CAAC,CAAC,WAAW,EAAE,EAAE;QACpB,IAAI,WAAW,CAAC,QAAQ,EAAE,KAAK,OAAO,EAAE,CAAC;YACxC,OAAO,UAAU,WAAW,CAAC,OAAO,EAAE,CAAC;QACxC,CAAC;aAAM,IAAI,WAAW,CAAC,QAAQ,EAAE,KAAK,IAAI,EAAE,CAAC;YAC5C,OAAO,cAAc,WAAW,CAAC,OAAO,EAAE,CAAC;QAC5C,CAAC;aAAM,CAAC;YACP,OAAO,GAAG,WAAW,CAAC,OAAO,EAAE,CAAC;QACjC,CAAC;IACF,CAAC,CAAC;SACD,IAAI,CAAC,IAAI,CAAC,CAAC;AACd,CAAC;AAEM,MAAM,iBAAiB,GAAG,KAAK,EACrC,GAAsB,EACtB,kBAA2B,EAC3B,wBAAiC,IAAI,EACpC,EAAE;IACH,MAAM,cAAc,GAClB,CAAC,MAAM,GAAG,CAAC,sBAAsB,YAA4B,CAAC,CAAC,CAAY,IAAI,EAAE,CAAC;IAEpF,IAAI,CAAC,kBAAkB;QAAE,OAAO,cAAc,CAAC;IAE/C,MAAM,SAAS,GAAG,IAAI,GAAG,EAAU,CAAC;IAEpC,MAAM,UAAU,GAAG,EAAE,CAAC;IAEtB,KAAK,MAAM,IAAI,IAAI,cAAc,EAAE,CAAC;QACnC,MAAM,EAAE,IAAI,EAAE,GAAG,IAAI,CAAC;QACtB,IAAI,SAAS,CAAC,GAAG,CAAC,IAAI,CAAC,EAAE,CAAC;YACzB,MAAM,IAAI,iCAAkB,CAC3B,GAAG,CAAC,OAAO,EAAE,EACb,gDAAgD,IAAI,0CAA0C,CAC9F,CAAC;QACH,CAAC;QACD,SAAS,CAAC,GAAG,CAAC,IAAI,CAAC,CAAC;QAEpB,IAAI,qBAAqB,IAAI,IAAI,YAAY,iBAAO,EAAE,CAAC;YACtD,UAAU,CAAC,IAAI,CAAC,IAAI,CAAC,aAAa,EAAE,CAAC,CAAC;QACvC,CAAC;aAAM,CAAC;YACP,UAAU,CAAC,IAAI,CAAC,IAAI,CAAC,CAAC;QACvB,CAAC;IACF,CAAC;IAED,OAAO,UAAU,CAAC;AACnB,CAAC,CAAC;AAhCW,QAAA,iBAAiB,qBAgC5B"}
@@ -119,7 +119,7 @@ function logWrapper(originalInstance, executeFunctions) {
119
119
  }));
120
120
  const payload = { action: 'getMessages', response };
121
121
  executeFunctions.addOutputData(connectionType, index, [[{ json: payload }]]);
122
- void (0, helpers_1.logAiEvent)(executeFunctions, 'n8n.ai.memory.get.messages', { response });
122
+ void (0, helpers_1.logAiEvent)(executeFunctions, 'ai-messages-retrieved-from-memory', { response });
123
123
  return response;
124
124
  };
125
125
  }
@@ -135,7 +135,7 @@ function logWrapper(originalInstance, executeFunctions) {
135
135
  method: target[prop],
136
136
  arguments: [message],
137
137
  });
138
- void (0, helpers_1.logAiEvent)(executeFunctions, 'n8n.ai.memory.added.message', { message });
138
+ void (0, helpers_1.logAiEvent)(executeFunctions, 'ai-message-added-to-memory', { message });
139
139
  executeFunctions.addOutputData(connectionType, index, [[{ json: payload }]]);
140
140
  };
141
141
  }
@@ -156,14 +156,14 @@ function logWrapper(originalInstance, executeFunctions) {
156
156
  method: target[prop],
157
157
  arguments: [stringifiedText],
158
158
  }));
159
- void (0, helpers_1.logAiEvent)(executeFunctions, 'n8n.ai.output.parser.parsed', { text, response });
159
+ void (0, helpers_1.logAiEvent)(executeFunctions, 'ai-output-parsed', { text, response });
160
160
  executeFunctions.addOutputData(connectionType, index, [
161
161
  [{ json: { action: 'parse', response } }],
162
162
  ]);
163
163
  return response;
164
164
  }
165
165
  catch (error) {
166
- void (0, helpers_1.logAiEvent)(executeFunctions, 'n8n.ai.output.parser.parsed', {
166
+ void (0, helpers_1.logAiEvent)(executeFunctions, 'ai-output-parsed', {
167
167
  text,
168
168
  response: error.message ?? error,
169
169
  });
@@ -189,7 +189,7 @@ function logWrapper(originalInstance, executeFunctions) {
189
189
  method: target[prop],
190
190
  arguments: [query, config],
191
191
  }));
192
- void (0, helpers_1.logAiEvent)(executeFunctions, 'n8n.ai.retriever.get.relevant.documents', { query });
192
+ void (0, helpers_1.logAiEvent)(executeFunctions, 'ai-documents-retrieved', { query });
193
193
  executeFunctions.addOutputData(connectionType, index, [[{ json: { response } }]]);
194
194
  return response;
195
195
  };
@@ -209,7 +209,7 @@ function logWrapper(originalInstance, executeFunctions) {
209
209
  method: target[prop],
210
210
  arguments: [documents],
211
211
  }));
212
- void (0, helpers_1.logAiEvent)(executeFunctions, 'n8n.ai.embeddings.embedded.document');
212
+ void (0, helpers_1.logAiEvent)(executeFunctions, 'ai-document-embedded');
213
213
  executeFunctions.addOutputData(connectionType, index, [[{ json: { response } }]]);
214
214
  return response;
215
215
  };
@@ -227,7 +227,7 @@ function logWrapper(originalInstance, executeFunctions) {
227
227
  method: target[prop],
228
228
  arguments: [query],
229
229
  }));
230
- void (0, helpers_1.logAiEvent)(executeFunctions, 'n8n.ai.embeddings.embedded.query');
230
+ void (0, helpers_1.logAiEvent)(executeFunctions, 'ai-query-embedded');
231
231
  executeFunctions.addOutputData(connectionType, index, [[{ json: { response } }]]);
232
232
  return response;
233
233
  };
@@ -261,7 +261,7 @@ function logWrapper(originalInstance, executeFunctions) {
261
261
  method: target[prop],
262
262
  arguments: [item, itemIndex],
263
263
  }));
264
- void (0, helpers_1.logAiEvent)(executeFunctions, 'n8n.ai.document.processed');
264
+ void (0, helpers_1.logAiEvent)(executeFunctions, 'ai-document-processed');
265
265
  executeFunctions.addOutputData(connectionType, index, [
266
266
  [{ json: { response }, pairedItem: { item: itemIndex } }],
267
267
  ]);
@@ -283,7 +283,7 @@ function logWrapper(originalInstance, executeFunctions) {
283
283
  method: target[prop],
284
284
  arguments: [text],
285
285
  }));
286
- void (0, helpers_1.logAiEvent)(executeFunctions, 'n8n.ai.text.splitter.split');
286
+ void (0, helpers_1.logAiEvent)(executeFunctions, 'ai-text-split');
287
287
  executeFunctions.addOutputData(connectionType, index, [[{ json: { response } }]]);
288
288
  return response;
289
289
  };
@@ -303,7 +303,7 @@ function logWrapper(originalInstance, executeFunctions) {
303
303
  method: target[prop],
304
304
  arguments: [query],
305
305
  }));
306
- void (0, helpers_1.logAiEvent)(executeFunctions, 'n8n.ai.tool.called', { query, response });
306
+ void (0, helpers_1.logAiEvent)(executeFunctions, 'ai-tool-called', { query, response });
307
307
  executeFunctions.addOutputData(connectionType, index, [[{ json: { response } }]]);
308
308
  return response;
309
309
  };
@@ -323,7 +323,7 @@ function logWrapper(originalInstance, executeFunctions) {
323
323
  method: target[prop],
324
324
  arguments: [query, k, filter, _callbacks],
325
325
  }));
326
- void (0, helpers_1.logAiEvent)(executeFunctions, 'n8n.ai.vector.store.searched', { query });
326
+ void (0, helpers_1.logAiEvent)(executeFunctions, 'ai-vector-store-searched', { query });
327
327
  executeFunctions.addOutputData(connectionType, index, [[{ json: { response } }]]);
328
328
  return response;
329
329
  };
@@ -1 +1 @@
1
- {"version":3,"file":"logWrapper.js","sourceRoot":"","sources":["../../utils/logWrapper.ts"],"names":[],"mappings":";;AA6BA,0CA6CC;AAED,wCA4BC;AAED,gCAmWC;AA7cD,+CAAsE;AAStE,2DAAwD;AACxD,+DAA2D;AAE3D,4DAAwD;AAExD,2DAA2D;AAC3D,mEAAyF;AACzF,mCAAkC;AAElC,mDAAgD;AAChD,uDAAoD;AACpD,uCAAoG;AAEpG,MAAM,SAAS,GAAgE;IAC9E,8EAA8E,EAAE;QAC/E,OAAO,EAAE,uBAAuB;QAChC,WAAW,EAAE,8EAA8E;KAC3F;CACD,CAAC;AAEK,KAAK,UAAU,eAAe,CAEpC,UAMC;IAED,IAAI,CAAC;QACJ,OAAO,MAAM,UAAU,CAAC,MAAM,CAAC,IAAI,CAAC,IAAI,EAAE,GAAG,UAAU,CAAC,SAAS,CAAC,CAAC;IACpE,CAAC;IAAC,OAAO,CAAC,EAAE,CAAC;QAGZ,IAAI,CAAC,YAAY,sCAAqB;YAAE,MAAM,CAAC,CAAC;QAGhD,IAAI,CAAC,CAAC,aAAa,KAAK,oBAAoB;YAAE,MAAM,CAAC,CAAC;QACtD,MAAM,aAAa,GAAG,UAAU,CAAC,gBAAgB,CAAC,OAAO,EAAE,CAAC;QAE5D,MAAM,KAAK,GAAG,IAAI,iCAAkB,CAAC,aAAa,EAAE,CAAC,EAAE;YACtD,aAAa,EAAE,oBAAoB;SACnC,CAAC,CAAC;QAEH,IAAI,SAAS,CAAC,KAAK,CAAC,OAAO,CAAC,EAAE,CAAC;YAC9B,KAAK,CAAC,WAAW,GAAG,SAAS,CAAC,KAAK,CAAC,OAAO,CAAC,CAAC,WAAW,CAAC;YACzD,KAAK,CAAC,OAAO,GAAG,SAAS,CAAC,KAAK,CAAC,OAAO,CAAC,CAAC,OAAO,CAAC;QAClD,CAAC;QAED,UAAU,CAAC,gBAAgB,CAAC,aAAa,CACxC,UAAU,CAAC,cAAc,EACzB,UAAU,CAAC,mBAAmB,EAC9B,KAAK,CACL,CAAC;QACF,IAAI,KAAK,CAAC,OAAO,EAAE,CAAC;YACnB,KAAK,CAAC,WAAW,GAAG,KAAK,CAAC,OAAO,CAAC;YAClC,MAAM,KAAK,CAAC;QACb,CAAC;QACD,MAAM,IAAI,iCAAkB,CAC3B,aAAa,EACb,kBAAkB,aAAa,CAAC,IAAI,mCAAmC,UAAU,CAAC,cAAc,GAAG,EACnG,EAAE,aAAa,EAAE,oBAAoB,EAAE,CACvC,CAAC;IACH,CAAC;AACF,CAAC;AAED,SAAgB,cAAc,CAE7B,UAMC;IAED,IAAI,CAAC;QACJ,OAAO,UAAU,CAAC,MAAM,CAAC,IAAI,CAAC,IAAI,EAAE,GAAG,UAAU,CAAC,SAAS,CAAC,CAAC;IAC9D,CAAC;IAAC,OAAO,CAAC,EAAE,CAAC;QAEZ,IAAI,CAAC,CAAC,aAAa,KAAK,oBAAoB;YAAE,MAAM,CAAC,CAAC;QACtD,MAAM,aAAa,GAAG,UAAU,CAAC,gBAAgB,CAAC,OAAO,EAAE,CAAC;QAC5D,MAAM,KAAK,GAAG,IAAI,iCAAkB,CAAC,aAAa,EAAE,CAAC,CAAC,CAAC;QACvD,UAAU,CAAC,gBAAgB,CAAC,aAAa,CACxC,UAAU,CAAC,cAAc,EACzB,UAAU,CAAC,mBAAmB,EAC9B,KAAK,CACL,CAAC;QACF,MAAM,IAAI,iCAAkB,CAC3B,aAAa,EACb,kBAAkB,aAAa,CAAC,IAAI,mCAAmC,UAAU,CAAC,cAAc,GAAG,EACnG,EAAE,aAAa,EAAE,oBAAoB,EAAE,CACvC,CAAC;IACH,CAAC;AACF,CAAC;AAED,SAAgB,UAAU,CACzB,gBAagB,EAChB,gBAAmC;IAEnC,OAAO,IAAI,KAAK,CAAC,gBAAgB,EAAE;QAClC,GAAG,EAAE,CAAC,MAAM,EAAE,IAAI,EAAE,EAAE;YACrB,IAAI,cAA8C,CAAC;YAEnD,IAAI,IAAA,0BAAgB,EAAC,gBAAgB,CAAC,EAAE,CAAC;gBACxC,IAAI,IAAI,KAAK,qBAAqB,IAAI,qBAAqB,IAAI,MAAM,EAAE,CAAC;oBACvE,OAAO,KAAK,EAAE,MAAmB,EAA4B,EAAE;wBAC9D,cAAc,cAA8B,CAAC;wBAE7C,MAAM,EAAE,KAAK,EAAE,GAAG,gBAAgB,CAAC,YAAY,CAAC,cAAc,EAAE;4BAC/D,CAAC,EAAE,IAAI,EAAE,EAAE,MAAM,EAAE,qBAAqB,EAAE,MAAM,EAAE,EAAE,CAAC;yBACrD,CAAC,CAAC;wBAEH,MAAM,QAAQ,GAAG,CAAC,MAAM,eAAe,CAAC,IAAI,CAAC,MAAM,EAAE;4BACpD,gBAAgB;4BAChB,cAAc;4BACd,mBAAmB,EAAE,KAAK;4BAC1B,MAAM,EAAE,MAAM,CAAC,IAAI,CAAC;4BACpB,SAAS,EAAE,CAAC,MAAM,CAAC;yBACnB,CAAC,CAAoB,CAAC;wBAEvB,MAAM,WAAW,GAAI,QAAQ,EAAE,YAA8B,IAAI,QAAQ,CAAC;wBAE1E,gBAAgB,CAAC,aAAa,CAAC,cAAc,EAAE,KAAK,EAAE;4BACrD,CAAC,EAAE,IAAI,EAAE,EAAE,MAAM,EAAE,qBAAqB,EAAE,WAAW,EAAE,EAAE,CAAC;yBAC1D,CAAC,CAAC;wBACH,OAAO,QAAQ,CAAC;oBACjB,CAAC,CAAC;gBACH,CAAC;qBAAM,IAAI,IAAI,KAAK,aAAa,IAAI,aAAa,IAAI,MAAM,EAAE,CAAC;oBAC9D,OAAO,KAAK,EAAE,KAAkB,EAAE,MAAoB,EAA4B,EAAE;wBACnF,cAAc,cAA8B,CAAC;wBAE7C,MAAM,EAAE,KAAK,EAAE,GAAG,gBAAgB,CAAC,YAAY,CAAC,cAAc,EAAE;4BAC/D,CAAC,EAAE,IAAI,EAAE,EAAE,MAAM,EAAE,aAAa,EAAE,KAAK,EAAE,MAAM,EAAE,EAAE,CAAC;yBACpD,CAAC,CAAC;wBAEH,MAAM,QAAQ,GAAG,CAAC,MAAM,eAAe,CAAC,IAAI,CAAC,MAAM,EAAE;4BACpD,gBAAgB;4BAChB,cAAc;4BACd,mBAAmB,EAAE,KAAK;4BAC1B,MAAM,EAAE,MAAM,CAAC,IAAI,CAAC;4BACpB,SAAS,EAAE,CAAC,KAAK,EAAE,MAAM,CAAC;yBAC1B,CAAC,CAAoB,CAAC;wBAEvB,MAAM,WAAW,GAAG,MAAM,MAAM,CAAC,WAAW,CAAC,WAAW,EAAE,CAAC;wBAE3D,gBAAgB,CAAC,aAAa,CAAC,cAAc,EAAE,KAAK,EAAE;4BACrD,CAAC,EAAE,IAAI,EAAE,EAAE,MAAM,EAAE,aAAa,EAAE,WAAW,EAAE,EAAE,CAAC;yBAClD,CAAC,CAAC;wBAEH,OAAO,QAAQ,CAAC;oBACjB,CAAC,CAAC;gBACH,CAAC;YACF,CAAC;YAGD,IAAI,IAAA,kCAAwB,EAAC,gBAAgB,CAAC,EAAE,CAAC;gBAChD,IAAI,IAAI,KAAK,aAAa,IAAI,aAAa,IAAI,MAAM,EAAE,CAAC;oBACvD,OAAO,KAAK,IAA4B,EAAE;wBACzC,cAAc,cAA8B,CAAC;wBAC7C,MAAM,EAAE,KAAK,EAAE,GAAG,gBAAgB,CAAC,YAAY,CAAC,cAAc,EAAE;4BAC/D,CAAC,EAAE,IAAI,EAAE,EAAE,MAAM,EAAE,aAAa,EAAE,EAAE,CAAC;yBACrC,CAAC,CAAC;wBAEH,MAAM,QAAQ,GAAG,CAAC,MAAM,eAAe,CAAC,IAAI,CAAC,MAAM,EAAE;4BACpD,gBAAgB;4BAChB,cAAc;4BACd,mBAAmB,EAAE,KAAK;4BAC1B,MAAM,EAAE,MAAM,CAAC,IAAI,CAAC;4BACpB,SAAS,EAAE,EAAE;yBACb,CAAC,CAAkB,CAAC;wBAErB,MAAM,OAAO,GAAG,EAAE,MAAM,EAAE,aAAa,EAAE,QAAQ,EAAE,CAAC;wBACpD,gBAAgB,CAAC,aAAa,CAAC,cAAc,EAAE,KAAK,EAAE,CAAC,CAAC,EAAE,IAAI,EAAE,OAAO,EAAE,CAAC,CAAC,CAAC,CAAC;wBAE7E,KAAK,IAAA,oBAAU,EAAC,gBAAgB,EAAE,4BAA4B,EAAE,EAAE,QAAQ,EAAE,CAAC,CAAC;wBAC9E,OAAO,QAAQ,CAAC;oBACjB,CAAC,CAAC;gBACH,CAAC;qBAAM,IAAI,IAAI,KAAK,YAAY,IAAI,YAAY,IAAI,MAAM,EAAE,CAAC;oBAC5D,OAAO,KAAK,EAAE,OAAoB,EAAiB,EAAE;wBACpD,cAAc,cAA8B,CAAC;wBAC7C,MAAM,OAAO,GAAG,EAAE,MAAM,EAAE,YAAY,EAAE,OAAO,EAAE,CAAC;wBAClD,MAAM,EAAE,KAAK,EAAE,GAAG,gBAAgB,CAAC,YAAY,CAAC,cAAc,EAAE,CAAC,CAAC,EAAE,IAAI,EAAE,OAAO,EAAE,CAAC,CAAC,CAAC,CAAC;wBAEvF,MAAM,eAAe,CAAC,IAAI,CAAC,MAAM,EAAE;4BAClC,gBAAgB;4BAChB,cAAc;4BACd,mBAAmB,EAAE,KAAK;4BAC1B,MAAM,EAAE,MAAM,CAAC,IAAI,CAAC;4BACpB,SAAS,EAAE,CAAC,OAAO,CAAC;yBACpB,CAAC,CAAC;wBAEH,KAAK,IAAA,oBAAU,EAAC,gBAAgB,EAAE,6BAA6B,EAAE,EAAE,OAAO,EAAE,CAAC,CAAC;wBAC9E,gBAAgB,CAAC,aAAa,CAAC,cAAc,EAAE,KAAK,EAAE,CAAC,CAAC,EAAE,IAAI,EAAE,OAAO,EAAE,CAAC,CAAC,CAAC,CAAC;oBAC9E,CAAC,CAAC;gBACH,CAAC;YACF,CAAC;YAGD,IAAI,gBAAgB,YAAY,iCAAgB,EAAE,CAAC;gBAClD,IAAI,IAAI,KAAK,OAAO,IAAI,OAAO,IAAI,MAAM,EAAE,CAAC;oBAC3C,OAAO,KAAK,EAAE,IAAsC,EAAoB,EAAE;wBACzE,cAAc,oBAAoC,CAAC;wBACnD,MAAM,eAAe,GAAG,IAAA,iBAAQ,EAAC,IAAI,CAAC,CAAC,CAAC,CAAC,IAAI,CAAC,SAAS,CAAC,IAAI,CAAC,CAAC,CAAC,CAAC,IAAI,CAAC;wBACrE,MAAM,EAAE,KAAK,EAAE,GAAG,gBAAgB,CAAC,YAAY,CAAC,cAAc,EAAE;4BAC/D,CAAC,EAAE,IAAI,EAAE,EAAE,MAAM,EAAE,OAAO,EAAE,IAAI,EAAE,eAAe,EAAE,EAAE,CAAC;yBACtD,CAAC,CAAC;wBAEH,IAAI,CAAC;4BACJ,MAAM,QAAQ,GAAG,CAAC,MAAM,eAAe,CAAC,IAAI,CAAC,MAAM,EAAE;gCACpD,gBAAgB;gCAChB,cAAc;gCACd,mBAAmB,EAAE,KAAK;gCAC1B,MAAM,EAAE,MAAM,CAAC,IAAI,CAAC;gCACpB,SAAS,EAAE,CAAC,eAAe,CAAC;6BAC5B,CAAC,CAAW,CAAC;4BAEd,KAAK,IAAA,oBAAU,EAAC,gBAAgB,EAAE,6BAA6B,EAAE,EAAE,IAAI,EAAE,QAAQ,EAAE,CAAC,CAAC;4BACrF,gBAAgB,CAAC,aAAa,CAAC,cAAc,EAAE,KAAK,EAAE;gCACrD,CAAC,EAAE,IAAI,EAAE,EAAE,MAAM,EAAE,OAAO,EAAE,QAAQ,EAAE,EAAE,CAAC;6BACzC,CAAC,CAAC;4BACH,OAAO,QAAQ,CAAC;wBACjB,CAAC;wBAAC,OAAO,KAAK,EAAE,CAAC;4BAChB,KAAK,IAAA,oBAAU,EAAC,gBAAgB,EAAE,6BAA6B,EAAE;gCAChE,IAAI;gCACJ,QAAQ,EAAE,KAAK,CAAC,OAAO,IAAI,KAAK;6BAChC,CAAC,CAAC;4BACH,gBAAgB,CAAC,aAAa,CAAC,cAAc,EAAE,KAAK,EAAE;gCACrD,CAAC,EAAE,IAAI,EAAE,EAAE,MAAM,EAAE,OAAO,EAAE,QAAQ,EAAE,KAAK,CAAC,OAAO,IAAI,KAAK,EAAE,EAAE,CAAC;6BACjE,CAAC,CAAC;4BACH,MAAM,KAAK,CAAC;wBACb,CAAC;oBACF,CAAC,CAAC;gBACH,CAAC;YACF,CAAC;YAGD,IAAI,gBAAgB,YAAY,0BAAa,EAAE,CAAC;gBAC/C,IAAI,IAAI,KAAK,sBAAsB,IAAI,sBAAsB,IAAI,MAAM,EAAE,CAAC;oBACzE,OAAO,KAAK,EACX,KAAa,EACb,MAAuC,EACjB,EAAE;wBACxB,cAAc,iBAAiC,CAAC;wBAChD,MAAM,EAAE,KAAK,EAAE,GAAG,gBAAgB,CAAC,YAAY,CAAC,cAAc,EAAE;4BAC/D,CAAC,EAAE,IAAI,EAAE,EAAE,KAAK,EAAE,MAAM,EAAE,EAAE,CAAC;yBAC7B,CAAC,CAAC;wBAEH,MAAM,QAAQ,GAAG,CAAC,MAAM,eAAe,CAAC,IAAI,CAAC,MAAM,EAAE;4BACpD,gBAAgB;4BAChB,cAAc;4BACd,mBAAmB,EAAE,KAAK;4BAC1B,MAAM,EAAE,MAAM,CAAC,IAAI,CAAC;4BACpB,SAAS,EAAE,CAAC,KAAK,EAAE,MAAM,CAAC;yBAC1B,CAAC,CAAyC,CAAC;wBAE5C,KAAK,IAAA,oBAAU,EAAC,gBAAgB,EAAE,yCAAyC,EAAE,EAAE,KAAK,EAAE,CAAC,CAAC;wBACxF,gBAAgB,CAAC,aAAa,CAAC,cAAc,EAAE,KAAK,EAAE,CAAC,CAAC,EAAE,IAAI,EAAE,EAAE,QAAQ,EAAE,EAAE,CAAC,CAAC,CAAC,CAAC;wBAClF,OAAO,QAAQ,CAAC;oBACjB,CAAC,CAAC;gBACH,CAAC;YACF,CAAC;YAGD,IAAI,gBAAgB,YAAY,uBAAU,EAAE,CAAC;gBAE5C,IAAI,IAAI,KAAK,gBAAgB,IAAI,gBAAgB,IAAI,MAAM,EAAE,CAAC;oBAC7D,OAAO,KAAK,EAAE,SAAmB,EAAuB,EAAE;wBACzD,cAAc,iBAAiC,CAAC;wBAChD,MAAM,EAAE,KAAK,EAAE,GAAG,gBAAgB,CAAC,YAAY,CAAC,cAAc,EAAE;4BAC/D,CAAC,EAAE,IAAI,EAAE,EAAE,SAAS,EAAE,EAAE,CAAC;yBACzB,CAAC,CAAC;wBAEH,MAAM,QAAQ,GAAG,CAAC,MAAM,eAAe,CAAC,IAAI,CAAC,MAAM,EAAE;4BACpD,gBAAgB;4BAChB,cAAc;4BACd,mBAAmB,EAAE,KAAK;4BAC1B,MAAM,EAAE,MAAM,CAAC,IAAI,CAAC;4BACpB,SAAS,EAAE,CAAC,SAAS,CAAC;yBACtB,CAAC,CAAe,CAAC;wBAElB,KAAK,IAAA,oBAAU,EAAC,gBAAgB,EAAE,qCAAqC,CAAC,CAAC;wBACzE,gBAAgB,CAAC,aAAa,CAAC,cAAc,EAAE,KAAK,EAAE,CAAC,CAAC,EAAE,IAAI,EAAE,EAAE,QAAQ,EAAE,EAAE,CAAC,CAAC,CAAC,CAAC;wBAClF,OAAO,QAAQ,CAAC;oBACjB,CAAC,CAAC;gBACH,CAAC;gBAED,IAAI,IAAI,KAAK,YAAY,IAAI,YAAY,IAAI,MAAM,EAAE,CAAC;oBACrD,OAAO,KAAK,EAAE,KAAa,EAAqB,EAAE;wBACjD,cAAc,iBAAiC,CAAC;wBAChD,MAAM,EAAE,KAAK,EAAE,GAAG,gBAAgB,CAAC,YAAY,CAAC,cAAc,EAAE;4BAC/D,CAAC,EAAE,IAAI,EAAE,EAAE,KAAK,EAAE,EAAE,CAAC;yBACrB,CAAC,CAAC;wBAEH,MAAM,QAAQ,GAAG,CAAC,MAAM,eAAe,CAAC,IAAI,CAAC,MAAM,EAAE;4BACpD,gBAAgB;4BAChB,cAAc;4BACd,mBAAmB,EAAE,KAAK;4BAC1B,MAAM,EAAE,MAAM,CAAC,IAAI,CAAC;4BACpB,SAAS,EAAE,CAAC,KAAK,CAAC;yBAClB,CAAC,CAAa,CAAC;wBAChB,KAAK,IAAA,oBAAU,EAAC,gBAAgB,EAAE,kCAAkC,CAAC,CAAC;wBACtE,gBAAgB,CAAC,aAAa,CAAC,cAAc,EAAE,KAAK,EAAE,CAAC,CAAC,EAAE,IAAI,EAAE,EAAE,QAAQ,EAAE,EAAE,CAAC,CAAC,CAAC,CAAC;wBAClF,OAAO,QAAQ,CAAC;oBACjB,CAAC,CAAC;gBACH,CAAC;YACF,CAAC;YAGD,IACC,gBAAgB,YAAY,6BAAa;gBACzC,gBAAgB,YAAY,iCAAe,EAC1C,CAAC;gBAEF,IAAI,IAAI,KAAK,YAAY,IAAI,YAAY,IAAI,MAAM,EAAE,CAAC;oBACrD,OAAO,KAAK,EAAE,KAA2B,EAAqB,EAAE;wBAC/D,cAAc,gBAAgC,CAAC;wBAC/C,MAAM,EAAE,KAAK,EAAE,GAAG,gBAAgB,CAAC,YAAY,CAAC,cAAc,EAAE,CAAC,KAAK,CAAC,CAAC,CAAC;wBAEzE,MAAM,QAAQ,GAAG,CAAC,MAAM,eAAe,CAAC,IAAI,CAAC,MAAM,EAAE;4BACpD,gBAAgB;4BAChB,cAAc;4BACd,mBAAmB,EAAE,KAAK;4BAC1B,MAAM,EAAE,MAAM,CAAC,IAAI,CAAC;4BACpB,SAAS,EAAE,CAAC,KAAK,CAAC;yBAClB,CAAC,CAAa,CAAC;wBAEhB,gBAAgB,CAAC,aAAa,CAAC,cAAc,EAAE,KAAK,EAAE,CAAC,CAAC,EAAE,IAAI,EAAE,EAAE,QAAQ,EAAE,EAAE,CAAC,CAAC,CAAC,CAAC;wBAClF,OAAO,QAAQ,CAAC;oBACjB,CAAC,CAAC;gBACH,CAAC;gBAGD,IAAI,IAAI,KAAK,aAAa,IAAI,aAAa,IAAI,MAAM,EAAE,CAAC;oBACvD,OAAO,KAAK,EAAE,IAAwB,EAAE,SAAiB,EAAqB,EAAE;wBAC/E,cAAc,gBAAgC,CAAC;wBAC/C,MAAM,EAAE,KAAK,EAAE,GAAG,gBAAgB,CAAC,YAAY,CAAC,cAAc,EAAE,CAAC,CAAC,IAAI,CAAC,CAAC,CAAC,CAAC;wBAE1E,MAAM,QAAQ,GAAG,CAAC,MAAM,eAAe,CAAC,IAAI,CAAC,MAAM,EAAE;4BACpD,gBAAgB;4BAChB,cAAc;4BACd,mBAAmB,EAAE,KAAK;4BAC1B,MAAM,EAAE,MAAM,CAAC,IAAI,CAAC;4BACpB,SAAS,EAAE,CAAC,IAAI,EAAE,SAAS,CAAC;yBAC5B,CAAC,CAAa,CAAC;wBAEhB,KAAK,IAAA,oBAAU,EAAC,gBAAgB,EAAE,2BAA2B,CAAC,CAAC;wBAC/D,gBAAgB,CAAC,aAAa,CAAC,cAAc,EAAE,KAAK,EAAE;4BACrD,CAAC,EAAE,IAAI,EAAE,EAAE,QAAQ,EAAE,EAAE,UAAU,EAAE,EAAE,IAAI,EAAE,SAAS,EAAE,EAAE,CAAC;yBACzD,CAAC,CAAC;wBACH,OAAO,QAAQ,CAAC;oBACjB,CAAC,CAAC;gBACH,CAAC;YACF,CAAC;YAGD,IAAI,gBAAgB,YAAY,4BAAY,EAAE,CAAC;gBAC9C,IAAI,IAAI,KAAK,WAAW,IAAI,WAAW,IAAI,MAAM,EAAE,CAAC;oBACnD,OAAO,KAAK,EAAE,IAAY,EAAqB,EAAE;wBAChD,cAAc,oBAAoC,CAAC;wBACnD,MAAM,EAAE,KAAK,EAAE,GAAG,gBAAgB,CAAC,YAAY,CAAC,cAAc,EAAE;4BAC/D,CAAC,EAAE,IAAI,EAAE,EAAE,YAAY,EAAE,IAAI,EAAE,EAAE,CAAC;yBAClC,CAAC,CAAC;wBAEH,MAAM,QAAQ,GAAG,CAAC,MAAM,eAAe,CAAC,IAAI,CAAC,MAAM,EAAE;4BACpD,gBAAgB;4BAChB,cAAc;4BACd,mBAAmB,EAAE,KAAK;4BAC1B,MAAM,EAAE,MAAM,CAAC,IAAI,CAAC;4BACpB,SAAS,EAAE,CAAC,IAAI,CAAC;yBACjB,CAAC,CAAa,CAAC;wBAEhB,KAAK,IAAA,oBAAU,EAAC,gBAAgB,EAAE,4BAA4B,CAAC,CAAC;wBAChE,gBAAgB,CAAC,aAAa,CAAC,cAAc,EAAE,KAAK,EAAE,CAAC,CAAC,EAAE,IAAI,EAAE,EAAE,QAAQ,EAAE,EAAE,CAAC,CAAC,CAAC,CAAC;wBAClF,OAAO,QAAQ,CAAC;oBACjB,CAAC,CAAC;gBACH,CAAC;YACF,CAAC;YAGD,IAAI,IAAA,yBAAe,EAAC,gBAAgB,CAAC,EAAE,CAAC;gBACvC,IAAI,IAAI,KAAK,OAAO,IAAI,OAAO,IAAI,MAAM,EAAE,CAAC;oBAC3C,OAAO,KAAK,EAAE,KAAa,EAAmB,EAAE;wBAC/C,cAAc,YAA4B,CAAC;wBAC3C,MAAM,EAAE,KAAK,EAAE,GAAG,gBAAgB,CAAC,YAAY,CAAC,cAAc,EAAE;4BAC/D,CAAC,EAAE,IAAI,EAAE,EAAE,KAAK,EAAE,EAAE,CAAC;yBACrB,CAAC,CAAC;wBAEH,MAAM,QAAQ,GAAG,CAAC,MAAM,eAAe,CAAC,IAAI,CAAC,MAAM,EAAE;4BACpD,gBAAgB;4BAChB,cAAc;4BACd,mBAAmB,EAAE,KAAK;4BAC1B,MAAM,EAAE,MAAM,CAAC,IAAI,CAAC;4BACpB,SAAS,EAAE,CAAC,KAAK,CAAC;yBAClB,CAAC,CAAW,CAAC;wBAEd,KAAK,IAAA,oBAAU,EAAC,gBAAgB,EAAE,oBAAoB,EAAE,EAAE,KAAK,EAAE,QAAQ,EAAE,CAAC,CAAC;wBAC7E,gBAAgB,CAAC,aAAa,CAAC,cAAc,EAAE,KAAK,EAAE,CAAC,CAAC,EAAE,IAAI,EAAE,EAAE,QAAQ,EAAE,EAAE,CAAC,CAAC,CAAC,CAAC;wBAClF,OAAO,QAAQ,CAAC;oBACjB,CAAC,CAAC;gBACH,CAAC;YACF,CAAC;YAGD,IAAI,gBAAgB,YAAY,0BAAW,EAAE,CAAC;gBAC7C,IAAI,IAAI,KAAK,kBAAkB,IAAI,kBAAkB,IAAI,MAAM,EAAE,CAAC;oBACjE,OAAO,KAAK,EACX,KAAa,EACb,CAAU,EAEV,MAAqC,EACrC,UAAkC,EACZ,EAAE;wBACxB,cAAc,mBAAmC,CAAC;wBAClD,MAAM,EAAE,KAAK,EAAE,GAAG,gBAAgB,CAAC,YAAY,CAAC,cAAc,EAAE;4BAC/D,CAAC,EAAE,IAAI,EAAE,EAAE,KAAK,EAAE,CAAC,EAAE,MAAM,EAAE,EAAE,CAAC;yBAChC,CAAC,CAAC;wBAEH,MAAM,QAAQ,GAAG,CAAC,MAAM,eAAe,CAAC,IAAI,CAAC,MAAM,EAAE;4BACpD,gBAAgB;4BAChB,cAAc;4BACd,mBAAmB,EAAE,KAAK;4BAC1B,MAAM,EAAE,MAAM,CAAC,IAAI,CAAC;4BACpB,SAAS,EAAE,CAAC,KAAK,EAAE,CAAC,EAAE,MAAM,EAAE,UAAU,CAAC;yBACzC,CAAC,CAAyC,CAAC;wBAE5C,KAAK,IAAA,oBAAU,EAAC,gBAAgB,EAAE,8BAA8B,EAAE,EAAE,KAAK,EAAE,CAAC,CAAC;wBAC7E,gBAAgB,CAAC,aAAa,CAAC,cAAc,EAAE,KAAK,EAAE,CAAC,CAAC,EAAE,IAAI,EAAE,EAAE,QAAQ,EAAE,EAAE,CAAC,CAAC,CAAC,CAAC;wBAElF,OAAO,QAAQ,CAAC;oBACjB,CAAC,CAAC;gBACH,CAAC;YACF,CAAC;YAGD,OAAQ,MAAc,CAAC,IAAI,CAAC,CAAC;QAC9B,CAAC;KACD,CAAC,CAAC;AACJ,CAAC"}
1
+ {"version":3,"file":"logWrapper.js","sourceRoot":"","sources":["../../utils/logWrapper.ts"],"names":[],"mappings":";;AA6BA,0CA6CC;AAED,wCA4BC;AAED,gCAmWC;AA7cD,+CAAsE;AAStE,2DAAwD;AACxD,+DAA2D;AAE3D,4DAAwD;AAExD,2DAA2D;AAC3D,mEAAyF;AACzF,mCAAkC;AAElC,mDAAgD;AAChD,uDAAoD;AACpD,uCAAoG;AAEpG,MAAM,SAAS,GAAgE;IAC9E,8EAA8E,EAAE;QAC/E,OAAO,EAAE,uBAAuB;QAChC,WAAW,EAAE,8EAA8E;KAC3F;CACD,CAAC;AAEK,KAAK,UAAU,eAAe,CAEpC,UAMC;IAED,IAAI,CAAC;QACJ,OAAO,MAAM,UAAU,CAAC,MAAM,CAAC,IAAI,CAAC,IAAI,EAAE,GAAG,UAAU,CAAC,SAAS,CAAC,CAAC;IACpE,CAAC;IAAC,OAAO,CAAC,EAAE,CAAC;QAGZ,IAAI,CAAC,YAAY,sCAAqB;YAAE,MAAM,CAAC,CAAC;QAGhD,IAAI,CAAC,CAAC,aAAa,KAAK,oBAAoB;YAAE,MAAM,CAAC,CAAC;QACtD,MAAM,aAAa,GAAG,UAAU,CAAC,gBAAgB,CAAC,OAAO,EAAE,CAAC;QAE5D,MAAM,KAAK,GAAG,IAAI,iCAAkB,CAAC,aAAa,EAAE,CAAC,EAAE;YACtD,aAAa,EAAE,oBAAoB;SACnC,CAAC,CAAC;QAEH,IAAI,SAAS,CAAC,KAAK,CAAC,OAAO,CAAC,EAAE,CAAC;YAC9B,KAAK,CAAC,WAAW,GAAG,SAAS,CAAC,KAAK,CAAC,OAAO,CAAC,CAAC,WAAW,CAAC;YACzD,KAAK,CAAC,OAAO,GAAG,SAAS,CAAC,KAAK,CAAC,OAAO,CAAC,CAAC,OAAO,CAAC;QAClD,CAAC;QAED,UAAU,CAAC,gBAAgB,CAAC,aAAa,CACxC,UAAU,CAAC,cAAc,EACzB,UAAU,CAAC,mBAAmB,EAC9B,KAAK,CACL,CAAC;QACF,IAAI,KAAK,CAAC,OAAO,EAAE,CAAC;YACnB,KAAK,CAAC,WAAW,GAAG,KAAK,CAAC,OAAO,CAAC;YAClC,MAAM,KAAK,CAAC;QACb,CAAC;QACD,MAAM,IAAI,iCAAkB,CAC3B,aAAa,EACb,kBAAkB,aAAa,CAAC,IAAI,mCAAmC,UAAU,CAAC,cAAc,GAAG,EACnG,EAAE,aAAa,EAAE,oBAAoB,EAAE,CACvC,CAAC;IACH,CAAC;AACF,CAAC;AAED,SAAgB,cAAc,CAE7B,UAMC;IAED,IAAI,CAAC;QACJ,OAAO,UAAU,CAAC,MAAM,CAAC,IAAI,CAAC,IAAI,EAAE,GAAG,UAAU,CAAC,SAAS,CAAC,CAAC;IAC9D,CAAC;IAAC,OAAO,CAAC,EAAE,CAAC;QAEZ,IAAI,CAAC,CAAC,aAAa,KAAK,oBAAoB;YAAE,MAAM,CAAC,CAAC;QACtD,MAAM,aAAa,GAAG,UAAU,CAAC,gBAAgB,CAAC,OAAO,EAAE,CAAC;QAC5D,MAAM,KAAK,GAAG,IAAI,iCAAkB,CAAC,aAAa,EAAE,CAAC,CAAC,CAAC;QACvD,UAAU,CAAC,gBAAgB,CAAC,aAAa,CACxC,UAAU,CAAC,cAAc,EACzB,UAAU,CAAC,mBAAmB,EAC9B,KAAK,CACL,CAAC;QACF,MAAM,IAAI,iCAAkB,CAC3B,aAAa,EACb,kBAAkB,aAAa,CAAC,IAAI,mCAAmC,UAAU,CAAC,cAAc,GAAG,EACnG,EAAE,aAAa,EAAE,oBAAoB,EAAE,CACvC,CAAC;IACH,CAAC;AACF,CAAC;AAED,SAAgB,UAAU,CACzB,gBAagB,EAChB,gBAAmC;IAEnC,OAAO,IAAI,KAAK,CAAC,gBAAgB,EAAE;QAClC,GAAG,EAAE,CAAC,MAAM,EAAE,IAAI,EAAE,EAAE;YACrB,IAAI,cAA8C,CAAC;YAEnD,IAAI,IAAA,0BAAgB,EAAC,gBAAgB,CAAC,EAAE,CAAC;gBACxC,IAAI,IAAI,KAAK,qBAAqB,IAAI,qBAAqB,IAAI,MAAM,EAAE,CAAC;oBACvE,OAAO,KAAK,EAAE,MAAmB,EAA4B,EAAE;wBAC9D,cAAc,cAA8B,CAAC;wBAE7C,MAAM,EAAE,KAAK,EAAE,GAAG,gBAAgB,CAAC,YAAY,CAAC,cAAc,EAAE;4BAC/D,CAAC,EAAE,IAAI,EAAE,EAAE,MAAM,EAAE,qBAAqB,EAAE,MAAM,EAAE,EAAE,CAAC;yBACrD,CAAC,CAAC;wBAEH,MAAM,QAAQ,GAAG,CAAC,MAAM,eAAe,CAAC,IAAI,CAAC,MAAM,EAAE;4BACpD,gBAAgB;4BAChB,cAAc;4BACd,mBAAmB,EAAE,KAAK;4BAC1B,MAAM,EAAE,MAAM,CAAC,IAAI,CAAC;4BACpB,SAAS,EAAE,CAAC,MAAM,CAAC;yBACnB,CAAC,CAAoB,CAAC;wBAEvB,MAAM,WAAW,GAAI,QAAQ,EAAE,YAA8B,IAAI,QAAQ,CAAC;wBAE1E,gBAAgB,CAAC,aAAa,CAAC,cAAc,EAAE,KAAK,EAAE;4BACrD,CAAC,EAAE,IAAI,EAAE,EAAE,MAAM,EAAE,qBAAqB,EAAE,WAAW,EAAE,EAAE,CAAC;yBAC1D,CAAC,CAAC;wBACH,OAAO,QAAQ,CAAC;oBACjB,CAAC,CAAC;gBACH,CAAC;qBAAM,IAAI,IAAI,KAAK,aAAa,IAAI,aAAa,IAAI,MAAM,EAAE,CAAC;oBAC9D,OAAO,KAAK,EAAE,KAAkB,EAAE,MAAoB,EAA4B,EAAE;wBACnF,cAAc,cAA8B,CAAC;wBAE7C,MAAM,EAAE,KAAK,EAAE,GAAG,gBAAgB,CAAC,YAAY,CAAC,cAAc,EAAE;4BAC/D,CAAC,EAAE,IAAI,EAAE,EAAE,MAAM,EAAE,aAAa,EAAE,KAAK,EAAE,MAAM,EAAE,EAAE,CAAC;yBACpD,CAAC,CAAC;wBAEH,MAAM,QAAQ,GAAG,CAAC,MAAM,eAAe,CAAC,IAAI,CAAC,MAAM,EAAE;4BACpD,gBAAgB;4BAChB,cAAc;4BACd,mBAAmB,EAAE,KAAK;4BAC1B,MAAM,EAAE,MAAM,CAAC,IAAI,CAAC;4BACpB,SAAS,EAAE,CAAC,KAAK,EAAE,MAAM,CAAC;yBAC1B,CAAC,CAAoB,CAAC;wBAEvB,MAAM,WAAW,GAAG,MAAM,MAAM,CAAC,WAAW,CAAC,WAAW,EAAE,CAAC;wBAE3D,gBAAgB,CAAC,aAAa,CAAC,cAAc,EAAE,KAAK,EAAE;4BACrD,CAAC,EAAE,IAAI,EAAE,EAAE,MAAM,EAAE,aAAa,EAAE,WAAW,EAAE,EAAE,CAAC;yBAClD,CAAC,CAAC;wBAEH,OAAO,QAAQ,CAAC;oBACjB,CAAC,CAAC;gBACH,CAAC;YACF,CAAC;YAGD,IAAI,IAAA,kCAAwB,EAAC,gBAAgB,CAAC,EAAE,CAAC;gBAChD,IAAI,IAAI,KAAK,aAAa,IAAI,aAAa,IAAI,MAAM,EAAE,CAAC;oBACvD,OAAO,KAAK,IAA4B,EAAE;wBACzC,cAAc,cAA8B,CAAC;wBAC7C,MAAM,EAAE,KAAK,EAAE,GAAG,gBAAgB,CAAC,YAAY,CAAC,cAAc,EAAE;4BAC/D,CAAC,EAAE,IAAI,EAAE,EAAE,MAAM,EAAE,aAAa,EAAE,EAAE,CAAC;yBACrC,CAAC,CAAC;wBAEH,MAAM,QAAQ,GAAG,CAAC,MAAM,eAAe,CAAC,IAAI,CAAC,MAAM,EAAE;4BACpD,gBAAgB;4BAChB,cAAc;4BACd,mBAAmB,EAAE,KAAK;4BAC1B,MAAM,EAAE,MAAM,CAAC,IAAI,CAAC;4BACpB,SAAS,EAAE,EAAE;yBACb,CAAC,CAAkB,CAAC;wBAErB,MAAM,OAAO,GAAG,EAAE,MAAM,EAAE,aAAa,EAAE,QAAQ,EAAE,CAAC;wBACpD,gBAAgB,CAAC,aAAa,CAAC,cAAc,EAAE,KAAK,EAAE,CAAC,CAAC,EAAE,IAAI,EAAE,OAAO,EAAE,CAAC,CAAC,CAAC,CAAC;wBAE7E,KAAK,IAAA,oBAAU,EAAC,gBAAgB,EAAE,mCAAmC,EAAE,EAAE,QAAQ,EAAE,CAAC,CAAC;wBACrF,OAAO,QAAQ,CAAC;oBACjB,CAAC,CAAC;gBACH,CAAC;qBAAM,IAAI,IAAI,KAAK,YAAY,IAAI,YAAY,IAAI,MAAM,EAAE,CAAC;oBAC5D,OAAO,KAAK,EAAE,OAAoB,EAAiB,EAAE;wBACpD,cAAc,cAA8B,CAAC;wBAC7C,MAAM,OAAO,GAAG,EAAE,MAAM,EAAE,YAAY,EAAE,OAAO,EAAE,CAAC;wBAClD,MAAM,EAAE,KAAK,EAAE,GAAG,gBAAgB,CAAC,YAAY,CAAC,cAAc,EAAE,CAAC,CAAC,EAAE,IAAI,EAAE,OAAO,EAAE,CAAC,CAAC,CAAC,CAAC;wBAEvF,MAAM,eAAe,CAAC,IAAI,CAAC,MAAM,EAAE;4BAClC,gBAAgB;4BAChB,cAAc;4BACd,mBAAmB,EAAE,KAAK;4BAC1B,MAAM,EAAE,MAAM,CAAC,IAAI,CAAC;4BACpB,SAAS,EAAE,CAAC,OAAO,CAAC;yBACpB,CAAC,CAAC;wBAEH,KAAK,IAAA,oBAAU,EAAC,gBAAgB,EAAE,4BAA4B,EAAE,EAAE,OAAO,EAAE,CAAC,CAAC;wBAC7E,gBAAgB,CAAC,aAAa,CAAC,cAAc,EAAE,KAAK,EAAE,CAAC,CAAC,EAAE,IAAI,EAAE,OAAO,EAAE,CAAC,CAAC,CAAC,CAAC;oBAC9E,CAAC,CAAC;gBACH,CAAC;YACF,CAAC;YAGD,IAAI,gBAAgB,YAAY,iCAAgB,EAAE,CAAC;gBAClD,IAAI,IAAI,KAAK,OAAO,IAAI,OAAO,IAAI,MAAM,EAAE,CAAC;oBAC3C,OAAO,KAAK,EAAE,IAAsC,EAAoB,EAAE;wBACzE,cAAc,oBAAoC,CAAC;wBACnD,MAAM,eAAe,GAAG,IAAA,iBAAQ,EAAC,IAAI,CAAC,CAAC,CAAC,CAAC,IAAI,CAAC,SAAS,CAAC,IAAI,CAAC,CAAC,CAAC,CAAC,IAAI,CAAC;wBACrE,MAAM,EAAE,KAAK,EAAE,GAAG,gBAAgB,CAAC,YAAY,CAAC,cAAc,EAAE;4BAC/D,CAAC,EAAE,IAAI,EAAE,EAAE,MAAM,EAAE,OAAO,EAAE,IAAI,EAAE,eAAe,EAAE,EAAE,CAAC;yBACtD,CAAC,CAAC;wBAEH,IAAI,CAAC;4BACJ,MAAM,QAAQ,GAAG,CAAC,MAAM,eAAe,CAAC,IAAI,CAAC,MAAM,EAAE;gCACpD,gBAAgB;gCAChB,cAAc;gCACd,mBAAmB,EAAE,KAAK;gCAC1B,MAAM,EAAE,MAAM,CAAC,IAAI,CAAC;gCACpB,SAAS,EAAE,CAAC,eAAe,CAAC;6BAC5B,CAAC,CAAW,CAAC;4BAEd,KAAK,IAAA,oBAAU,EAAC,gBAAgB,EAAE,kBAAkB,EAAE,EAAE,IAAI,EAAE,QAAQ,EAAE,CAAC,CAAC;4BAC1E,gBAAgB,CAAC,aAAa,CAAC,cAAc,EAAE,KAAK,EAAE;gCACrD,CAAC,EAAE,IAAI,EAAE,EAAE,MAAM,EAAE,OAAO,EAAE,QAAQ,EAAE,EAAE,CAAC;6BACzC,CAAC,CAAC;4BACH,OAAO,QAAQ,CAAC;wBACjB,CAAC;wBAAC,OAAO,KAAK,EAAE,CAAC;4BAChB,KAAK,IAAA,oBAAU,EAAC,gBAAgB,EAAE,kBAAkB,EAAE;gCACrD,IAAI;gCACJ,QAAQ,EAAE,KAAK,CAAC,OAAO,IAAI,KAAK;6BAChC,CAAC,CAAC;4BACH,gBAAgB,CAAC,aAAa,CAAC,cAAc,EAAE,KAAK,EAAE;gCACrD,CAAC,EAAE,IAAI,EAAE,EAAE,MAAM,EAAE,OAAO,EAAE,QAAQ,EAAE,KAAK,CAAC,OAAO,IAAI,KAAK,EAAE,EAAE,CAAC;6BACjE,CAAC,CAAC;4BACH,MAAM,KAAK,CAAC;wBACb,CAAC;oBACF,CAAC,CAAC;gBACH,CAAC;YACF,CAAC;YAGD,IAAI,gBAAgB,YAAY,0BAAa,EAAE,CAAC;gBAC/C,IAAI,IAAI,KAAK,sBAAsB,IAAI,sBAAsB,IAAI,MAAM,EAAE,CAAC;oBACzE,OAAO,KAAK,EACX,KAAa,EACb,MAAuC,EACjB,EAAE;wBACxB,cAAc,iBAAiC,CAAC;wBAChD,MAAM,EAAE,KAAK,EAAE,GAAG,gBAAgB,CAAC,YAAY,CAAC,cAAc,EAAE;4BAC/D,CAAC,EAAE,IAAI,EAAE,EAAE,KAAK,EAAE,MAAM,EAAE,EAAE,CAAC;yBAC7B,CAAC,CAAC;wBAEH,MAAM,QAAQ,GAAG,CAAC,MAAM,eAAe,CAAC,IAAI,CAAC,MAAM,EAAE;4BACpD,gBAAgB;4BAChB,cAAc;4BACd,mBAAmB,EAAE,KAAK;4BAC1B,MAAM,EAAE,MAAM,CAAC,IAAI,CAAC;4BACpB,SAAS,EAAE,CAAC,KAAK,EAAE,MAAM,CAAC;yBAC1B,CAAC,CAAyC,CAAC;wBAE5C,KAAK,IAAA,oBAAU,EAAC,gBAAgB,EAAE,wBAAwB,EAAE,EAAE,KAAK,EAAE,CAAC,CAAC;wBACvE,gBAAgB,CAAC,aAAa,CAAC,cAAc,EAAE,KAAK,EAAE,CAAC,CAAC,EAAE,IAAI,EAAE,EAAE,QAAQ,EAAE,EAAE,CAAC,CAAC,CAAC,CAAC;wBAClF,OAAO,QAAQ,CAAC;oBACjB,CAAC,CAAC;gBACH,CAAC;YACF,CAAC;YAGD,IAAI,gBAAgB,YAAY,uBAAU,EAAE,CAAC;gBAE5C,IAAI,IAAI,KAAK,gBAAgB,IAAI,gBAAgB,IAAI,MAAM,EAAE,CAAC;oBAC7D,OAAO,KAAK,EAAE,SAAmB,EAAuB,EAAE;wBACzD,cAAc,iBAAiC,CAAC;wBAChD,MAAM,EAAE,KAAK,EAAE,GAAG,gBAAgB,CAAC,YAAY,CAAC,cAAc,EAAE;4BAC/D,CAAC,EAAE,IAAI,EAAE,EAAE,SAAS,EAAE,EAAE,CAAC;yBACzB,CAAC,CAAC;wBAEH,MAAM,QAAQ,GAAG,CAAC,MAAM,eAAe,CAAC,IAAI,CAAC,MAAM,EAAE;4BACpD,gBAAgB;4BAChB,cAAc;4BACd,mBAAmB,EAAE,KAAK;4BAC1B,MAAM,EAAE,MAAM,CAAC,IAAI,CAAC;4BACpB,SAAS,EAAE,CAAC,SAAS,CAAC;yBACtB,CAAC,CAAe,CAAC;wBAElB,KAAK,IAAA,oBAAU,EAAC,gBAAgB,EAAE,sBAAsB,CAAC,CAAC;wBAC1D,gBAAgB,CAAC,aAAa,CAAC,cAAc,EAAE,KAAK,EAAE,CAAC,CAAC,EAAE,IAAI,EAAE,EAAE,QAAQ,EAAE,EAAE,CAAC,CAAC,CAAC,CAAC;wBAClF,OAAO,QAAQ,CAAC;oBACjB,CAAC,CAAC;gBACH,CAAC;gBAED,IAAI,IAAI,KAAK,YAAY,IAAI,YAAY,IAAI,MAAM,EAAE,CAAC;oBACrD,OAAO,KAAK,EAAE,KAAa,EAAqB,EAAE;wBACjD,cAAc,iBAAiC,CAAC;wBAChD,MAAM,EAAE,KAAK,EAAE,GAAG,gBAAgB,CAAC,YAAY,CAAC,cAAc,EAAE;4BAC/D,CAAC,EAAE,IAAI,EAAE,EAAE,KAAK,EAAE,EAAE,CAAC;yBACrB,CAAC,CAAC;wBAEH,MAAM,QAAQ,GAAG,CAAC,MAAM,eAAe,CAAC,IAAI,CAAC,MAAM,EAAE;4BACpD,gBAAgB;4BAChB,cAAc;4BACd,mBAAmB,EAAE,KAAK;4BAC1B,MAAM,EAAE,MAAM,CAAC,IAAI,CAAC;4BACpB,SAAS,EAAE,CAAC,KAAK,CAAC;yBAClB,CAAC,CAAa,CAAC;wBAChB,KAAK,IAAA,oBAAU,EAAC,gBAAgB,EAAE,mBAAmB,CAAC,CAAC;wBACvD,gBAAgB,CAAC,aAAa,CAAC,cAAc,EAAE,KAAK,EAAE,CAAC,CAAC,EAAE,IAAI,EAAE,EAAE,QAAQ,EAAE,EAAE,CAAC,CAAC,CAAC,CAAC;wBAClF,OAAO,QAAQ,CAAC;oBACjB,CAAC,CAAC;gBACH,CAAC;YACF,CAAC;YAGD,IACC,gBAAgB,YAAY,6BAAa;gBACzC,gBAAgB,YAAY,iCAAe,EAC1C,CAAC;gBAEF,IAAI,IAAI,KAAK,YAAY,IAAI,YAAY,IAAI,MAAM,EAAE,CAAC;oBACrD,OAAO,KAAK,EAAE,KAA2B,EAAqB,EAAE;wBAC/D,cAAc,gBAAgC,CAAC;wBAC/C,MAAM,EAAE,KAAK,EAAE,GAAG,gBAAgB,CAAC,YAAY,CAAC,cAAc,EAAE,CAAC,KAAK,CAAC,CAAC,CAAC;wBAEzE,MAAM,QAAQ,GAAG,CAAC,MAAM,eAAe,CAAC,IAAI,CAAC,MAAM,EAAE;4BACpD,gBAAgB;4BAChB,cAAc;4BACd,mBAAmB,EAAE,KAAK;4BAC1B,MAAM,EAAE,MAAM,CAAC,IAAI,CAAC;4BACpB,SAAS,EAAE,CAAC,KAAK,CAAC;yBAClB,CAAC,CAAa,CAAC;wBAEhB,gBAAgB,CAAC,aAAa,CAAC,cAAc,EAAE,KAAK,EAAE,CAAC,CAAC,EAAE,IAAI,EAAE,EAAE,QAAQ,EAAE,EAAE,CAAC,CAAC,CAAC,CAAC;wBAClF,OAAO,QAAQ,CAAC;oBACjB,CAAC,CAAC;gBACH,CAAC;gBAGD,IAAI,IAAI,KAAK,aAAa,IAAI,aAAa,IAAI,MAAM,EAAE,CAAC;oBACvD,OAAO,KAAK,EAAE,IAAwB,EAAE,SAAiB,EAAqB,EAAE;wBAC/E,cAAc,gBAAgC,CAAC;wBAC/C,MAAM,EAAE,KAAK,EAAE,GAAG,gBAAgB,CAAC,YAAY,CAAC,cAAc,EAAE,CAAC,CAAC,IAAI,CAAC,CAAC,CAAC,CAAC;wBAE1E,MAAM,QAAQ,GAAG,CAAC,MAAM,eAAe,CAAC,IAAI,CAAC,MAAM,EAAE;4BACpD,gBAAgB;4BAChB,cAAc;4BACd,mBAAmB,EAAE,KAAK;4BAC1B,MAAM,EAAE,MAAM,CAAC,IAAI,CAAC;4BACpB,SAAS,EAAE,CAAC,IAAI,EAAE,SAAS,CAAC;yBAC5B,CAAC,CAAa,CAAC;wBAEhB,KAAK,IAAA,oBAAU,EAAC,gBAAgB,EAAE,uBAAuB,CAAC,CAAC;wBAC3D,gBAAgB,CAAC,aAAa,CAAC,cAAc,EAAE,KAAK,EAAE;4BACrD,CAAC,EAAE,IAAI,EAAE,EAAE,QAAQ,EAAE,EAAE,UAAU,EAAE,EAAE,IAAI,EAAE,SAAS,EAAE,EAAE,CAAC;yBACzD,CAAC,CAAC;wBACH,OAAO,QAAQ,CAAC;oBACjB,CAAC,CAAC;gBACH,CAAC;YACF,CAAC;YAGD,IAAI,gBAAgB,YAAY,4BAAY,EAAE,CAAC;gBAC9C,IAAI,IAAI,KAAK,WAAW,IAAI,WAAW,IAAI,MAAM,EAAE,CAAC;oBACnD,OAAO,KAAK,EAAE,IAAY,EAAqB,EAAE;wBAChD,cAAc,oBAAoC,CAAC;wBACnD,MAAM,EAAE,KAAK,EAAE,GAAG,gBAAgB,CAAC,YAAY,CAAC,cAAc,EAAE;4BAC/D,CAAC,EAAE,IAAI,EAAE,EAAE,YAAY,EAAE,IAAI,EAAE,EAAE,CAAC;yBAClC,CAAC,CAAC;wBAEH,MAAM,QAAQ,GAAG,CAAC,MAAM,eAAe,CAAC,IAAI,CAAC,MAAM,EAAE;4BACpD,gBAAgB;4BAChB,cAAc;4BACd,mBAAmB,EAAE,KAAK;4BAC1B,MAAM,EAAE,MAAM,CAAC,IAAI,CAAC;4BACpB,SAAS,EAAE,CAAC,IAAI,CAAC;yBACjB,CAAC,CAAa,CAAC;wBAEhB,KAAK,IAAA,oBAAU,EAAC,gBAAgB,EAAE,eAAe,CAAC,CAAC;wBACnD,gBAAgB,CAAC,aAAa,CAAC,cAAc,EAAE,KAAK,EAAE,CAAC,CAAC,EAAE,IAAI,EAAE,EAAE,QAAQ,EAAE,EAAE,CAAC,CAAC,CAAC,CAAC;wBAClF,OAAO,QAAQ,CAAC;oBACjB,CAAC,CAAC;gBACH,CAAC;YACF,CAAC;YAGD,IAAI,IAAA,yBAAe,EAAC,gBAAgB,CAAC,EAAE,CAAC;gBACvC,IAAI,IAAI,KAAK,OAAO,IAAI,OAAO,IAAI,MAAM,EAAE,CAAC;oBAC3C,OAAO,KAAK,EAAE,KAAa,EAAmB,EAAE;wBAC/C,cAAc,YAA4B,CAAC;wBAC3C,MAAM,EAAE,KAAK,EAAE,GAAG,gBAAgB,CAAC,YAAY,CAAC,cAAc,EAAE;4BAC/D,CAAC,EAAE,IAAI,EAAE,EAAE,KAAK,EAAE,EAAE,CAAC;yBACrB,CAAC,CAAC;wBAEH,MAAM,QAAQ,GAAG,CAAC,MAAM,eAAe,CAAC,IAAI,CAAC,MAAM,EAAE;4BACpD,gBAAgB;4BAChB,cAAc;4BACd,mBAAmB,EAAE,KAAK;4BAC1B,MAAM,EAAE,MAAM,CAAC,IAAI,CAAC;4BACpB,SAAS,EAAE,CAAC,KAAK,CAAC;yBAClB,CAAC,CAAW,CAAC;wBAEd,KAAK,IAAA,oBAAU,EAAC,gBAAgB,EAAE,gBAAgB,EAAE,EAAE,KAAK,EAAE,QAAQ,EAAE,CAAC,CAAC;wBACzE,gBAAgB,CAAC,aAAa,CAAC,cAAc,EAAE,KAAK,EAAE,CAAC,CAAC,EAAE,IAAI,EAAE,EAAE,QAAQ,EAAE,EAAE,CAAC,CAAC,CAAC,CAAC;wBAClF,OAAO,QAAQ,CAAC;oBACjB,CAAC,CAAC;gBACH,CAAC;YACF,CAAC;YAGD,IAAI,gBAAgB,YAAY,0BAAW,EAAE,CAAC;gBAC7C,IAAI,IAAI,KAAK,kBAAkB,IAAI,kBAAkB,IAAI,MAAM,EAAE,CAAC;oBACjE,OAAO,KAAK,EACX,KAAa,EACb,CAAU,EAEV,MAAqC,EACrC,UAAkC,EACZ,EAAE;wBACxB,cAAc,mBAAmC,CAAC;wBAClD,MAAM,EAAE,KAAK,EAAE,GAAG,gBAAgB,CAAC,YAAY,CAAC,cAAc,EAAE;4BAC/D,CAAC,EAAE,IAAI,EAAE,EAAE,KAAK,EAAE,CAAC,EAAE,MAAM,EAAE,EAAE,CAAC;yBAChC,CAAC,CAAC;wBAEH,MAAM,QAAQ,GAAG,CAAC,MAAM,eAAe,CAAC,IAAI,CAAC,MAAM,EAAE;4BACpD,gBAAgB;4BAChB,cAAc;4BACd,mBAAmB,EAAE,KAAK;4BAC1B,MAAM,EAAE,MAAM,CAAC,IAAI,CAAC;4BACpB,SAAS,EAAE,CAAC,KAAK,EAAE,CAAC,EAAE,MAAM,EAAE,UAAU,CAAC;yBACzC,CAAC,CAAyC,CAAC;wBAE5C,KAAK,IAAA,oBAAU,EAAC,gBAAgB,EAAE,0BAA0B,EAAE,EAAE,KAAK,EAAE,CAAC,CAAC;wBACzE,gBAAgB,CAAC,aAAa,CAAC,cAAc,EAAE,KAAK,EAAE,CAAC,CAAC,EAAE,IAAI,EAAE,EAAE,QAAQ,EAAE,EAAE,CAAC,CAAC,CAAC,CAAC;wBAElF,OAAO,QAAQ,CAAC;oBACjB,CAAC,CAAC;gBACH,CAAC;YACF,CAAC;YAGD,OAAQ,MAAc,CAAC,IAAI,CAAC,CAAC;QAC9B,CAAC;KACD,CAAC,CAAC;AACJ,CAAC"}
package/package.json CHANGED
@@ -1,12 +1,12 @@
1
1
  {
2
2
  "name": "@n8n/n8n-nodes-langchain",
3
- "version": "1.59.0",
3
+ "version": "1.60.0",
4
4
  "description": "",
5
5
  "main": "index.js",
6
6
  "files": [
7
7
  "dist",
8
- "LICENSE_EE.md",
9
- "LICENSE.md"
8
+ "LICENSE.md",
9
+ "LICENSE_EE.md"
10
10
  ],
11
11
  "n8n": {
12
12
  "n8nNodesApiVersion": 1,
@@ -116,7 +116,7 @@
116
116
  "@types/html-to-text": "^9.0.1",
117
117
  "@types/json-schema": "^7.0.15",
118
118
  "@types/temp": "^0.9.1",
119
- "n8n-core": "1.59.0"
119
+ "n8n-core": "1.60.0"
120
120
  },
121
121
  "dependencies": {
122
122
  "@aws-sdk/client-bedrock-runtime": "3.645.0",
@@ -172,8 +172,8 @@
172
172
  "tmp-promise": "3.0.3",
173
173
  "zod": "3.23.8",
174
174
  "zod-to-json-schema": "3.23.2",
175
- "n8n-workflow": "1.58.0",
176
- "n8n-nodes-base": "1.59.0"
175
+ "n8n-workflow": "1.59.0",
176
+ "n8n-nodes-base": "1.60.0"
177
177
  },
178
178
  "license": "SEE LICENSE IN LICENSE.md",
179
179
  "homepage": "https://n8n.io",
@@ -191,7 +191,8 @@
191
191
  "typecheck": "tsc --noEmit",
192
192
  "build": "tsc -p tsconfig.build.json && pnpm n8n-copy-icons && pnpm build:metadata",
193
193
  "build:metadata": "pnpm n8n-generate-known && pnpm n8n-generate-ui-types",
194
- "format": "prettier nodes credentials --write",
194
+ "format": "biome format --write .",
195
+ "format:check": "biome ci .",
195
196
  "lint": "eslint nodes credentials --quiet",
196
197
  "lintfix": "eslint nodes credentials --fix",
197
198
  "watch": "tsc-watch -p tsconfig.build.json --onCompilationComplete \"tsc-alias -p tsconfig.build.json\" --onSuccess \"pnpm n8n-generate-ui-types\"",