@n8n/n8n-nodes-langchain 0.19.0 → 0.19.2

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -17,6 +17,10 @@ const modelField = {
17
17
  name: 'Claude 3 Sonnet(20240229)',
18
18
  value: 'claude-3-sonnet-20240229',
19
19
  },
20
+ {
21
+ name: 'Claude 3 Haiku(20240307)',
22
+ value: 'claude-3-haiku-20240307',
23
+ },
20
24
  {
21
25
  name: 'LEGACY: Claude 2',
22
26
  value: 'claude-2',
@@ -1 +1 @@
1
- {"version":3,"file":"LmChatAnthropic.node.js","sourceRoot":"","sources":["../../../../nodes/llms/LMChatAnthropic/LmChatAnthropic.node.ts"],"names":[],"mappings":";;;AAUA,oDAAqD;AACrD,0DAAuD;AACvD,8DAA2E;AAE3E,MAAM,UAAU,GAAoB;IACnC,WAAW,EAAE,OAAO;IACpB,IAAI,EAAE,OAAO;IACb,IAAI,EAAE,SAAS;IAEf,OAAO,EAAE;QACR;YACC,IAAI,EAAE,yBAAyB;YAC/B,KAAK,EAAE,wBAAwB;SAC/B;QACD;YACC,IAAI,EAAE,2BAA2B;YACjC,KAAK,EAAE,0BAA0B;SACjC;QACD;YACC,IAAI,EAAE,kBAAkB;YACxB,KAAK,EAAE,UAAU;SACjB;QACD;YACC,IAAI,EAAE,oBAAoB;YAC1B,KAAK,EAAE,YAAY;SACnB;QACD;YACC,IAAI,EAAE,4BAA4B;YAClC,KAAK,EAAE,oBAAoB;SAC3B;QACD;YACC,IAAI,EAAE,0BAA0B;YAChC,KAAK,EAAE,kBAAkB;SACzB;KACD;IACD,WAAW,EACV,gIAAgI;IACjI,OAAO,EAAE,UAAU;CACnB,CAAC;AAEF,MAAa,eAAe;IAA5B;QACC,gBAAW,GAAyB;YACnC,WAAW,EAAE,sBAAsB;YAEnC,IAAI,EAAE,iBAAiB;YACvB,IAAI,EAAE,oBAAoB;YAC1B,KAAK,EAAE,CAAC,WAAW,CAAC;YACpB,OAAO,EAAE,CAAC,CAAC,EAAE,GAAG,CAAC;YACjB,WAAW,EAAE,0BAA0B;YACvC,QAAQ,EAAE;gBACT,IAAI,EAAE,sBAAsB;aAC5B;YACD,KAAK,EAAE;gBACN,UAAU,EAAE,CAAC,IAAI,CAAC;gBAClB,aAAa,EAAE;oBACd,EAAE,EAAE,CAAC,iBAAiB,CAAC;iBACvB;gBACD,SAAS,EAAE;oBACV,oBAAoB,EAAE;wBACrB;4BACC,GAAG,EAAE,uGAAuG;yBAC5G;qBACD;iBACD;gBACD,KAAK,EAAE,CAAC,QAAQ,EAAE,QAAQ,EAAE,MAAM,CAAC;aACnC;YAED,MAAM,EAAE,EAAE;YAEV,OAAO,EAAE,oBAAoC;YAC7C,WAAW,EAAE,CAAC,OAAO,CAAC;YACtB,WAAW,EAAE;gBACZ;oBACC,IAAI,EAAE,cAAc;oBACpB,QAAQ,EAAE,IAAI;iBACd;aACD;YACD,UAAU,EAAE;gBACX,IAAA,2CAA4B,EAAC,wBAAwD,CAAC;gBACtF;oBACC,GAAG,UAAU;oBACb,cAAc,EAAE;wBACf,IAAI,EAAE;4BACL,UAAU,EAAE,CAAC,CAAC,CAAC;yBACf;qBACD;iBACD;gBACD;oBACC,GAAG,UAAU;oBACb,OAAO,EAAE,0BAA0B;oBACnC,cAAc,EAAE;wBACf,IAAI,EAAE;4BACL,UAAU,EAAE,CAAC,CAAC,CAAC;yBACf;qBACD;iBACD;gBACD;oBACC,WAAW,EAAE,SAAS;oBACtB,IAAI,EAAE,SAAS;oBACf,WAAW,EAAE,YAAY;oBACzB,WAAW,EAAE,2BAA2B;oBACxC,IAAI,EAAE,YAAY;oBAClB,OAAO,EAAE,EAAE;oBACX,OAAO,EAAE;wBACR;4BACC,WAAW,EAAE,0BAA0B;4BACvC,IAAI,EAAE,mBAAmB;4BACzB,OAAO,EAAE,IAAI;4BACb,WAAW,EAAE,4DAA4D;4BACzE,IAAI,EAAE,QAAQ;yBACd;wBACD;4BACC,WAAW,EAAE,sBAAsB;4BACnC,IAAI,EAAE,aAAa;4BACnB,OAAO,EAAE,GAAG;4BACZ,WAAW,EAAE,EAAE,QAAQ,EAAE,CAAC,EAAE,QAAQ,EAAE,CAAC,EAAE,eAAe,EAAE,CAAC,EAAE;4BAC7D,WAAW,EACV,2JAA2J;4BAC5J,IAAI,EAAE,QAAQ;yBACd;wBACD;4BACC,WAAW,EAAE,OAAO;4BACpB,IAAI,EAAE,MAAM;4BACZ,OAAO,EAAE,CAAC,CAAC;4BACX,WAAW,EAAE,EAAE,QAAQ,EAAE,CAAC,EAAE,QAAQ,EAAE,CAAC,CAAC,EAAE,eAAe,EAAE,CAAC,EAAE;4BAC9D,WAAW,EACV,0FAA0F;4BAC3F,IAAI,EAAE,QAAQ;yBACd;wBACD;4BACC,WAAW,EAAE,OAAO;4BACpB,IAAI,EAAE,MAAM;4BACZ,OAAO,EAAE,CAAC;4BACV,WAAW,EAAE,EAAE,QAAQ,EAAE,CAAC,EAAE,QAAQ,EAAE,CAAC,EAAE,eAAe,EAAE,CAAC,EAAE;4BAC7D,WAAW,EACV,8KAA8K;4BAC/K,IAAI,EAAE,QAAQ;yBACd;qBACD;iBACD;aACD;SACD,CAAC;IA0BH,CAAC;IAxBA,KAAK,CAAC,UAAU,CAA0B,SAAiB;QAC1D,MAAM,WAAW,GAAG,MAAM,IAAI,CAAC,cAAc,CAAC,cAAc,CAAC,CAAC;QAE9D,MAAM,SAAS,GAAG,IAAI,CAAC,gBAAgB,CAAC,OAAO,EAAE,SAAS,CAAW,CAAC;QACtE,MAAM,OAAO,GAAG,IAAI,CAAC,gBAAgB,CAAC,SAAS,EAAE,SAAS,EAAE,EAAE,CAK7D,CAAC;QAEF,MAAM,KAAK,GAAG,IAAI,yBAAa,CAAC;YAC/B,eAAe,EAAE,WAAW,CAAC,MAAgB;YAC7C,SAAS;YACT,SAAS,EAAE,OAAO,CAAC,iBAAiB;YACpC,WAAW,EAAE,OAAO,CAAC,WAAW;YAChC,IAAI,EAAE,OAAO,CAAC,IAAI;YAClB,IAAI,EAAE,OAAO,CAAC,IAAI;SAClB,CAAC,CAAC;QAEH,OAAO;YACN,QAAQ,EAAE,IAAA,uBAAU,EAAC,KAAK,EAAE,IAAI,CAAC;SACjC,CAAC;IACH,CAAC;CACD;AA/HD,0CA+HC"}
1
+ {"version":3,"file":"LmChatAnthropic.node.js","sourceRoot":"","sources":["../../../../nodes/llms/LMChatAnthropic/LmChatAnthropic.node.ts"],"names":[],"mappings":";;;AAUA,oDAAqD;AACrD,0DAAuD;AACvD,8DAA2E;AAE3E,MAAM,UAAU,GAAoB;IACnC,WAAW,EAAE,OAAO;IACpB,IAAI,EAAE,OAAO;IACb,IAAI,EAAE,SAAS;IAEf,OAAO,EAAE;QACR;YACC,IAAI,EAAE,yBAAyB;YAC/B,KAAK,EAAE,wBAAwB;SAC/B;QACD;YACC,IAAI,EAAE,2BAA2B;YACjC,KAAK,EAAE,0BAA0B;SACjC;QACD;YACC,IAAI,EAAE,0BAA0B;YAChC,KAAK,EAAE,yBAAyB;SAChC;QACD;YACC,IAAI,EAAE,kBAAkB;YACxB,KAAK,EAAE,UAAU;SACjB;QACD;YACC,IAAI,EAAE,oBAAoB;YAC1B,KAAK,EAAE,YAAY;SACnB;QACD;YACC,IAAI,EAAE,4BAA4B;YAClC,KAAK,EAAE,oBAAoB;SAC3B;QACD;YACC,IAAI,EAAE,0BAA0B;YAChC,KAAK,EAAE,kBAAkB;SACzB;KACD;IACD,WAAW,EACV,gIAAgI;IACjI,OAAO,EAAE,UAAU;CACnB,CAAC;AAEF,MAAa,eAAe;IAA5B;QACC,gBAAW,GAAyB;YACnC,WAAW,EAAE,sBAAsB;YAEnC,IAAI,EAAE,iBAAiB;YACvB,IAAI,EAAE,oBAAoB;YAC1B,KAAK,EAAE,CAAC,WAAW,CAAC;YACpB,OAAO,EAAE,CAAC,CAAC,EAAE,GAAG,CAAC;YACjB,WAAW,EAAE,0BAA0B;YACvC,QAAQ,EAAE;gBACT,IAAI,EAAE,sBAAsB;aAC5B;YACD,KAAK,EAAE;gBACN,UAAU,EAAE,CAAC,IAAI,CAAC;gBAClB,aAAa,EAAE;oBACd,EAAE,EAAE,CAAC,iBAAiB,CAAC;iBACvB;gBACD,SAAS,EAAE;oBACV,oBAAoB,EAAE;wBACrB;4BACC,GAAG,EAAE,uGAAuG;yBAC5G;qBACD;iBACD;gBACD,KAAK,EAAE,CAAC,QAAQ,EAAE,QAAQ,EAAE,MAAM,CAAC;aACnC;YAED,MAAM,EAAE,EAAE;YAEV,OAAO,EAAE,oBAAoC;YAC7C,WAAW,EAAE,CAAC,OAAO,CAAC;YACtB,WAAW,EAAE;gBACZ;oBACC,IAAI,EAAE,cAAc;oBACpB,QAAQ,EAAE,IAAI;iBACd;aACD;YACD,UAAU,EAAE;gBACX,IAAA,2CAA4B,EAAC,wBAAwD,CAAC;gBACtF;oBACC,GAAG,UAAU;oBACb,cAAc,EAAE;wBACf,IAAI,EAAE;4BACL,UAAU,EAAE,CAAC,CAAC,CAAC;yBACf;qBACD;iBACD;gBACD;oBACC,GAAG,UAAU;oBACb,OAAO,EAAE,0BAA0B;oBACnC,cAAc,EAAE;wBACf,IAAI,EAAE;4BACL,UAAU,EAAE,CAAC,CAAC,CAAC;yBACf;qBACD;iBACD;gBACD;oBACC,WAAW,EAAE,SAAS;oBACtB,IAAI,EAAE,SAAS;oBACf,WAAW,EAAE,YAAY;oBACzB,WAAW,EAAE,2BAA2B;oBACxC,IAAI,EAAE,YAAY;oBAClB,OAAO,EAAE,EAAE;oBACX,OAAO,EAAE;wBACR;4BACC,WAAW,EAAE,0BAA0B;4BACvC,IAAI,EAAE,mBAAmB;4BACzB,OAAO,EAAE,IAAI;4BACb,WAAW,EAAE,4DAA4D;4BACzE,IAAI,EAAE,QAAQ;yBACd;wBACD;4BACC,WAAW,EAAE,sBAAsB;4BACnC,IAAI,EAAE,aAAa;4BACnB,OAAO,EAAE,GAAG;4BACZ,WAAW,EAAE,EAAE,QAAQ,EAAE,CAAC,EAAE,QAAQ,EAAE,CAAC,EAAE,eAAe,EAAE,CAAC,EAAE;4BAC7D,WAAW,EACV,2JAA2J;4BAC5J,IAAI,EAAE,QAAQ;yBACd;wBACD;4BACC,WAAW,EAAE,OAAO;4BACpB,IAAI,EAAE,MAAM;4BACZ,OAAO,EAAE,CAAC,CAAC;4BACX,WAAW,EAAE,EAAE,QAAQ,EAAE,CAAC,EAAE,QAAQ,EAAE,CAAC,CAAC,EAAE,eAAe,EAAE,CAAC,EAAE;4BAC9D,WAAW,EACV,0FAA0F;4BAC3F,IAAI,EAAE,QAAQ;yBACd;wBACD;4BACC,WAAW,EAAE,OAAO;4BACpB,IAAI,EAAE,MAAM;4BACZ,OAAO,EAAE,CAAC;4BACV,WAAW,EAAE,EAAE,QAAQ,EAAE,CAAC,EAAE,QAAQ,EAAE,CAAC,EAAE,eAAe,EAAE,CAAC,EAAE;4BAC7D,WAAW,EACV,8KAA8K;4BAC/K,IAAI,EAAE,QAAQ;yBACd;qBACD;iBACD;aACD;SACD,CAAC;IA0BH,CAAC;IAxBA,KAAK,CAAC,UAAU,CAA0B,SAAiB;QAC1D,MAAM,WAAW,GAAG,MAAM,IAAI,CAAC,cAAc,CAAC,cAAc,CAAC,CAAC;QAE9D,MAAM,SAAS,GAAG,IAAI,CAAC,gBAAgB,CAAC,OAAO,EAAE,SAAS,CAAW,CAAC;QACtE,MAAM,OAAO,GAAG,IAAI,CAAC,gBAAgB,CAAC,SAAS,EAAE,SAAS,EAAE,EAAE,CAK7D,CAAC;QAEF,MAAM,KAAK,GAAG,IAAI,yBAAa,CAAC;YAC/B,eAAe,EAAE,WAAW,CAAC,MAAgB;YAC7C,SAAS;YACT,SAAS,EAAE,OAAO,CAAC,iBAAiB;YACpC,WAAW,EAAE,OAAO,CAAC,WAAW;YAChC,IAAI,EAAE,OAAO,CAAC,IAAI;YAClB,IAAI,EAAE,OAAO,CAAC,IAAI;SAClB,CAAC,CAAC;QAEH,OAAO;YACN,QAAQ,EAAE,IAAA,uBAAU,EAAC,KAAK,EAAE,IAAI,CAAC;SACjC,CAAC;IACH,CAAC;CACD;AA/HD,0CA+HC"}
@@ -19,7 +19,7 @@
19
19
  {"displayName":"Embeddings Mistral Cloud","name":"@n8n/n8n-nodes-langchain.embeddingsMistralCloud","credentials":[{"name":"mistralCloudApi","required":true}],"group":["transform"],"version":1,"description":"Use Embeddings Mistral Cloud","defaults":{"name":"Embeddings Mistral Cloud"},"codex":{"categories":["AI"],"subcategories":{"AI":["Embeddings"]},"resources":{"primaryDocumentation":[{"url":"https://docs.n8n.io/integrations/builtin/cluster-nodes/sub-nodes/n8n-nodes-langchain.embeddingsmistralcloud/"}]}},"inputs":[],"outputs":["ai_embedding"],"outputNames":["Embeddings"],"requestDefaults":{"ignoreHttpStatusErrors":true,"baseURL":"https://api.mistral.ai/v1"},"properties":[{"displayName":"This node must be connected to a vector store. <a data-action='openSelectiveNodeCreator' data-action-parameter-connectiontype='ai_vectorStore'>Insert one</a>","name":"notice","type":"notice","default":"","typeOptions":{"containerClass":"ndv-connection-hint-notice"}},{"displayName":"Model","name":"model","type":"options","description":"The model which will compute the embeddings. <a href=\"https://docs.mistral.ai/platform/endpoints/\">Learn more</a>.","typeOptions":{"loadOptions":{"routing":{"request":{"method":"GET","url":"/models"},"output":{"postReceive":[{"type":"rootProperty","properties":{"property":"data"}},{"type":"filter","properties":{"pass":"={{ $responseItem.id.includes('embed') }}"}},{"type":"setKeyValue","properties":{"name":"={{ $responseItem.id }}","value":"={{ $responseItem.id }}"}},{"type":"sort","properties":{"key":"name"}}]}}}},"routing":{"send":{"type":"body","property":"model"}},"default":"mistral-embed"},{"displayName":"Options","name":"options","placeholder":"Add Option","description":"Additional options to add","type":"collection","default":{},"options":[{"displayName":"Batch Size","name":"batchSize","default":512,"typeOptions":{"maxValue":2048},"description":"Maximum number of documents to send in each request","type":"number"},{"displayName":"Strip New Lines","name":"stripNewLines","default":true,"description":"Whether to strip new lines from the input text","type":"boolean"}]}],"iconUrl":"icons/@n8n/n8n-nodes-langchain/dist/nodes/embeddings/EmbeddingsMistralCloud/mistral.svg"},
20
20
  {"displayName":"Embeddings OpenAI","name":"@n8n/n8n-nodes-langchain.embeddingsOpenAi","credentials":[{"name":"openAiApi","required":true}],"group":["transform"],"version":1,"description":"Use Embeddings OpenAI","defaults":{"name":"Embeddings OpenAI"},"codex":{"categories":["AI"],"subcategories":{"AI":["Embeddings"]},"resources":{"primaryDocumentation":[{"url":"https://docs.n8n.io/integrations/builtin/cluster-nodes/sub-nodes/n8n-nodes-langchain.embeddingsopenai/"}]}},"inputs":[],"outputs":["ai_embedding"],"outputNames":["Embeddings"],"requestDefaults":{"ignoreHttpStatusErrors":true,"baseURL":"={{ $parameter.options?.baseURL?.split(\"/\").slice(0,-1).join(\"/\") || \"https://api.openai.com\" }}"},"properties":[{"displayName":"This node must be connected to a vector store. <a data-action='openSelectiveNodeCreator' data-action-parameter-connectiontype='ai_vectorStore'>Insert one</a>","name":"notice","type":"notice","default":"","typeOptions":{"containerClass":"ndv-connection-hint-notice"}},{"displayName":"Model","name":"model","type":"options","description":"The model which will generate the embeddings. <a href=\"https://platform.openai.com/docs/models/overview\">Learn more</a>.","typeOptions":{"loadOptions":{"routing":{"request":{"method":"GET","url":"={{ $parameter.options?.baseURL?.split(\"/\").slice(-1).pop() || \"v1\" }}/models"},"output":{"postReceive":[{"type":"rootProperty","properties":{"property":"data"}},{"type":"filter","properties":{"pass":"={{ $responseItem.id.includes('embed') }}"}},{"type":"setKeyValue","properties":{"name":"={{$responseItem.id}}","value":"={{$responseItem.id}}"}},{"type":"sort","properties":{"key":"name"}}]}}}},"routing":{"send":{"type":"body","property":"model"}},"default":"text-embedding-ada-002","displayOptions":{"show":{"@version":[1]}}},{"displayName":"Model","name":"model","type":"options","description":"The model which will generate the embeddings. <a href=\"https://platform.openai.com/docs/models/overview\">Learn more</a>.","typeOptions":{"loadOptions":{"routing":{"request":{"method":"GET","url":"={{ $parameter.options?.baseURL?.split(\"/\").slice(-1).pop() || \"v1\" }}/models"},"output":{"postReceive":[{"type":"rootProperty","properties":{"property":"data"}},{"type":"filter","properties":{"pass":"={{ $responseItem.id.includes('embed') }}"}},{"type":"setKeyValue","properties":{"name":"={{$responseItem.id}}","value":"={{$responseItem.id}}"}},{"type":"sort","properties":{"key":"name"}}]}}}},"routing":{"send":{"type":"body","property":"model"}},"default":"text-embedding-3-small","displayOptions":{"hide":{"@version":[1]}}},{"displayName":"Options","name":"options","placeholder":"Add Option","description":"Additional options to add","type":"collection","default":{},"options":[{"displayName":"Base URL","name":"baseURL","default":"https://api.openai.com/v1","description":"Override the default base URL for the API","type":"string"},{"displayName":"Batch Size","name":"batchSize","default":512,"typeOptions":{"maxValue":2048},"description":"Maximum number of documents to send in each request","type":"number"},{"displayName":"Strip New Lines","name":"stripNewLines","default":true,"description":"Whether to strip new lines from the input text","type":"boolean"},{"displayName":"Timeout","name":"timeout","default":-1,"description":"Maximum amount of time a request is allowed to take in seconds. Set to -1 for no timeout.","type":"number"}]}],"iconUrl":"icons/@n8n/n8n-nodes-langchain/dist/nodes/embeddings/EmbeddingsOpenAI/openAi.svg"},
21
21
  {"displayName":"Embeddings Ollama","name":"@n8n/n8n-nodes-langchain.embeddingsOllama","group":["transform"],"version":1,"description":"Use Ollama Embeddings","defaults":{"name":"Embeddings Ollama"},"credentials":[{"name":"ollamaApi","required":true}],"requestDefaults":{"ignoreHttpStatusErrors":true,"baseURL":"={{ $credentials.baseUrl.replace(new RegExp(\"/$\"), \"\") }}"},"codex":{"categories":["AI"],"subcategories":{"AI":["Embeddings"]},"resources":{"primaryDocumentation":[{"url":"https://docs.n8n.io/integrations/builtin/cluster-nodes/sub-nodes/n8n-nodes-langchain.embeddingsollama/"}]}},"inputs":[],"outputs":["ai_embedding"],"outputNames":["Embeddings"],"properties":[{"displayName":"This node must be connected to a vector store. <a data-action='openSelectiveNodeCreator' data-action-parameter-connectiontype='ai_vectorStore'>Insert one</a>","name":"notice","type":"notice","default":"","typeOptions":{"containerClass":"ndv-connection-hint-notice"}},{"displayName":"Model","name":"model","type":"options","default":"llama2","description":"The model which will generate the completion. To download models, visit <a href=\"https://ollama.ai/library\">Ollama Models Library</a>.","typeOptions":{"loadOptions":{"routing":{"request":{"method":"GET","url":"/api/tags"},"output":{"postReceive":[{"type":"rootProperty","properties":{"property":"models"}},{"type":"setKeyValue","properties":{"name":"={{$responseItem.name}}","value":"={{$responseItem.name}}"}},{"type":"sort","properties":{"key":"name"}}]}}}},"routing":{"send":{"type":"body","property":"model"}},"required":true}],"iconUrl":"icons/@n8n/n8n-nodes-langchain/dist/nodes/embeddings/EmbeddingsOllama/ollama.svg"},
22
- {"displayName":"Anthropic Chat Model","name":"@n8n/n8n-nodes-langchain.lmChatAnthropic","group":["transform"],"version":[1,1.1],"description":"Language Model Anthropic","defaults":{"name":"Anthropic Chat Model"},"codex":{"categories":["AI"],"subcategories":{"AI":["Language Models"]},"resources":{"primaryDocumentation":[{"url":"https://docs.n8n.io/integrations/builtin/cluster-nodes/sub-nodes/n8n-nodes-langchain.lmchatanthropic/"}]},"alias":["claude","sonnet","opus"]},"inputs":[],"outputs":["ai_languageModel"],"outputNames":["Model"],"credentials":[{"name":"anthropicApi","required":true}],"properties":[{"displayName":"This node must be connected to an AI chain. <a data-action='openSelectiveNodeCreator' data-action-parameter-creatorview='AI'>Insert one</a>","name":"notice","type":"notice","default":"","typeOptions":{"containerClass":"ndv-connection-hint-notice"}},{"displayName":"Model","name":"model","type":"options","options":[{"name":"Claude 3 Opus(20240229)","value":"claude-3-opus-20240229"},{"name":"Claude 3 Sonnet(20240229)","value":"claude-3-sonnet-20240229"},{"name":"LEGACY: Claude 2","value":"claude-2"},{"name":"LEGACY: Claude 2.1","value":"claude-2.1"},{"name":"LEGACY: Claude Instant 1.2","value":"claude-instant-1.2"},{"name":"LEGACY: Claude Instant 1","value":"claude-instant-1"}],"description":"The model which will generate the completion. <a href=\"https://docs.anthropic.com/claude/docs/models-overview\">Learn more</a>.","default":"claude-2","displayOptions":{"show":{"@version":[1]}}},{"displayName":"Model","name":"model","type":"options","options":[{"name":"Claude 3 Opus(20240229)","value":"claude-3-opus-20240229"},{"name":"Claude 3 Sonnet(20240229)","value":"claude-3-sonnet-20240229"},{"name":"LEGACY: Claude 2","value":"claude-2"},{"name":"LEGACY: Claude 2.1","value":"claude-2.1"},{"name":"LEGACY: Claude Instant 1.2","value":"claude-instant-1.2"},{"name":"LEGACY: Claude Instant 1","value":"claude-instant-1"}],"description":"The model which will generate the completion. <a href=\"https://docs.anthropic.com/claude/docs/models-overview\">Learn more</a>.","default":"claude-3-sonnet-20240229","displayOptions":{"hide":{"@version":[1]}}},{"displayName":"Options","name":"options","placeholder":"Add Option","description":"Additional options to add","type":"collection","default":{},"options":[{"displayName":"Maximum Number of Tokens","name":"maxTokensToSample","default":4096,"description":"The maximum number of tokens to generate in the completion","type":"number"},{"displayName":"Sampling Temperature","name":"temperature","default":0.7,"typeOptions":{"maxValue":1,"minValue":0,"numberPrecision":1},"description":"Controls randomness: Lowering results in less random completions. As the temperature approaches zero, the model will become deterministic and repetitive.","type":"number"},{"displayName":"Top K","name":"topK","default":-1,"typeOptions":{"maxValue":1,"minValue":-1,"numberPrecision":1},"description":"Used to remove \"long tail\" low probability responses. Defaults to -1, which disables it.","type":"number"},{"displayName":"Top P","name":"topP","default":1,"typeOptions":{"maxValue":1,"minValue":0,"numberPrecision":1},"description":"Controls diversity via nucleus sampling: 0.5 means half of all likelihood-weighted options are considered. We generally recommend altering this or temperature but not both.","type":"number"}]}],"iconUrl":"icons/@n8n/n8n-nodes-langchain/dist/nodes/llms/LMChatAnthropic/anthropic.svg"},
22
+ {"displayName":"Anthropic Chat Model","name":"@n8n/n8n-nodes-langchain.lmChatAnthropic","group":["transform"],"version":[1,1.1],"description":"Language Model Anthropic","defaults":{"name":"Anthropic Chat Model"},"codex":{"categories":["AI"],"subcategories":{"AI":["Language Models"]},"resources":{"primaryDocumentation":[{"url":"https://docs.n8n.io/integrations/builtin/cluster-nodes/sub-nodes/n8n-nodes-langchain.lmchatanthropic/"}]},"alias":["claude","sonnet","opus"]},"inputs":[],"outputs":["ai_languageModel"],"outputNames":["Model"],"credentials":[{"name":"anthropicApi","required":true}],"properties":[{"displayName":"This node must be connected to an AI chain. <a data-action='openSelectiveNodeCreator' data-action-parameter-creatorview='AI'>Insert one</a>","name":"notice","type":"notice","default":"","typeOptions":{"containerClass":"ndv-connection-hint-notice"}},{"displayName":"Model","name":"model","type":"options","options":[{"name":"Claude 3 Opus(20240229)","value":"claude-3-opus-20240229"},{"name":"Claude 3 Sonnet(20240229)","value":"claude-3-sonnet-20240229"},{"name":"Claude 3 Haiku(20240307)","value":"claude-3-haiku-20240307"},{"name":"LEGACY: Claude 2","value":"claude-2"},{"name":"LEGACY: Claude 2.1","value":"claude-2.1"},{"name":"LEGACY: Claude Instant 1.2","value":"claude-instant-1.2"},{"name":"LEGACY: Claude Instant 1","value":"claude-instant-1"}],"description":"The model which will generate the completion. <a href=\"https://docs.anthropic.com/claude/docs/models-overview\">Learn more</a>.","default":"claude-2","displayOptions":{"show":{"@version":[1]}}},{"displayName":"Model","name":"model","type":"options","options":[{"name":"Claude 3 Opus(20240229)","value":"claude-3-opus-20240229"},{"name":"Claude 3 Sonnet(20240229)","value":"claude-3-sonnet-20240229"},{"name":"Claude 3 Haiku(20240307)","value":"claude-3-haiku-20240307"},{"name":"LEGACY: Claude 2","value":"claude-2"},{"name":"LEGACY: Claude 2.1","value":"claude-2.1"},{"name":"LEGACY: Claude Instant 1.2","value":"claude-instant-1.2"},{"name":"LEGACY: Claude Instant 1","value":"claude-instant-1"}],"description":"The model which will generate the completion. <a href=\"https://docs.anthropic.com/claude/docs/models-overview\">Learn more</a>.","default":"claude-3-sonnet-20240229","displayOptions":{"hide":{"@version":[1]}}},{"displayName":"Options","name":"options","placeholder":"Add Option","description":"Additional options to add","type":"collection","default":{},"options":[{"displayName":"Maximum Number of Tokens","name":"maxTokensToSample","default":4096,"description":"The maximum number of tokens to generate in the completion","type":"number"},{"displayName":"Sampling Temperature","name":"temperature","default":0.7,"typeOptions":{"maxValue":1,"minValue":0,"numberPrecision":1},"description":"Controls randomness: Lowering results in less random completions. As the temperature approaches zero, the model will become deterministic and repetitive.","type":"number"},{"displayName":"Top K","name":"topK","default":-1,"typeOptions":{"maxValue":1,"minValue":-1,"numberPrecision":1},"description":"Used to remove \"long tail\" low probability responses. Defaults to -1, which disables it.","type":"number"},{"displayName":"Top P","name":"topP","default":1,"typeOptions":{"maxValue":1,"minValue":0,"numberPrecision":1},"description":"Controls diversity via nucleus sampling: 0.5 means half of all likelihood-weighted options are considered. We generally recommend altering this or temperature but not both.","type":"number"}]}],"iconUrl":"icons/@n8n/n8n-nodes-langchain/dist/nodes/llms/LMChatAnthropic/anthropic.svg"},
23
23
  {"displayName":"Azure OpenAI Chat Model","name":"@n8n/n8n-nodes-langchain.lmChatAzureOpenAi","group":["transform"],"version":1,"description":"For advanced usage with an AI chain","defaults":{"name":"Azure OpenAI Chat Model"},"codex":{"categories":["AI"],"subcategories":{"AI":["Language Models"]},"resources":{"primaryDocumentation":[{"url":"https://docs.n8n.io/integrations/builtin/cluster-nodes/sub-nodes/n8n-nodes-langchain.lmchatazureopenai/"}]}},"inputs":[],"outputs":["ai_languageModel"],"outputNames":["Model"],"credentials":[{"name":"azureOpenAiApi","required":true}],"properties":[{"displayName":"This node must be connected to an AI chain. <a data-action='openSelectiveNodeCreator' data-action-parameter-creatorview='AI'>Insert one</a>","name":"notice","type":"notice","default":"","typeOptions":{"containerClass":"ndv-connection-hint-notice"}},{"displayName":"Model (Deployment) Name","name":"model","type":"string","description":"The name of the model(deployment) to use","default":""},{"displayName":"Options","name":"options","placeholder":"Add Option","description":"Additional options to add","type":"collection","default":{},"options":[{"displayName":"Frequency Penalty","name":"frequencyPenalty","default":0,"typeOptions":{"maxValue":2,"minValue":-2,"numberPrecision":1},"description":"Positive values penalize new tokens based on their existing frequency in the text so far, decreasing the model's likelihood to repeat the same line verbatim","type":"number"},{"displayName":"Maximum Number of Tokens","name":"maxTokens","default":-1,"description":"The maximum number of tokens to generate in the completion. Most models have a context length of 2048 tokens (except for the newest models, which support 32,768).","type":"number","typeOptions":{"maxValue":32768}},{"displayName":"Presence Penalty","name":"presencePenalty","default":0,"typeOptions":{"maxValue":2,"minValue":-2,"numberPrecision":1},"description":"Positive values penalize new tokens based on whether they appear in the text so far, increasing the model's likelihood to talk about new topics","type":"number"},{"displayName":"Sampling Temperature","name":"temperature","default":0.7,"typeOptions":{"maxValue":1,"minValue":0,"numberPrecision":1},"description":"Controls randomness: Lowering results in less random completions. As the temperature approaches zero, the model will become deterministic and repetitive.","type":"number"},{"displayName":"Timeout","name":"timeout","default":60000,"description":"Maximum amount of time a request is allowed to take in milliseconds","type":"number"},{"displayName":"Max Retries","name":"maxRetries","default":2,"description":"Maximum number of retries to attempt","type":"number"},{"displayName":"Top P","name":"topP","default":1,"typeOptions":{"maxValue":1,"minValue":0,"numberPrecision":1},"description":"Controls diversity via nucleus sampling: 0.5 means half of all likelihood-weighted options are considered. We generally recommend altering this or temperature but not both.","type":"number"}]}],"iconUrl":"icons/@n8n/n8n-nodes-langchain/dist/nodes/llms/LmChatAzureOpenAi/azure.svg"},
24
24
  {"displayName":"Google PaLM Language Model","name":"@n8n/n8n-nodes-langchain.lmGooglePalm","group":["transform"],"version":1,"description":"Language Model Google PaLM","defaults":{"name":"Google PaLM Language Model"},"codex":{"categories":["AI"],"subcategories":{"AI":["Language Models"]},"resources":{"primaryDocumentation":[{"url":"https://docs.n8n.io/integrations/builtin/cluster-nodes/sub-nodes/n8n-nodes-langchain.lmgooglepalm/"}]}},"inputs":[],"outputs":["ai_languageModel"],"outputNames":["Model"],"credentials":[{"name":"googlePalmApi","required":true}],"requestDefaults":{"ignoreHttpStatusErrors":true,"baseURL":"={{ $credentials.host }}"},"properties":[{"displayName":"This node must be connected to an AI chain. <a data-action='openSelectiveNodeCreator' data-action-parameter-creatorview='AI'>Insert one</a>","name":"notice","type":"notice","default":"","typeOptions":{"containerClass":"ndv-connection-hint-notice"}},{"displayName":"Model","name":"modelName","type":"options","description":"The model which will generate the completion. <a href=\"https://developers.generativeai.google/api/rest/generativelanguage/models/list\">Learn more</a>.","typeOptions":{"loadOptions":{"routing":{"request":{"method":"GET","url":"/v1beta3/models"},"output":{"postReceive":[{"type":"rootProperty","properties":{"property":"models"}},{"type":"filter","properties":{"pass":"={{ $responseItem.name.startsWith('models/text') }}"}},{"type":"setKeyValue","properties":{"name":"={{$responseItem.name}}","value":"={{$responseItem.name}}","description":"={{$responseItem.description}}"}},{"type":"sort","properties":{"key":"name"}}]}}}},"routing":{"send":{"type":"body","property":"model"}},"default":"models/text-bison-001"},{"displayName":"Options","name":"options","placeholder":"Add Option","description":"Additional options to add","type":"collection","default":{},"options":[{"displayName":"Maximum Number of Tokens","name":"maxOutputTokens","default":1024,"description":"The maximum number of tokens to generate in the completion","type":"number"},{"displayName":"Sampling Temperature","name":"temperature","default":0.7,"typeOptions":{"maxValue":1,"minValue":0,"numberPrecision":1},"description":"Controls randomness: Lowering results in less random completions. As the temperature approaches zero, the model will become deterministic and repetitive.","type":"number"},{"displayName":"Top K","name":"topK","default":40,"typeOptions":{"maxValue":1,"minValue":-1,"numberPrecision":1},"description":"Used to remove \"long tail\" low probability responses. Defaults to -1, which disables it.","type":"number"},{"displayName":"Top P","name":"topP","default":0.9,"typeOptions":{"maxValue":1,"minValue":0,"numberPrecision":1},"description":"Controls diversity via nucleus sampling: 0.5 means half of all likelihood-weighted options are considered. We generally recommend altering this or temperature but not both.","type":"number"}]}],"iconUrl":"icons/@n8n/n8n-nodes-langchain/dist/nodes/llms/LmGooglePalm/google.svg"},
25
25
  {"displayName":"AWS Bedrock Chat Model","name":"@n8n/n8n-nodes-langchain.lmChatAwsBedrock","group":["transform"],"version":1,"description":"Language Model AWS Bedrock","defaults":{"name":"AWS Bedrock Chat Model"},"codex":{"categories":["AI"],"subcategories":{"AI":["Language Models"]},"resources":{"primaryDocumentation":[{"url":"https://docs.n8n.io/integrations/builtin/cluster-nodes/sub-nodes/n8n-nodes-langchain.lmchatawsbedrock/"}]}},"inputs":[],"outputs":["ai_languageModel"],"outputNames":["Model"],"credentials":[{"name":"aws","required":true}],"requestDefaults":{"ignoreHttpStatusErrors":true,"baseURL":"=https://bedrock.{{$credentials?.region ?? \"eu-central-1\"}}.amazonaws.com"},"properties":[{"displayName":"This node must be connected to an AI chain. <a data-action='openSelectiveNodeCreator' data-action-parameter-creatorview='AI'>Insert one</a>","name":"notice","type":"notice","default":"","typeOptions":{"containerClass":"ndv-connection-hint-notice"}},{"displayName":"Model","name":"model","type":"options","description":"The model which will generate the completion. <a href=\"https://docs.aws.amazon.com/bedrock/latest/userguide/foundation-models.html\">Learn more</a>.","typeOptions":{"loadOptions":{"routing":{"request":{"method":"GET","url":"/foundation-models?&byOutputModality=TEXT"},"output":{"postReceive":[{"type":"rootProperty","properties":{"property":"modelSummaries"}},{"type":"filter","properties":{"pass":"={{ !['anthropic.claude-instant-v1-100k'].includes($responseItem.modelId) }}"}},{"type":"setKeyValue","properties":{"name":"={{$responseItem.modelName}}","description":"={{$responseItem.modelArn}}","value":"={{$responseItem.modelId}}"}},{"type":"sort","properties":{"key":"name"}}]}}}},"routing":{"send":{"type":"body","property":"model"}},"default":""},{"displayName":"Options","name":"options","placeholder":"Add Option","description":"Additional options to add","type":"collection","default":{},"options":[{"displayName":"Maximum Number of Tokens","name":"maxTokensToSample","default":2000,"description":"The maximum number of tokens to generate in the completion","type":"number"},{"displayName":"Sampling Temperature","name":"temperature","default":0.7,"typeOptions":{"maxValue":1,"minValue":0,"numberPrecision":1},"description":"Controls randomness: Lowering results in less random completions. As the temperature approaches zero, the model will become deterministic and repetitive.","type":"number"}]}],"iconUrl":"icons/@n8n/n8n-nodes-langchain/dist/nodes/llms/LmChatAwsBedrock/bedrock.svg"},
@@ -1,10 +1,10 @@
1
1
  import type { EventNamesAiNodesType, IDataObject, IExecuteFunctions } from 'n8n-workflow';
2
- import { BaseChatModel } from '@langchain/core/language_models/chat_models';
2
+ import type { BaseChatModel } from '@langchain/core/language_models/chat_models';
3
3
  import type { BaseOutputParser } from '@langchain/core/output_parsers';
4
4
  import type { BaseMessage } from '@langchain/core/messages';
5
5
  import { type Tool } from '@langchain/core/tools';
6
6
  export declare function getMetadataFiltersValues(ctx: IExecuteFunctions, itemIndex: number): Record<string, never> | undefined;
7
- export declare function isChatInstance(model: any): model is BaseChatModel;
7
+ export declare function isChatInstance(model: unknown): model is BaseChatModel;
8
8
  export declare function getOptionalOutputParsers(ctx: IExecuteFunctions): Promise<Array<BaseOutputParser<unknown>>>;
9
9
  export declare function getPromptInputByType(options: {
10
10
  ctx: IExecuteFunctions;
@@ -2,7 +2,6 @@
2
2
  Object.defineProperty(exports, "__esModule", { value: true });
3
3
  exports.getConnectedTools = exports.serializeChatHistory = exports.logAiEvent = exports.getSessionId = exports.getPromptInputByType = exports.getOptionalOutputParsers = exports.isChatInstance = exports.getMetadataFiltersValues = void 0;
4
4
  const n8n_workflow_1 = require("n8n-workflow");
5
- const chat_models_1 = require("@langchain/core/language_models/chat_models");
6
5
  const tools_1 = require("@langchain/core/tools");
7
6
  function getMetadataFiltersValues(ctx, itemIndex) {
8
7
  const metadata = ctx.getNodeParameter('options.metadata.metadataValues', itemIndex, []);
@@ -13,7 +12,9 @@ function getMetadataFiltersValues(ctx, itemIndex) {
13
12
  }
14
13
  exports.getMetadataFiltersValues = getMetadataFiltersValues;
15
14
  function isChatInstance(model) {
16
- return model instanceof chat_models_1.BaseChatModel;
15
+ var _a;
16
+ const namespace = (_a = model === null || model === void 0 ? void 0 : model.lc_namespace) !== null && _a !== void 0 ? _a : [];
17
+ return namespace.includes('chat_models');
17
18
  }
18
19
  exports.isChatInstance = isChatInstance;
19
20
  async function getOptionalOutputParsers(ctx) {
@@ -1 +1 @@
1
- {"version":3,"file":"helpers.js","sourceRoot":"","sources":["../../utils/helpers.ts"],"names":[],"mappings":";;;AAAA,+CAAqF;AAErF,6EAA4E;AAG5E,iDAA+D;AAE/D,SAAgB,wBAAwB,CACvC,GAAsB,EACtB,SAAiB;IAEjB,MAAM,QAAQ,GAAG,GAAG,CAAC,gBAAgB,CAAC,iCAAiC,EAAE,SAAS,EAAE,EAAE,CAGpF,CAAC;IACH,IAAI,QAAQ,CAAC,MAAM,GAAG,CAAC,EAAE,CAAC;QACzB,OAAO,QAAQ,CAAC,MAAM,CAAC,CAAC,GAAG,EAAE,EAAE,IAAI,EAAE,KAAK,EAAE,EAAE,EAAE,CAAC,CAAC,EAAE,GAAG,GAAG,EAAE,CAAC,IAAI,CAAC,EAAE,KAAK,EAAE,CAAC,EAAE,EAAE,CAAC,CAAC;IACnF,CAAC;IAED,OAAO,SAAS,CAAC;AAClB,CAAC;AAbD,4DAaC;AAID,SAAgB,cAAc,CAAC,KAAU;IACxC,OAAO,KAAK,YAAY,2BAAa,CAAC;AACvC,CAAC;AAFD,wCAEC;AAEM,KAAK,UAAU,wBAAwB,CAC7C,GAAsB;IAEtB,IAAI,aAAa,GAAuB,EAAE,CAAC;IAE3C,IAAI,GAAG,CAAC,gBAAgB,CAAC,iBAAiB,EAAE,CAAC,EAAE,IAAI,CAAC,KAAK,IAAI,EAAE,CAAC;QAC/D,aAAa,GAAG,CAAC,MAAM,GAAG,CAAC,sBAAsB,oBAEhD,CAAC,CACD,CAAuB,CAAC;IAC1B,CAAC;IAED,OAAO,aAAa,CAAC;AACtB,CAAC;AAbD,4DAaC;AAED,SAAgB,oBAAoB,CAAC,OAKpC;IACA,MAAM,EAAE,GAAG,EAAE,CAAC,EAAE,aAAa,EAAE,QAAQ,EAAE,GAAG,OAAO,CAAC;IACpD,MAAM,MAAM,GAAG,GAAG,CAAC,gBAAgB,CAAC,aAAa,EAAE,CAAC,CAAW,CAAC;IAEhE,IAAI,KAAK,CAAC;IACV,IAAI,MAAM,KAAK,MAAM,EAAE,CAAC;QACvB,KAAK,GAAG,GAAG,CAAC,kBAAkB,CAAC,0BAA0B,EAAE,CAAC,CAAW,CAAC;IACzE,CAAC;SAAM,CAAC;QACP,KAAK,GAAG,GAAG,CAAC,gBAAgB,CAAC,QAAQ,EAAE,CAAC,CAAW,CAAC;IACrD,CAAC;IAED,IAAI,KAAK,KAAK,SAAS,EAAE,CAAC;QACzB,MAAM,IAAI,iCAAkB,CAAC,GAAG,CAAC,OAAO,EAAE,EAAE,qBAAqB,EAAE;YAClE,WAAW,EACV,qKAAqK;SACtK,CAAC,CAAC;IACJ,CAAC;IAED,OAAO,KAAK,CAAC;AACd,CAAC;AAxBD,oDAwBC;AAED,SAAgB,YAAY,CAC3B,GAAsB,EACtB,SAAiB,EACjB,WAAW,GAAG,eAAe,EAC7B,UAAU,GAAG,WAAW,EACxB,SAAS,GAAG,YAAY;IAExB,IAAI,SAAS,GAAG,EAAE,CAAC;IACnB,MAAM,YAAY,GAAG,GAAG,CAAC,gBAAgB,CAAC,WAAW,EAAE,SAAS,CAAW,CAAC;IAE5E,IAAI,YAAY,KAAK,UAAU,EAAE,CAAC;QACjC,SAAS,GAAG,GAAG,CAAC,kBAAkB,CAAC,uBAAuB,EAAE,SAAS,CAAW,CAAC;QACjF,IAAI,SAAS,KAAK,EAAE,IAAI,SAAS,KAAK,SAAS,EAAE,CAAC;YACjD,MAAM,IAAI,iCAAkB,CAAC,GAAG,CAAC,OAAO,EAAE,EAAE,qBAAqB,EAAE;gBAClE,WAAW,EACV,6KAA6K;gBAC9K,SAAS;aACT,CAAC,CAAC;QACJ,CAAC;IACF,CAAC;SAAM,CAAC;QACP,SAAS,GAAG,GAAG,CAAC,gBAAgB,CAAC,SAAS,EAAE,SAAS,EAAE,EAAE,CAAW,CAAC;QACrE,IAAI,SAAS,KAAK,EAAE,IAAI,SAAS,KAAK,SAAS,EAAE,CAAC;YACjD,MAAM,IAAI,iCAAkB,CAAC,GAAG,CAAC,OAAO,EAAE,EAAE,wBAAwB,EAAE;gBACrE,WAAW,EACV,0LAA0L;gBAC3L,SAAS;aACT,CAAC,CAAC;QACJ,CAAC;IACF,CAAC;IAED,OAAO,SAAS,CAAC;AAClB,CAAC;AA/BD,oCA+BC;AAEM,KAAK,UAAU,UAAU,CAC/B,gBAAmC,EACnC,KAA4B,EAC5B,IAAkB;IAElB,IAAI,CAAC;QACJ,MAAM,gBAAgB,CAAC,UAAU,CAAC,KAAK,EAAE,IAAI,CAAC,CAAC,CAAC,IAAA,4BAAa,EAAC,IAAI,CAAC,CAAC,CAAC,CAAC,SAAS,CAAC,CAAC;IAClF,CAAC;IAAC,OAAO,KAAK,EAAE,CAAC;QAChB,gBAAgB,CAAC,MAAM,CAAC,KAAK,CAAC,2BAA2B,KAAK,EAAE,CAAC,CAAC;IACnE,CAAC;AACF,CAAC;AAVD,gCAUC;AAED,SAAgB,oBAAoB,CAAC,WAA0B;IAC9D,OAAO,WAAW;SAChB,GAAG,CAAC,CAAC,WAAW,EAAE,EAAE;QACpB,IAAI,WAAW,CAAC,QAAQ,EAAE,KAAK,OAAO,EAAE,CAAC;YACxC,OAAO,UAAU,WAAW,CAAC,OAAO,EAAE,CAAC;QACxC,CAAC;aAAM,IAAI,WAAW,CAAC,QAAQ,EAAE,KAAK,IAAI,EAAE,CAAC;YAC5C,OAAO,cAAc,WAAW,CAAC,OAAO,EAAE,CAAC;QAC5C,CAAC;aAAM,CAAC;YACP,OAAO,GAAG,WAAW,CAAC,OAAO,EAAE,CAAC;QACjC,CAAC;IACF,CAAC,CAAC;SACD,IAAI,CAAC,IAAI,CAAC,CAAC;AACd,CAAC;AAZD,oDAYC;AAEM,MAAM,iBAAiB,GAAG,KAAK,EAAE,GAAsB,EAAE,kBAA2B,EAAE,EAAE;IAC9F,MAAM,cAAc,GAClB,CAAC,MAAM,GAAG,CAAC,sBAAsB,YAA4B,CAAC,CAAC,CAAY,IAAI,EAAE,CAAC;IAEpF,IAAI,CAAC,kBAAkB;QAAE,OAAO,cAAc,CAAC;IAE/C,MAAM,SAAS,GAAG,IAAI,GAAG,EAAU,CAAC;IAEpC,KAAK,MAAM,IAAI,IAAI,cAAc,EAAE,CAAC;QACnC,IAAI,CAAC,CAAC,IAAI,YAAY,mBAAW,CAAC;YAAE,SAAS;QAE7C,MAAM,EAAE,IAAI,EAAE,GAAG,IAAI,CAAC;QACtB,IAAI,SAAS,CAAC,GAAG,CAAC,IAAI,CAAC,EAAE,CAAC;YACzB,MAAM,IAAI,iCAAkB,CAC3B,GAAG,CAAC,OAAO,EAAE,EACb,gDAAgD,IAAI,0CAA0C,CAC9F,CAAC;QACH,CAAC;QACD,SAAS,CAAC,GAAG,CAAC,IAAI,CAAC,CAAC;IACrB,CAAC;IAED,OAAO,cAAc,CAAC;AACvB,CAAC,CAAC;AAtBW,QAAA,iBAAiB,qBAsB5B"}
1
+ {"version":3,"file":"helpers.js","sourceRoot":"","sources":["../../utils/helpers.ts"],"names":[],"mappings":";;;AAAA,+CAAqF;AAKrF,iDAA+D;AAG/D,SAAgB,wBAAwB,CACvC,GAAsB,EACtB,SAAiB;IAEjB,MAAM,QAAQ,GAAG,GAAG,CAAC,gBAAgB,CAAC,iCAAiC,EAAE,SAAS,EAAE,EAAE,CAGpF,CAAC;IACH,IAAI,QAAQ,CAAC,MAAM,GAAG,CAAC,EAAE,CAAC;QACzB,OAAO,QAAQ,CAAC,MAAM,CAAC,CAAC,GAAG,EAAE,EAAE,IAAI,EAAE,KAAK,EAAE,EAAE,EAAE,CAAC,CAAC,EAAE,GAAG,GAAG,EAAE,CAAC,IAAI,CAAC,EAAE,KAAK,EAAE,CAAC,EAAE,EAAE,CAAC,CAAC;IACnF,CAAC;IAED,OAAO,SAAS,CAAC;AAClB,CAAC;AAbD,4DAaC;AAED,SAAgB,cAAc,CAAC,KAAc;;IAC5C,MAAM,SAAS,GAAG,MAAC,KAAiC,aAAjC,KAAK,uBAAL,KAAK,CAA8B,YAAY,mCAAI,EAAE,CAAC;IAEzE,OAAO,SAAS,CAAC,QAAQ,CAAC,aAAa,CAAC,CAAC;AAC1C,CAAC;AAJD,wCAIC;AAEM,KAAK,UAAU,wBAAwB,CAC7C,GAAsB;IAEtB,IAAI,aAAa,GAAuB,EAAE,CAAC;IAE3C,IAAI,GAAG,CAAC,gBAAgB,CAAC,iBAAiB,EAAE,CAAC,EAAE,IAAI,CAAC,KAAK,IAAI,EAAE,CAAC;QAC/D,aAAa,GAAG,CAAC,MAAM,GAAG,CAAC,sBAAsB,oBAEhD,CAAC,CACD,CAAuB,CAAC;IAC1B,CAAC;IAED,OAAO,aAAa,CAAC;AACtB,CAAC;AAbD,4DAaC;AAED,SAAgB,oBAAoB,CAAC,OAKpC;IACA,MAAM,EAAE,GAAG,EAAE,CAAC,EAAE,aAAa,EAAE,QAAQ,EAAE,GAAG,OAAO,CAAC;IACpD,MAAM,MAAM,GAAG,GAAG,CAAC,gBAAgB,CAAC,aAAa,EAAE,CAAC,CAAW,CAAC;IAEhE,IAAI,KAAK,CAAC;IACV,IAAI,MAAM,KAAK,MAAM,EAAE,CAAC;QACvB,KAAK,GAAG,GAAG,CAAC,kBAAkB,CAAC,0BAA0B,EAAE,CAAC,CAAW,CAAC;IACzE,CAAC;SAAM,CAAC;QACP,KAAK,GAAG,GAAG,CAAC,gBAAgB,CAAC,QAAQ,EAAE,CAAC,CAAW,CAAC;IACrD,CAAC;IAED,IAAI,KAAK,KAAK,SAAS,EAAE,CAAC;QACzB,MAAM,IAAI,iCAAkB,CAAC,GAAG,CAAC,OAAO,EAAE,EAAE,qBAAqB,EAAE;YAClE,WAAW,EACV,qKAAqK;SACtK,CAAC,CAAC;IACJ,CAAC;IAED,OAAO,KAAK,CAAC;AACd,CAAC;AAxBD,oDAwBC;AAED,SAAgB,YAAY,CAC3B,GAAsB,EACtB,SAAiB,EACjB,WAAW,GAAG,eAAe,EAC7B,UAAU,GAAG,WAAW,EACxB,SAAS,GAAG,YAAY;IAExB,IAAI,SAAS,GAAG,EAAE,CAAC;IACnB,MAAM,YAAY,GAAG,GAAG,CAAC,gBAAgB,CAAC,WAAW,EAAE,SAAS,CAAW,CAAC;IAE5E,IAAI,YAAY,KAAK,UAAU,EAAE,CAAC;QACjC,SAAS,GAAG,GAAG,CAAC,kBAAkB,CAAC,uBAAuB,EAAE,SAAS,CAAW,CAAC;QACjF,IAAI,SAAS,KAAK,EAAE,IAAI,SAAS,KAAK,SAAS,EAAE,CAAC;YACjD,MAAM,IAAI,iCAAkB,CAAC,GAAG,CAAC,OAAO,EAAE,EAAE,qBAAqB,EAAE;gBAClE,WAAW,EACV,6KAA6K;gBAC9K,SAAS;aACT,CAAC,CAAC;QACJ,CAAC;IACF,CAAC;SAAM,CAAC;QACP,SAAS,GAAG,GAAG,CAAC,gBAAgB,CAAC,SAAS,EAAE,SAAS,EAAE,EAAE,CAAW,CAAC;QACrE,IAAI,SAAS,KAAK,EAAE,IAAI,SAAS,KAAK,SAAS,EAAE,CAAC;YACjD,MAAM,IAAI,iCAAkB,CAAC,GAAG,CAAC,OAAO,EAAE,EAAE,wBAAwB,EAAE;gBACrE,WAAW,EACV,0LAA0L;gBAC3L,SAAS;aACT,CAAC,CAAC;QACJ,CAAC;IACF,CAAC;IAED,OAAO,SAAS,CAAC;AAClB,CAAC;AA/BD,oCA+BC;AAEM,KAAK,UAAU,UAAU,CAC/B,gBAAmC,EACnC,KAA4B,EAC5B,IAAkB;IAElB,IAAI,CAAC;QACJ,MAAM,gBAAgB,CAAC,UAAU,CAAC,KAAK,EAAE,IAAI,CAAC,CAAC,CAAC,IAAA,4BAAa,EAAC,IAAI,CAAC,CAAC,CAAC,CAAC,SAAS,CAAC,CAAC;IAClF,CAAC;IAAC,OAAO,KAAK,EAAE,CAAC;QAChB,gBAAgB,CAAC,MAAM,CAAC,KAAK,CAAC,2BAA2B,KAAK,EAAE,CAAC,CAAC;IACnE,CAAC;AACF,CAAC;AAVD,gCAUC;AAED,SAAgB,oBAAoB,CAAC,WAA0B;IAC9D,OAAO,WAAW;SAChB,GAAG,CAAC,CAAC,WAAW,EAAE,EAAE;QACpB,IAAI,WAAW,CAAC,QAAQ,EAAE,KAAK,OAAO,EAAE,CAAC;YACxC,OAAO,UAAU,WAAW,CAAC,OAAO,EAAE,CAAC;QACxC,CAAC;aAAM,IAAI,WAAW,CAAC,QAAQ,EAAE,KAAK,IAAI,EAAE,CAAC;YAC5C,OAAO,cAAc,WAAW,CAAC,OAAO,EAAE,CAAC;QAC5C,CAAC;aAAM,CAAC;YACP,OAAO,GAAG,WAAW,CAAC,OAAO,EAAE,CAAC;QACjC,CAAC;IACF,CAAC,CAAC;SACD,IAAI,CAAC,IAAI,CAAC,CAAC;AACd,CAAC;AAZD,oDAYC;AAEM,MAAM,iBAAiB,GAAG,KAAK,EAAE,GAAsB,EAAE,kBAA2B,EAAE,EAAE;IAC9F,MAAM,cAAc,GAClB,CAAC,MAAM,GAAG,CAAC,sBAAsB,YAA4B,CAAC,CAAC,CAAY,IAAI,EAAE,CAAC;IAEpF,IAAI,CAAC,kBAAkB;QAAE,OAAO,cAAc,CAAC;IAE/C,MAAM,SAAS,GAAG,IAAI,GAAG,EAAU,CAAC;IAEpC,KAAK,MAAM,IAAI,IAAI,cAAc,EAAE,CAAC;QACnC,IAAI,CAAC,CAAC,IAAI,YAAY,mBAAW,CAAC;YAAE,SAAS;QAE7C,MAAM,EAAE,IAAI,EAAE,GAAG,IAAI,CAAC;QACtB,IAAI,SAAS,CAAC,GAAG,CAAC,IAAI,CAAC,EAAE,CAAC;YACzB,MAAM,IAAI,iCAAkB,CAC3B,GAAG,CAAC,OAAO,EAAE,EACb,gDAAgD,IAAI,0CAA0C,CAC9F,CAAC;QACH,CAAC;QACD,SAAS,CAAC,GAAG,CAAC,IAAI,CAAC,CAAC;IACrB,CAAC;IAED,OAAO,cAAc,CAAC;AACvB,CAAC,CAAC;AAtBW,QAAA,iBAAiB,qBAsB5B"}
@@ -27,4 +27,4 @@ export declare function callMethodSync<T>(this: T, parameters: {
27
27
  method: (...args: any[]) => T;
28
28
  arguments: unknown[];
29
29
  }): unknown;
30
- export declare function logWrapper(originalInstance: Tool | BaseChatModel | BaseChatMemory | BaseLLM | BaseChatMessageHistory | BaseOutputParser | BaseRetriever | Embeddings | Document[] | Document | BaseDocumentLoader | TextSplitter | VectorStore | N8nBinaryLoader | N8nJsonLoader, executeFunctions: IExecuteFunctions): BaseChatModel<import("langchain/base_language").BaseLanguageModelCallOptions> | BaseOutputParser<unknown> | Tool | BaseChatMemory | BaseChatMessageHistory | BaseRetriever<Record<string, any>> | N8nJsonLoader | TextSplitter | Document<Record<string, any>> | Document<Record<string, any>>[] | BaseDocumentLoader | N8nBinaryLoader | BaseLLM<import("@langchain/core/language_models/llms").BaseLLMCallOptions> | Embeddings | VectorStore;
30
+ export declare function logWrapper(originalInstance: Tool | BaseChatModel | BaseChatMemory | BaseLLM | BaseChatMessageHistory | BaseOutputParser | BaseRetriever | Embeddings | Document[] | Document | BaseDocumentLoader | TextSplitter | VectorStore | N8nBinaryLoader | N8nJsonLoader, executeFunctions: IExecuteFunctions): BaseChatModel<import("langchain/base_language").BaseLanguageModelCallOptions> | BaseLLM<import("@langchain/core/language_models/llms").BaseLLMCallOptions> | BaseOutputParser<unknown> | Tool | BaseChatMemory | BaseChatMessageHistory | BaseRetriever<Record<string, any>> | N8nJsonLoader | TextSplitter | Document<Record<string, any>> | Document<Record<string, any>>[] | BaseDocumentLoader | N8nBinaryLoader | Embeddings | VectorStore;
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@n8n/n8n-nodes-langchain",
3
- "version": "0.19.0",
3
+ "version": "0.19.2",
4
4
  "description": "",
5
5
  "license": "SEE LICENSE IN LICENSE.md",
6
6
  "homepage": "https://n8n.io",
@@ -111,7 +111,7 @@
111
111
  "@types/temp": "^0.9.1",
112
112
  "eslint-plugin-n8n-nodes-base": "^1.16.0",
113
113
  "gulp": "^4.0.2",
114
- "n8n-core": "1.33.0"
114
+ "n8n-core": "1.33.1"
115
115
  },
116
116
  "dependencies": {
117
117
  "@aws-sdk/client-bedrock-runtime": "3.454.0",
@@ -152,8 +152,8 @@
152
152
  "temp": "0.9.4",
153
153
  "zod": "3.22.4",
154
154
  "zod-to-json-schema": "3.22.4",
155
- "n8n-workflow": "1.33.0",
156
- "n8n-nodes-base": "1.34.0"
155
+ "n8n-workflow": "1.33.1",
156
+ "n8n-nodes-base": "1.34.1"
157
157
  },
158
158
  "scripts": {
159
159
  "clean": "rimraf dist .turbo",