@n8n/n8n-nodes-langchain 2.19.2 → 2.20.1
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/node-definitions/.nodes-hash +1 -1
- package/dist/node-definitions/nodes/n8n-nodes-langchain/agent/v1.schema.js +3 -3
- package/dist/node-definitions/nodes/n8n-nodes-langchain/agent/v1.ts +0 -1
- package/dist/node-definitions/nodes/n8n-nodes-langchain/agent/v11.schema.js +3 -3
- package/dist/node-definitions/nodes/n8n-nodes-langchain/agent/v11.ts +7 -7
- package/dist/node-definitions/nodes/n8n-nodes-langchain/agent/v12.schema.js +3 -3
- package/dist/node-definitions/nodes/n8n-nodes-langchain/agent/v12.ts +7 -7
- package/dist/node-definitions/nodes/n8n-nodes-langchain/agent/v13.schema.js +4 -4
- package/dist/node-definitions/nodes/n8n-nodes-langchain/agent/v13.ts +7 -8
- package/dist/node-definitions/nodes/n8n-nodes-langchain/agent/v14.schema.js +4 -4
- package/dist/node-definitions/nodes/n8n-nodes-langchain/agent/v14.ts +7 -8
- package/dist/node-definitions/nodes/n8n-nodes-langchain/agent/v15.schema.js +4 -4
- package/dist/node-definitions/nodes/n8n-nodes-langchain/agent/v15.ts +7 -8
- package/dist/node-definitions/nodes/n8n-nodes-langchain/agent/v16.schema.js +8 -8
- package/dist/node-definitions/nodes/n8n-nodes-langchain/agent/v16.ts +9 -10
- package/dist/node-definitions/nodes/n8n-nodes-langchain/agent/v17.schema.js +8 -8
- package/dist/node-definitions/nodes/n8n-nodes-langchain/agent/v17.ts +9 -10
- package/dist/node-definitions/nodes/n8n-nodes-langchain/agent/v18.schema.js +8 -8
- package/dist/node-definitions/nodes/n8n-nodes-langchain/agent/v18.ts +9 -10
- package/dist/node-definitions/nodes/n8n-nodes-langchain/agent/v19.schema.js +8 -8
- package/dist/node-definitions/nodes/n8n-nodes-langchain/agent/v19.ts +9 -10
- package/dist/node-definitions/nodes/n8n-nodes-langchain/chat/v1.schema.js +2 -2
- package/dist/node-definitions/nodes/n8n-nodes-langchain/chat/v11.schema.js +2 -2
- package/dist/node-definitions/nodes/n8n-nodes-langchain/chat/v12.schema.js +2 -2
- package/dist/node-definitions/nodes/n8n-nodes-langchain/chat/v13.schema.js +2 -2
- package/dist/node-definitions/nodes/n8n-nodes-langchain/chatTrigger/v1.schema.js +2 -2
- package/dist/node-definitions/nodes/n8n-nodes-langchain/chatTrigger/v1.ts +1 -0
- package/dist/node-definitions/nodes/n8n-nodes-langchain/chatTrigger/v11.schema.js +2 -2
- package/dist/node-definitions/nodes/n8n-nodes-langchain/chatTrigger/v11.ts +2 -1
- package/dist/node-definitions/nodes/n8n-nodes-langchain/chatTrigger/v12.ts +1 -0
- package/dist/node-definitions/nodes/n8n-nodes-langchain/chatTrigger/v13.schema.js +2 -2
- package/dist/node-definitions/nodes/n8n-nodes-langchain/chatTrigger/v13.ts +1 -0
- package/dist/node-definitions/nodes/n8n-nodes-langchain/chatTrigger/v14.schema.js +2 -2
- package/dist/node-definitions/nodes/n8n-nodes-langchain/chatTrigger/v14.ts +2 -1
- package/dist/node-definitions/nodes/n8n-nodes-langchain/lmChatAnthropic/index.ts +3 -1
- package/dist/node-definitions/nodes/n8n-nodes-langchain/lmChatAnthropic/v1.schema.js +1 -1
- package/dist/node-definitions/nodes/n8n-nodes-langchain/lmChatAnthropic/v1.ts +23 -4
- package/dist/node-definitions/nodes/n8n-nodes-langchain/lmChatAnthropic/v11.schema.js +1 -1
- package/dist/node-definitions/nodes/n8n-nodes-langchain/lmChatAnthropic/v11.ts +24 -5
- package/dist/node-definitions/nodes/n8n-nodes-langchain/lmChatAnthropic/v12.schema.js +2 -2
- package/dist/node-definitions/nodes/n8n-nodes-langchain/lmChatAnthropic/v12.ts +25 -6
- package/dist/node-definitions/nodes/n8n-nodes-langchain/lmChatAnthropic/v13.schema.js +1 -1
- package/dist/node-definitions/nodes/n8n-nodes-langchain/lmChatAnthropic/v13.ts +23 -4
- package/dist/node-definitions/nodes/n8n-nodes-langchain/lmChatAnthropic/v14.schema.js +1 -1
- package/dist/node-definitions/nodes/n8n-nodes-langchain/lmChatAnthropic/v14.ts +23 -4
- package/dist/node-definitions/nodes/n8n-nodes-langchain/lmChatAnthropic/v15.schema.js +25 -0
- package/dist/node-definitions/nodes/n8n-nodes-langchain/lmChatAnthropic/v15.ts +85 -0
- package/dist/node-definitions/nodes/n8n-nodes-langchain/lmChatAwsBedrock/v1.schema.js +2 -2
- package/dist/node-definitions/nodes/n8n-nodes-langchain/lmChatAwsBedrock/v1.ts +0 -1
- package/dist/node-definitions/nodes/n8n-nodes-langchain/lmChatAwsBedrock/v11.schema.js +2 -2
- package/dist/node-definitions/nodes/n8n-nodes-langchain/lmChatAwsBedrock/v11.ts +0 -1
- package/dist/node-definitions/nodes/n8n-nodes-langchain/lmChatOllama/v1.schema.js +1 -1
- package/dist/node-definitions/nodes/n8n-nodes-langchain/lmChatOllama/v1.ts +4 -0
- package/dist/node-definitions/nodes/n8n-nodes-langchain/lmOllama/v1.schema.js +1 -1
- package/dist/node-definitions/nodes/n8n-nodes-langchain/lmOllama/v1.ts +4 -0
- package/dist/node-definitions/nodes/n8n-nodes-langchain/mcpTrigger/v1.ts +1 -0
- package/dist/node-definitions/nodes/n8n-nodes-langchain/mcpTrigger/v11.ts +1 -0
- package/dist/node-definitions/nodes/n8n-nodes-langchain/mcpTrigger/v2.ts +1 -0
- package/dist/node-definitions/nodes/n8n-nodes-langchain/memoryBufferWindow/index.ts +3 -1
- package/dist/node-definitions/nodes/n8n-nodes-langchain/memoryBufferWindow/v1.schema.js +1 -1
- package/dist/node-definitions/nodes/n8n-nodes-langchain/memoryBufferWindow/v11.schema.js +1 -1
- package/dist/node-definitions/nodes/n8n-nodes-langchain/memoryBufferWindow/v14.schema.js +26 -0
- package/dist/node-definitions/nodes/n8n-nodes-langchain/memoryBufferWindow/v14.ts +37 -0
- package/dist/node-definitions/nodes/n8n-nodes-langchain/memoryMongoDbChat/index.ts +3 -1
- package/dist/node-definitions/nodes/n8n-nodes-langchain/memoryMongoDbChat/v11.schema.js +28 -0
- package/dist/node-definitions/nodes/n8n-nodes-langchain/memoryMongoDbChat/v11.ts +51 -0
- package/dist/node-definitions/nodes/n8n-nodes-langchain/memoryPostgresChat/index.ts +3 -1
- package/dist/node-definitions/nodes/n8n-nodes-langchain/memoryPostgresChat/v14.schema.js +27 -0
- package/dist/node-definitions/nodes/n8n-nodes-langchain/memoryPostgresChat/v14.ts +47 -0
- package/dist/node-definitions/nodes/n8n-nodes-langchain/memoryRedisChat/index.ts +3 -1
- package/dist/node-definitions/nodes/n8n-nodes-langchain/memoryRedisChat/v1.schema.js +1 -1
- package/dist/node-definitions/nodes/n8n-nodes-langchain/memoryRedisChat/v11.schema.js +1 -1
- package/dist/node-definitions/nodes/n8n-nodes-langchain/memoryRedisChat/v16.schema.js +27 -0
- package/dist/node-definitions/nodes/n8n-nodes-langchain/memoryRedisChat/v16.ts +47 -0
- package/dist/node-definitions/nodes/n8n-nodes-langchain/memoryXata/index.ts +3 -1
- package/dist/node-definitions/nodes/n8n-nodes-langchain/memoryXata/v15.schema.js +26 -0
- package/dist/node-definitions/nodes/n8n-nodes-langchain/memoryXata/v15.ts +42 -0
- package/dist/node-definitions/nodes/n8n-nodes-langchain/ollama/v1/resource_image/operation_analyze.schema.js +1 -1
- package/dist/node-definitions/nodes/n8n-nodes-langchain/ollama/v1/resource_image/operation_analyze.ts +4 -0
- package/dist/node-definitions/nodes/n8n-nodes-langchain/ollama/v1/resource_text/operation_message.schema.js +1 -1
- package/dist/node-definitions/nodes/n8n-nodes-langchain/ollama/v1/resource_text/operation_message.ts +4 -0
- package/dist/node-definitions/nodes/n8n-nodes-langchain/openAi/v1/resource_assistant/operation_create.schema.js +2 -2
- package/dist/node-definitions/nodes/n8n-nodes-langchain/openAi/v1/resource_assistant/operation_create.ts +0 -1
- package/dist/node-definitions/nodes/n8n-nodes-langchain/openAi/v11/resource_assistant/operation_create.schema.js +2 -2
- package/dist/node-definitions/nodes/n8n-nodes-langchain/openAi/v11/resource_assistant/operation_create.ts +1 -2
- package/dist/node-definitions/nodes/n8n-nodes-langchain/openAi/v12/resource_assistant/operation_create.schema.js +2 -2
- package/dist/node-definitions/nodes/n8n-nodes-langchain/openAi/v12/resource_assistant/operation_create.ts +1 -2
- package/dist/node-definitions/nodes/n8n-nodes-langchain/openAi/v13/resource_assistant/operation_create.schema.js +2 -2
- package/dist/node-definitions/nodes/n8n-nodes-langchain/openAi/v13/resource_assistant/operation_create.ts +1 -2
- package/dist/node-definitions/nodes/n8n-nodes-langchain/openAi/v14/resource_assistant/operation_create.schema.js +2 -2
- package/dist/node-definitions/nodes/n8n-nodes-langchain/openAi/v14/resource_assistant/operation_create.ts +1 -2
- package/dist/node-definitions/nodes/n8n-nodes-langchain/openAi/v15/resource_assistant/operation_create.schema.js +2 -2
- package/dist/node-definitions/nodes/n8n-nodes-langchain/openAi/v15/resource_assistant/operation_create.ts +1 -2
- package/dist/node-definitions/nodes/n8n-nodes-langchain/openAi/v16/resource_assistant/operation_create.schema.js +2 -2
- package/dist/node-definitions/nodes/n8n-nodes-langchain/openAi/v16/resource_assistant/operation_create.ts +1 -2
- package/dist/node-definitions/nodes/n8n-nodes-langchain/openAi/v17/resource_assistant/operation_create.schema.js +2 -2
- package/dist/node-definitions/nodes/n8n-nodes-langchain/openAi/v17/resource_assistant/operation_create.ts +1 -2
- package/dist/node-definitions/nodes/n8n-nodes-langchain/openAi/v18/resource_assistant/operation_create.schema.js +2 -2
- package/dist/node-definitions/nodes/n8n-nodes-langchain/openAi/v18/resource_assistant/operation_create.ts +1 -2
- package/dist/node-definitions/nodes/n8n-nodes-langchain/openAi/v2/resource_image/operation_edit.schema.js +7 -7
- package/dist/node-definitions/nodes/n8n-nodes-langchain/openAi/v21/resource_image/operation_edit.schema.js +7 -7
- package/dist/node-definitions/nodes/n8n-nodes-langchain/openAi/v21/resource_image/operation_edit.ts +6 -6
- package/dist/node-definitions/nodes/n8n-nodes-langchain/openAi/v22/resource_image/operation_edit.schema.js +7 -7
- package/dist/node-definitions/nodes/n8n-nodes-langchain/openAi/v22/resource_image/operation_edit.ts +6 -6
- package/dist/node-definitions/nodes/n8n-nodes-langchain/openAi/v23/resource_image/operation_edit.schema.js +7 -7
- package/dist/node-definitions/nodes/n8n-nodes-langchain/openAi/v23/resource_image/operation_edit.ts +6 -6
- package/dist/node-definitions/nodes/n8n-nodes-langchain/retrieverWorkflow/v11.schema.js +2 -2
- package/dist/nodes/llms/LMChatAnthropic/LmChatAnthropic.node.js +169 -20
- package/dist/nodes/llms/LMChatAnthropic/LmChatAnthropic.node.js.map +1 -1
- package/dist/nodes/llms/LMOllama/LmOllama.node.js.map +1 -1
- package/dist/nodes/llms/LMOllama/description.js +7 -0
- package/dist/nodes/llms/LMOllama/description.js.map +1 -1
- package/dist/nodes/mcp/McpTrigger/McpTrigger.node.js +3 -0
- package/dist/nodes/mcp/McpTrigger/McpTrigger.node.js.map +1 -1
- package/dist/nodes/memory/MemoryBufferWindow/MemoryBufferWindow.node.js +2 -1
- package/dist/nodes/memory/MemoryBufferWindow/MemoryBufferWindow.node.js.map +1 -1
- package/dist/nodes/memory/MemoryMongoDbChat/MemoryMongoDbChat.node.js +2 -1
- package/dist/nodes/memory/MemoryMongoDbChat/MemoryMongoDbChat.node.js.map +1 -1
- package/dist/nodes/memory/MemoryMotorhead/MemoryMotorhead.node.js +2 -1
- package/dist/nodes/memory/MemoryMotorhead/MemoryMotorhead.node.js.map +1 -1
- package/dist/nodes/memory/MemoryPostgresChat/MemoryPostgresChat.node.js +2 -1
- package/dist/nodes/memory/MemoryPostgresChat/MemoryPostgresChat.node.js.map +1 -1
- package/dist/nodes/memory/MemoryRedisChat/MemoryRedisChat.node.js +2 -1
- package/dist/nodes/memory/MemoryRedisChat/MemoryRedisChat.node.js.map +1 -1
- package/dist/nodes/memory/MemoryXata/MemoryXata.node.js +2 -1
- package/dist/nodes/memory/MemoryXata/MemoryXata.node.js.map +1 -1
- package/dist/nodes/memory/MemoryZep/MemoryZep.node.js +2 -1
- package/dist/nodes/memory/MemoryZep/MemoryZep.node.js.map +1 -1
- package/dist/nodes/memory/descriptions.d.ts +1 -0
- package/dist/nodes/memory/descriptions.js +14 -1
- package/dist/nodes/memory/descriptions.js.map +1 -1
- package/dist/nodes/trigger/ChatTrigger/ChatTrigger.node.js +3 -0
- package/dist/nodes/trigger/ChatTrigger/ChatTrigger.node.js.map +1 -1
- package/dist/nodes/vector_store/VectorStorePGVector/VectorStorePGVector.node.d.ts +1 -1
- package/dist/nodes/vendors/Ollama/actions/image/analyze.operation.js +9 -1
- package/dist/nodes/vendors/Ollama/actions/image/analyze.operation.js.map +1 -1
- package/dist/nodes/vendors/Ollama/actions/text/message.operation.js +11 -4
- package/dist/nodes/vendors/Ollama/actions/text/message.operation.js.map +1 -1
- package/dist/nodes/vendors/Ollama/transport/index.js +1 -1
- package/dist/nodes/vendors/Ollama/transport/index.js.map +1 -1
- package/dist/nodes/vendors/OpenAi/v2/actions/text/response.operation.js.map +1 -1
- package/dist/typecheck.tsbuildinfo +1 -1
- package/dist/types/nodes.json +28 -28
- package/dist/utils/helpers.js +24 -0
- package/dist/utils/helpers.js.map +1 -1
- package/dist/utils/output_parsers/N8nOutputFixingParser.d.ts +1 -1
- package/package.json +18 -18
|
@@ -18,6 +18,7 @@ export interface LcChatTriggerV14Params {
|
|
|
18
18
|
mode?: 'hostedChat' | 'webhook' | Expression<string>;
|
|
19
19
|
/**
|
|
20
20
|
* The way to authenticate
|
|
21
|
+
* @builderHint Default to 'none'. n8n exposes inbound trigger URLs publicly by design. Only select an authentication method when the user explicitly asks to authenticate inbound traffic.
|
|
21
22
|
* @displayOptions.show { public: [true] }
|
|
22
23
|
* @default none
|
|
23
24
|
*/
|
|
@@ -68,7 +69,7 @@ export interface LcChatTriggerV14Params {
|
|
|
68
69
|
};
|
|
69
70
|
/**
|
|
70
71
|
* Options
|
|
71
|
-
* @displayOptions.show { public: [false
|
|
72
|
+
* @displayOptions.show { public: [false] }
|
|
72
73
|
* @default {}
|
|
73
74
|
*/
|
|
74
75
|
options?: {
|
|
@@ -4,12 +4,14 @@
|
|
|
4
4
|
* Re-exports all version-specific types and provides combined union type.
|
|
5
5
|
*/
|
|
6
6
|
|
|
7
|
+
import type { LcLmChatAnthropicV15Node } from './v15';
|
|
7
8
|
import type { LcLmChatAnthropicV14Node } from './v14';
|
|
8
9
|
import type { LcLmChatAnthropicV13Node } from './v13';
|
|
9
10
|
import type { LcLmChatAnthropicV12Node } from './v12';
|
|
10
11
|
import type { LcLmChatAnthropicV11Node } from './v11';
|
|
11
12
|
import type { LcLmChatAnthropicV1Node } from './v1';
|
|
12
13
|
|
|
14
|
+
export * from './v15';
|
|
13
15
|
export * from './v14';
|
|
14
16
|
export * from './v13';
|
|
15
17
|
export * from './v12';
|
|
@@ -17,4 +19,4 @@ export * from './v11';
|
|
|
17
19
|
export * from './v1';
|
|
18
20
|
|
|
19
21
|
// Combined union type for all versions
|
|
20
|
-
export type LcLmChatAnthropicNode = LcLmChatAnthropicV14Node | LcLmChatAnthropicV13Node | LcLmChatAnthropicV12Node | LcLmChatAnthropicV11Node | LcLmChatAnthropicV1Node;
|
|
22
|
+
export type LcLmChatAnthropicNode = LcLmChatAnthropicV15Node | LcLmChatAnthropicV14Node | LcLmChatAnthropicV13Node | LcLmChatAnthropicV12Node | LcLmChatAnthropicV11Node | LcLmChatAnthropicV1Node;
|
|
@@ -15,7 +15,7 @@ module.exports = function getSchema({ parameters, z, expressionSchema, stringOrE
|
|
|
15
15
|
// Parameters schema
|
|
16
16
|
const parametersSchema = z.object({
|
|
17
17
|
model: z.union([z.literal('claude-3-5-sonnet-20241022'), z.literal('claude-3-opus-20240229'), z.literal('claude-3-5-sonnet-20240620'), z.literal('claude-3-sonnet-20240229'), z.literal('claude-3-5-haiku-20241022'), z.literal('claude-3-haiku-20240307'), z.literal('claude-2'), z.literal('claude-2.1'), z.literal('claude-instant-1.2'), z.literal('claude-instant-1'), expressionSchema]).optional(),
|
|
18
|
-
options: z.object({ maxTokensToSample: numberOrExpression.optional(), temperature: numberOrExpression.optional(), topK: numberOrExpression.optional(), topP: numberOrExpression.optional(), thinking: booleanOrExpression.optional(), thinkingBudget: numberOrExpression.optional() }).optional(),
|
|
18
|
+
options: z.object({ maxTokensToSample: numberOrExpression.optional(), temperature: numberOrExpression.optional(), topK: numberOrExpression.optional(), topP: numberOrExpression.optional(), thinking: booleanOrExpression.optional(), thinkingBudget: numberOrExpression.optional(), thinkingMode: z.union([z.literal('disabled'), z.literal('adaptive'), z.literal('manual'), expressionSchema]).optional(), effort: z.union([z.literal('low'), z.literal('medium'), z.literal('high'), z.literal('xhigh'), z.literal('max'), expressionSchema]).optional(), effort: z.union([z.literal('low'), z.literal('medium'), z.literal('high'), expressionSchema]).optional(), thinkingBudget: numberOrExpression.optional() }).optional(),
|
|
19
19
|
});
|
|
20
20
|
|
|
21
21
|
// Return combined config schema
|
|
@@ -7,7 +7,7 @@
|
|
|
7
7
|
export interface LcLmChatAnthropicV1Params {
|
|
8
8
|
/**
|
|
9
9
|
* The model which will generate the completion. <a href="https://docs.anthropic.com/claude/docs/models-overview">Learn more</a>.
|
|
10
|
-
* @builderHint Default to claude-sonnet-4-6 (latest Sonnet); use claude-opus-4-7 when the user needs the most capable model. Never use Claude Sonnet 4.5, Claude 3.x, Claude 2, or LEGACY options — those are superseded and are not valid choices.
|
|
10
|
+
* @builderHint Default to claude-sonnet-4-6 (latest Sonnet); use claude-opus-4-7 when the user needs the most capable model. Never use Claude Sonnet 4.5, Claude 3.x, Claude 2, or LEGACY options — those are superseded and are not valid choices. When extended thinking is needed on Opus 4.7+, set Thinking Mode to Adaptive and choose an Effort level. The legacy Manual thinking mode is rejected by Opus 4.7.
|
|
11
11
|
* @default claude-2
|
|
12
12
|
*/
|
|
13
13
|
model?: 'claude-3-5-sonnet-20241022' | 'claude-3-opus-20240229' | 'claude-3-5-sonnet-20240620' | 'claude-3-sonnet-20240229' | 'claude-3-5-haiku-20241022' | 'claude-3-haiku-20240307' | 'claude-2' | 'claude-2.1' | 'claude-instant-1.2' | 'claude-instant-1' | Expression<string>;
|
|
@@ -21,17 +21,17 @@ export interface LcLmChatAnthropicV1Params {
|
|
|
21
21
|
*/
|
|
22
22
|
maxTokensToSample?: number | Expression<number>;
|
|
23
23
|
/** Controls randomness: Lowering results in less random completions. As the temperature approaches zero, the model will become deterministic and repetitive.
|
|
24
|
-
* @displayOptions.hide { thinking: [true] }
|
|
24
|
+
* @displayOptions.hide { thinking: [true], thinkingMode: ["adaptive", "manual"] }
|
|
25
25
|
* @default 0.7
|
|
26
26
|
*/
|
|
27
27
|
temperature?: number | Expression<number>;
|
|
28
28
|
/** Used to remove "long tail" low probability responses. Defaults to -1, which disables it.
|
|
29
|
-
* @displayOptions.hide { thinking: [true] }
|
|
29
|
+
* @displayOptions.hide { thinking: [true], thinkingMode: ["adaptive", "manual"] }
|
|
30
30
|
* @default -1
|
|
31
31
|
*/
|
|
32
32
|
topK?: number | Expression<number>;
|
|
33
33
|
/** Controls diversity via nucleus sampling: 0.5 means half of all likelihood-weighted options are considered. We generally recommend altering this or temperature but not both.
|
|
34
|
-
* @displayOptions.hide { thinking: [true] }
|
|
34
|
+
* @displayOptions.hide { thinking: [true], thinkingMode: ["adaptive", "manual"] }
|
|
35
35
|
* @default 1
|
|
36
36
|
*/
|
|
37
37
|
topP?: number | Expression<number>;
|
|
@@ -44,6 +44,25 @@ export interface LcLmChatAnthropicV1Params {
|
|
|
44
44
|
* @default 1024
|
|
45
45
|
*/
|
|
46
46
|
thinkingBudget?: number | Expression<number>;
|
|
47
|
+
/** How extended thinking should be configured for the model
|
|
48
|
+
* @default disabled
|
|
49
|
+
*/
|
|
50
|
+
thinkingMode?: 'disabled' | 'adaptive' | 'manual' | Expression<string>;
|
|
51
|
+
/** Effort level for adaptive thinking
|
|
52
|
+
* @displayOptions.show { thinkingMode: ["adaptive"], /model.value: [{"_cnd":{"includes":"opus"}}] }
|
|
53
|
+
* @default medium
|
|
54
|
+
*/
|
|
55
|
+
effort?: 'low' | 'medium' | 'high' | 'xhigh' | 'max' | Expression<string>;
|
|
56
|
+
/** Effort level for adaptive thinking
|
|
57
|
+
* @displayOptions.show { thinkingMode: ["adaptive"], /model.value: [{"_cnd":{"regex":"^(?!.*opus).*"}}] }
|
|
58
|
+
* @default medium
|
|
59
|
+
*/
|
|
60
|
+
effort?: 'low' | 'medium' | 'high' | Expression<string>;
|
|
61
|
+
/** Maximum tokens used for thinking. Manual mode is rejected by Opus 4.7+.
|
|
62
|
+
* @displayOptions.show { thinkingMode: ["manual"] }
|
|
63
|
+
* @default 1024
|
|
64
|
+
*/
|
|
65
|
+
thinkingBudget?: number | Expression<number>;
|
|
47
66
|
};
|
|
48
67
|
}
|
|
49
68
|
|
|
@@ -15,7 +15,7 @@ module.exports = function getSchema({ parameters, z, expressionSchema, stringOrE
|
|
|
15
15
|
// Parameters schema
|
|
16
16
|
const parametersSchema = z.object({
|
|
17
17
|
model: z.union([z.literal('claude-3-5-sonnet-20241022'), z.literal('claude-3-opus-20240229'), z.literal('claude-3-5-sonnet-20240620'), z.literal('claude-3-sonnet-20240229'), z.literal('claude-3-5-haiku-20241022'), z.literal('claude-3-haiku-20240307'), z.literal('claude-2'), z.literal('claude-2.1'), z.literal('claude-instant-1.2'), z.literal('claude-instant-1'), expressionSchema]).optional(),
|
|
18
|
-
options: z.object({ maxTokensToSample: numberOrExpression.optional(), temperature: numberOrExpression.optional(), topK: numberOrExpression.optional(), topP: numberOrExpression.optional(), thinking: booleanOrExpression.optional(), thinkingBudget: numberOrExpression.optional() }).optional(),
|
|
18
|
+
options: z.object({ maxTokensToSample: numberOrExpression.optional(), temperature: numberOrExpression.optional(), topK: numberOrExpression.optional(), topP: numberOrExpression.optional(), thinking: booleanOrExpression.optional(), thinkingBudget: numberOrExpression.optional(), thinkingMode: z.union([z.literal('disabled'), z.literal('adaptive'), z.literal('manual'), expressionSchema]).optional(), effort: z.union([z.literal('low'), z.literal('medium'), z.literal('high'), z.literal('xhigh'), z.literal('max'), expressionSchema]).optional(), effort: z.union([z.literal('low'), z.literal('medium'), z.literal('high'), expressionSchema]).optional(), thinkingBudget: numberOrExpression.optional() }).optional(),
|
|
19
19
|
});
|
|
20
20
|
|
|
21
21
|
// Return combined config schema
|
|
@@ -7,8 +7,8 @@
|
|
|
7
7
|
export interface LcLmChatAnthropicV11Params {
|
|
8
8
|
/**
|
|
9
9
|
* The model which will generate the completion. <a href="https://docs.anthropic.com/claude/docs/models-overview">Learn more</a>.
|
|
10
|
-
* @builderHint Default to claude-sonnet-4-6 (latest Sonnet); use claude-opus-4-7 when the user needs the most capable model. Never use Claude Sonnet 4.5, Claude 3.x, Claude 2, or LEGACY options — those are superseded and are not valid choices.
|
|
11
|
-
* @default claude-
|
|
10
|
+
* @builderHint Default to claude-sonnet-4-6 (latest Sonnet); use claude-opus-4-7 when the user needs the most capable model. Never use Claude Sonnet 4.5, Claude 3.x, Claude 2, or LEGACY options — those are superseded and are not valid choices. When extended thinking is needed on Opus 4.7+, set Thinking Mode to Adaptive and choose an Effort level. The legacy Manual thinking mode is rejected by Opus 4.7.
|
|
11
|
+
* @default claude-3-sonnet-20240229
|
|
12
12
|
*/
|
|
13
13
|
model?: 'claude-3-5-sonnet-20241022' | 'claude-3-opus-20240229' | 'claude-3-5-sonnet-20240620' | 'claude-3-sonnet-20240229' | 'claude-3-5-haiku-20241022' | 'claude-3-haiku-20240307' | 'claude-2' | 'claude-2.1' | 'claude-instant-1.2' | 'claude-instant-1' | Expression<string>;
|
|
14
14
|
/**
|
|
@@ -21,17 +21,17 @@ export interface LcLmChatAnthropicV11Params {
|
|
|
21
21
|
*/
|
|
22
22
|
maxTokensToSample?: number | Expression<number>;
|
|
23
23
|
/** Controls randomness: Lowering results in less random completions. As the temperature approaches zero, the model will become deterministic and repetitive.
|
|
24
|
-
* @displayOptions.hide { thinking: [true] }
|
|
24
|
+
* @displayOptions.hide { thinking: [true], thinkingMode: ["adaptive", "manual"] }
|
|
25
25
|
* @default 0.7
|
|
26
26
|
*/
|
|
27
27
|
temperature?: number | Expression<number>;
|
|
28
28
|
/** Used to remove "long tail" low probability responses. Defaults to -1, which disables it.
|
|
29
|
-
* @displayOptions.hide { thinking: [true] }
|
|
29
|
+
* @displayOptions.hide { thinking: [true], thinkingMode: ["adaptive", "manual"] }
|
|
30
30
|
* @default -1
|
|
31
31
|
*/
|
|
32
32
|
topK?: number | Expression<number>;
|
|
33
33
|
/** Controls diversity via nucleus sampling: 0.5 means half of all likelihood-weighted options are considered. We generally recommend altering this or temperature but not both.
|
|
34
|
-
* @displayOptions.hide { thinking: [true] }
|
|
34
|
+
* @displayOptions.hide { thinking: [true], thinkingMode: ["adaptive", "manual"] }
|
|
35
35
|
* @default 1
|
|
36
36
|
*/
|
|
37
37
|
topP?: number | Expression<number>;
|
|
@@ -44,6 +44,25 @@ export interface LcLmChatAnthropicV11Params {
|
|
|
44
44
|
* @default 1024
|
|
45
45
|
*/
|
|
46
46
|
thinkingBudget?: number | Expression<number>;
|
|
47
|
+
/** How extended thinking should be configured for the model
|
|
48
|
+
* @default disabled
|
|
49
|
+
*/
|
|
50
|
+
thinkingMode?: 'disabled' | 'adaptive' | 'manual' | Expression<string>;
|
|
51
|
+
/** Effort level for adaptive thinking
|
|
52
|
+
* @displayOptions.show { thinkingMode: ["adaptive"], /model.value: [{"_cnd":{"includes":"opus"}}] }
|
|
53
|
+
* @default medium
|
|
54
|
+
*/
|
|
55
|
+
effort?: 'low' | 'medium' | 'high' | 'xhigh' | 'max' | Expression<string>;
|
|
56
|
+
/** Effort level for adaptive thinking
|
|
57
|
+
* @displayOptions.show { thinkingMode: ["adaptive"], /model.value: [{"_cnd":{"regex":"^(?!.*opus).*"}}] }
|
|
58
|
+
* @default medium
|
|
59
|
+
*/
|
|
60
|
+
effort?: 'low' | 'medium' | 'high' | Expression<string>;
|
|
61
|
+
/** Maximum tokens used for thinking. Manual mode is rejected by Opus 4.7+.
|
|
62
|
+
* @displayOptions.show { thinkingMode: ["manual"] }
|
|
63
|
+
* @default 1024
|
|
64
|
+
*/
|
|
65
|
+
thinkingBudget?: number | Expression<number>;
|
|
47
66
|
};
|
|
48
67
|
}
|
|
49
68
|
|
|
@@ -14,8 +14,8 @@ module.exports = function getSchema({ parameters, z, expressionSchema, stringOrE
|
|
|
14
14
|
|
|
15
15
|
// Parameters schema
|
|
16
16
|
const parametersSchema = z.object({
|
|
17
|
-
model: z.union([z.literal('claude-3-5-sonnet-20241022'), z.literal('claude-3-opus-20240229'), z.literal('claude-3-5-sonnet-20240620'), z.literal('claude-3-sonnet-20240229'), z.literal('claude-3-5-haiku-20241022'), z.literal('claude-3-haiku-20240307'),
|
|
18
|
-
options: z.object({ maxTokensToSample: numberOrExpression.optional(), temperature: numberOrExpression.optional(), topK: numberOrExpression.optional(), topP: numberOrExpression.optional(), thinking: booleanOrExpression.optional(), thinkingBudget: numberOrExpression.optional() }).optional(),
|
|
17
|
+
model: z.union([z.literal('claude-3-5-sonnet-20241022'), z.literal('claude-3-opus-20240229'), z.literal('claude-3-5-sonnet-20240620'), z.literal('claude-3-sonnet-20240229'), z.literal('claude-3-5-haiku-20241022'), z.literal('claude-3-haiku-20240307'), expressionSchema]).optional(),
|
|
18
|
+
options: z.object({ maxTokensToSample: numberOrExpression.optional(), temperature: numberOrExpression.optional(), topK: numberOrExpression.optional(), topP: numberOrExpression.optional(), thinking: booleanOrExpression.optional(), thinkingBudget: numberOrExpression.optional(), thinkingMode: z.union([z.literal('disabled'), z.literal('adaptive'), z.literal('manual'), expressionSchema]).optional(), effort: z.union([z.literal('low'), z.literal('medium'), z.literal('high'), z.literal('xhigh'), z.literal('max'), expressionSchema]).optional(), effort: z.union([z.literal('low'), z.literal('medium'), z.literal('high'), expressionSchema]).optional(), thinkingBudget: numberOrExpression.optional() }).optional(),
|
|
19
19
|
});
|
|
20
20
|
|
|
21
21
|
// Return combined config schema
|
|
@@ -7,10 +7,10 @@
|
|
|
7
7
|
export interface LcLmChatAnthropicV12Params {
|
|
8
8
|
/**
|
|
9
9
|
* The model which will generate the completion. <a href="https://docs.anthropic.com/claude/docs/models-overview">Learn more</a>.
|
|
10
|
-
* @builderHint Default to claude-sonnet-4-6 (latest Sonnet); use claude-opus-4-7 when the user needs the most capable model. Never use Claude Sonnet 4.5, Claude 3.x, Claude 2, or LEGACY options — those are superseded and are not valid choices.
|
|
11
|
-
* @default claude-
|
|
10
|
+
* @builderHint Default to claude-sonnet-4-6 (latest Sonnet); use claude-opus-4-7 when the user needs the most capable model. Never use Claude Sonnet 4.5, Claude 3.x, Claude 2, or LEGACY options — those are superseded and are not valid choices. When extended thinking is needed on Opus 4.7+, set Thinking Mode to Adaptive and choose an Effort level. The legacy Manual thinking mode is rejected by Opus 4.7.
|
|
11
|
+
* @default claude-3-5-sonnet-20240620
|
|
12
12
|
*/
|
|
13
|
-
model?: 'claude-3-5-sonnet-20241022' | 'claude-3-opus-20240229' | 'claude-3-5-sonnet-20240620' | 'claude-3-sonnet-20240229' | 'claude-3-5-haiku-20241022' | 'claude-3-haiku-20240307' |
|
|
13
|
+
model?: 'claude-3-5-sonnet-20241022' | 'claude-3-opus-20240229' | 'claude-3-5-sonnet-20240620' | 'claude-3-sonnet-20240229' | 'claude-3-5-haiku-20241022' | 'claude-3-haiku-20240307' | Expression<string>;
|
|
14
14
|
/**
|
|
15
15
|
* Additional options to add
|
|
16
16
|
* @default {}
|
|
@@ -21,17 +21,17 @@ export interface LcLmChatAnthropicV12Params {
|
|
|
21
21
|
*/
|
|
22
22
|
maxTokensToSample?: number | Expression<number>;
|
|
23
23
|
/** Controls randomness: Lowering results in less random completions. As the temperature approaches zero, the model will become deterministic and repetitive.
|
|
24
|
-
* @displayOptions.hide { thinking: [true] }
|
|
24
|
+
* @displayOptions.hide { thinking: [true], thinkingMode: ["adaptive", "manual"] }
|
|
25
25
|
* @default 0.7
|
|
26
26
|
*/
|
|
27
27
|
temperature?: number | Expression<number>;
|
|
28
28
|
/** Used to remove "long tail" low probability responses. Defaults to -1, which disables it.
|
|
29
|
-
* @displayOptions.hide { thinking: [true] }
|
|
29
|
+
* @displayOptions.hide { thinking: [true], thinkingMode: ["adaptive", "manual"] }
|
|
30
30
|
* @default -1
|
|
31
31
|
*/
|
|
32
32
|
topK?: number | Expression<number>;
|
|
33
33
|
/** Controls diversity via nucleus sampling: 0.5 means half of all likelihood-weighted options are considered. We generally recommend altering this or temperature but not both.
|
|
34
|
-
* @displayOptions.hide { thinking: [true] }
|
|
34
|
+
* @displayOptions.hide { thinking: [true], thinkingMode: ["adaptive", "manual"] }
|
|
35
35
|
* @default 1
|
|
36
36
|
*/
|
|
37
37
|
topP?: number | Expression<number>;
|
|
@@ -44,6 +44,25 @@ export interface LcLmChatAnthropicV12Params {
|
|
|
44
44
|
* @default 1024
|
|
45
45
|
*/
|
|
46
46
|
thinkingBudget?: number | Expression<number>;
|
|
47
|
+
/** How extended thinking should be configured for the model
|
|
48
|
+
* @default disabled
|
|
49
|
+
*/
|
|
50
|
+
thinkingMode?: 'disabled' | 'adaptive' | 'manual' | Expression<string>;
|
|
51
|
+
/** Effort level for adaptive thinking
|
|
52
|
+
* @displayOptions.show { thinkingMode: ["adaptive"], /model.value: [{"_cnd":{"includes":"opus"}}] }
|
|
53
|
+
* @default medium
|
|
54
|
+
*/
|
|
55
|
+
effort?: 'low' | 'medium' | 'high' | 'xhigh' | 'max' | Expression<string>;
|
|
56
|
+
/** Effort level for adaptive thinking
|
|
57
|
+
* @displayOptions.show { thinkingMode: ["adaptive"], /model.value: [{"_cnd":{"regex":"^(?!.*opus).*"}}] }
|
|
58
|
+
* @default medium
|
|
59
|
+
*/
|
|
60
|
+
effort?: 'low' | 'medium' | 'high' | Expression<string>;
|
|
61
|
+
/** Maximum tokens used for thinking. Manual mode is rejected by Opus 4.7+.
|
|
62
|
+
* @displayOptions.show { thinkingMode: ["manual"] }
|
|
63
|
+
* @default 1024
|
|
64
|
+
*/
|
|
65
|
+
thinkingBudget?: number | Expression<number>;
|
|
47
66
|
};
|
|
48
67
|
}
|
|
49
68
|
|
|
@@ -15,7 +15,7 @@ module.exports = function getSchema({ parameters, z, expressionSchema, stringOrE
|
|
|
15
15
|
// Parameters schema
|
|
16
16
|
const parametersSchema = z.object({
|
|
17
17
|
model: z.union([z.object({ __rl: z.literal(true), mode: z.union([z.literal('list'), z.literal('id')]), value: z.union([z.string(), z.number()]), cachedResultName: z.string().optional(), cachedResultUrl: z.string().optional() }), expressionSchema]).optional(),
|
|
18
|
-
options: z.object({ maxTokensToSample: numberOrExpression.optional(), temperature: numberOrExpression.optional(), topK: numberOrExpression.optional(), topP: numberOrExpression.optional(), thinking: booleanOrExpression.optional(), thinkingBudget: numberOrExpression.optional() }).optional(),
|
|
18
|
+
options: z.object({ maxTokensToSample: numberOrExpression.optional(), temperature: numberOrExpression.optional(), topK: numberOrExpression.optional(), topP: numberOrExpression.optional(), thinking: booleanOrExpression.optional(), thinkingBudget: numberOrExpression.optional(), thinkingMode: z.union([z.literal('disabled'), z.literal('adaptive'), z.literal('manual'), expressionSchema]).optional(), effort: z.union([z.literal('low'), z.literal('medium'), z.literal('high'), z.literal('xhigh'), z.literal('max'), expressionSchema]).optional(), effort: z.union([z.literal('low'), z.literal('medium'), z.literal('high'), expressionSchema]).optional(), thinkingBudget: numberOrExpression.optional() }).optional(),
|
|
19
19
|
});
|
|
20
20
|
|
|
21
21
|
// Return combined config schema
|
|
@@ -7,7 +7,7 @@
|
|
|
7
7
|
export interface LcLmChatAnthropicV13Params {
|
|
8
8
|
/**
|
|
9
9
|
* The model. Choose from the list, or specify an ID. <a href="https://docs.anthropic.com/claude/docs/models-overview">Learn more</a>.
|
|
10
|
-
* @builderHint Default to claude-sonnet-4-6 (latest Sonnet); use claude-opus-4-7 when the user needs the most capable model. Never use Claude Sonnet 4.5, Claude 3.x, Claude 2, or LEGACY options — those are superseded and are not valid choices.
|
|
10
|
+
* @builderHint Default to claude-sonnet-4-6 (latest Sonnet); use claude-opus-4-7 when the user needs the most capable model. Never use Claude Sonnet 4.5, Claude 3.x, Claude 2, or LEGACY options — those are superseded and are not valid choices. When extended thinking is needed on Opus 4.7+, set Thinking Mode to Adaptive and choose an Effort level. The legacy Manual thinking mode is rejected by Opus 4.7.
|
|
11
11
|
* @searchListMethod searchModels
|
|
12
12
|
* @default {"mode":"list","value":"claude-sonnet-4-5-20250929","cachedResultName":"Claude Sonnet 4.5"}
|
|
13
13
|
*/
|
|
@@ -22,17 +22,17 @@ export interface LcLmChatAnthropicV13Params {
|
|
|
22
22
|
*/
|
|
23
23
|
maxTokensToSample?: number | Expression<number>;
|
|
24
24
|
/** Controls randomness: Lowering results in less random completions. As the temperature approaches zero, the model will become deterministic and repetitive.
|
|
25
|
-
* @displayOptions.hide { thinking: [true] }
|
|
25
|
+
* @displayOptions.hide { thinking: [true], thinkingMode: ["adaptive", "manual"] }
|
|
26
26
|
* @default 0.7
|
|
27
27
|
*/
|
|
28
28
|
temperature?: number | Expression<number>;
|
|
29
29
|
/** Used to remove "long tail" low probability responses. Defaults to -1, which disables it.
|
|
30
|
-
* @displayOptions.hide { thinking: [true] }
|
|
30
|
+
* @displayOptions.hide { thinking: [true], thinkingMode: ["adaptive", "manual"] }
|
|
31
31
|
* @default -1
|
|
32
32
|
*/
|
|
33
33
|
topK?: number | Expression<number>;
|
|
34
34
|
/** Controls diversity via nucleus sampling: 0.5 means half of all likelihood-weighted options are considered. We generally recommend altering this or temperature but not both.
|
|
35
|
-
* @displayOptions.hide { thinking: [true] }
|
|
35
|
+
* @displayOptions.hide { thinking: [true], thinkingMode: ["adaptive", "manual"] }
|
|
36
36
|
* @default 1
|
|
37
37
|
*/
|
|
38
38
|
topP?: number | Expression<number>;
|
|
@@ -45,6 +45,25 @@ export interface LcLmChatAnthropicV13Params {
|
|
|
45
45
|
* @default 1024
|
|
46
46
|
*/
|
|
47
47
|
thinkingBudget?: number | Expression<number>;
|
|
48
|
+
/** How extended thinking should be configured for the model
|
|
49
|
+
* @default disabled
|
|
50
|
+
*/
|
|
51
|
+
thinkingMode?: 'disabled' | 'adaptive' | 'manual' | Expression<string>;
|
|
52
|
+
/** Effort level for adaptive thinking
|
|
53
|
+
* @displayOptions.show { thinkingMode: ["adaptive"], /model.value: [{"_cnd":{"includes":"opus"}}] }
|
|
54
|
+
* @default medium
|
|
55
|
+
*/
|
|
56
|
+
effort?: 'low' | 'medium' | 'high' | 'xhigh' | 'max' | Expression<string>;
|
|
57
|
+
/** Effort level for adaptive thinking
|
|
58
|
+
* @displayOptions.show { thinkingMode: ["adaptive"], /model.value: [{"_cnd":{"regex":"^(?!.*opus).*"}}] }
|
|
59
|
+
* @default medium
|
|
60
|
+
*/
|
|
61
|
+
effort?: 'low' | 'medium' | 'high' | Expression<string>;
|
|
62
|
+
/** Maximum tokens used for thinking. Manual mode is rejected by Opus 4.7+.
|
|
63
|
+
* @displayOptions.show { thinkingMode: ["manual"] }
|
|
64
|
+
* @default 1024
|
|
65
|
+
*/
|
|
66
|
+
thinkingBudget?: number | Expression<number>;
|
|
48
67
|
};
|
|
49
68
|
}
|
|
50
69
|
|
|
@@ -15,7 +15,7 @@ module.exports = function getSchema({ parameters, z, expressionSchema, stringOrE
|
|
|
15
15
|
// Parameters schema
|
|
16
16
|
const parametersSchema = z.object({
|
|
17
17
|
model: z.union([z.object({ __rl: z.literal(true), mode: z.union([z.literal('list'), z.literal('id')]), value: z.union([z.string(), z.number()]), cachedResultName: z.string().optional(), cachedResultUrl: z.string().optional() }), expressionSchema]).optional(),
|
|
18
|
-
options: z.object({ maxTokensToSample: numberOrExpression.optional(), temperature: numberOrExpression.optional(), topK: numberOrExpression.optional(), topP: numberOrExpression.optional(), thinking: booleanOrExpression.optional(), thinkingBudget: numberOrExpression.optional() }).optional(),
|
|
18
|
+
options: z.object({ maxTokensToSample: numberOrExpression.optional(), temperature: numberOrExpression.optional(), topK: numberOrExpression.optional(), topP: numberOrExpression.optional(), thinking: booleanOrExpression.optional(), thinkingBudget: numberOrExpression.optional(), thinkingMode: z.union([z.literal('disabled'), z.literal('adaptive'), z.literal('manual'), expressionSchema]).optional(), effort: z.union([z.literal('low'), z.literal('medium'), z.literal('high'), z.literal('xhigh'), z.literal('max'), expressionSchema]).optional(), effort: z.union([z.literal('low'), z.literal('medium'), z.literal('high'), expressionSchema]).optional(), thinkingBudget: numberOrExpression.optional() }).optional(),
|
|
19
19
|
});
|
|
20
20
|
|
|
21
21
|
// Return combined config schema
|
|
@@ -7,7 +7,7 @@
|
|
|
7
7
|
export interface LcLmChatAnthropicV14Params {
|
|
8
8
|
/**
|
|
9
9
|
* The model. Choose from the list, or specify an ID. <a href="https://docs.anthropic.com/claude/docs/models-overview">Learn more</a>.
|
|
10
|
-
* @builderHint Default to claude-sonnet-4-6 (latest Sonnet); use claude-opus-4-7 when the user needs the most capable model. Never use Claude Sonnet 4.5, Claude 3.x, Claude 2, or LEGACY options — those are superseded and are not valid choices.
|
|
10
|
+
* @builderHint Default to claude-sonnet-4-6 (latest Sonnet); use claude-opus-4-7 when the user needs the most capable model. Never use Claude Sonnet 4.5, Claude 3.x, Claude 2, or LEGACY options — those are superseded and are not valid choices. When extended thinking is needed on Opus 4.7+, set Thinking Mode to Adaptive and choose an Effort level. The legacy Manual thinking mode is rejected by Opus 4.7.
|
|
11
11
|
* @searchListMethod searchModels
|
|
12
12
|
* @default {"mode":"list","value":"claude-sonnet-4-6","cachedResultName":"Claude Sonnet 4.6"}
|
|
13
13
|
*/
|
|
@@ -22,17 +22,17 @@ export interface LcLmChatAnthropicV14Params {
|
|
|
22
22
|
*/
|
|
23
23
|
maxTokensToSample?: number | Expression<number>;
|
|
24
24
|
/** Controls randomness: Lowering results in less random completions. As the temperature approaches zero, the model will become deterministic and repetitive.
|
|
25
|
-
* @displayOptions.hide { thinking: [true] }
|
|
25
|
+
* @displayOptions.hide { thinking: [true], thinkingMode: ["adaptive", "manual"] }
|
|
26
26
|
* @default 0.7
|
|
27
27
|
*/
|
|
28
28
|
temperature?: number | Expression<number>;
|
|
29
29
|
/** Used to remove "long tail" low probability responses. Defaults to -1, which disables it.
|
|
30
|
-
* @displayOptions.hide { thinking: [true] }
|
|
30
|
+
* @displayOptions.hide { thinking: [true], thinkingMode: ["adaptive", "manual"] }
|
|
31
31
|
* @default -1
|
|
32
32
|
*/
|
|
33
33
|
topK?: number | Expression<number>;
|
|
34
34
|
/** Controls diversity via nucleus sampling: 0.5 means half of all likelihood-weighted options are considered. We generally recommend altering this or temperature but not both.
|
|
35
|
-
* @displayOptions.hide { thinking: [true] }
|
|
35
|
+
* @displayOptions.hide { thinking: [true], thinkingMode: ["adaptive", "manual"] }
|
|
36
36
|
* @default 1
|
|
37
37
|
*/
|
|
38
38
|
topP?: number | Expression<number>;
|
|
@@ -45,6 +45,25 @@ export interface LcLmChatAnthropicV14Params {
|
|
|
45
45
|
* @default 1024
|
|
46
46
|
*/
|
|
47
47
|
thinkingBudget?: number | Expression<number>;
|
|
48
|
+
/** How extended thinking should be configured for the model
|
|
49
|
+
* @default disabled
|
|
50
|
+
*/
|
|
51
|
+
thinkingMode?: 'disabled' | 'adaptive' | 'manual' | Expression<string>;
|
|
52
|
+
/** Effort level for adaptive thinking
|
|
53
|
+
* @displayOptions.show { thinkingMode: ["adaptive"], /model.value: [{"_cnd":{"includes":"opus"}}] }
|
|
54
|
+
* @default medium
|
|
55
|
+
*/
|
|
56
|
+
effort?: 'low' | 'medium' | 'high' | 'xhigh' | 'max' | Expression<string>;
|
|
57
|
+
/** Effort level for adaptive thinking
|
|
58
|
+
* @displayOptions.show { thinkingMode: ["adaptive"], /model.value: [{"_cnd":{"regex":"^(?!.*opus).*"}}] }
|
|
59
|
+
* @default medium
|
|
60
|
+
*/
|
|
61
|
+
effort?: 'low' | 'medium' | 'high' | Expression<string>;
|
|
62
|
+
/** Maximum tokens used for thinking. Manual mode is rejected by Opus 4.7+.
|
|
63
|
+
* @displayOptions.show { thinkingMode: ["manual"] }
|
|
64
|
+
* @default 1024
|
|
65
|
+
*/
|
|
66
|
+
thinkingBudget?: number | Expression<number>;
|
|
48
67
|
};
|
|
49
68
|
}
|
|
50
69
|
|
|
@@ -0,0 +1,25 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* Anthropic Chat Model Node - Version 1.5 - Zod Validation Schemas
|
|
3
|
+
*
|
|
4
|
+
* These schemas validate node configuration at runtime.
|
|
5
|
+
* Use .parse() for strict validation or .safeParse() for error handling.
|
|
6
|
+
*
|
|
7
|
+
* Schema helpers (z, expressionSchema, etc.) are passed as parameters
|
|
8
|
+
* by the schema-validator, not imported from external files.
|
|
9
|
+
*
|
|
10
|
+
* @generated - CommonJS JavaScript for runtime loading
|
|
11
|
+
*/
|
|
12
|
+
|
|
13
|
+
module.exports = function getSchema({ parameters, z, expressionSchema, stringOrExpression, numberOrExpression, booleanOrExpression, resourceLocatorValueSchema, resourceMapperValueSchema, filterValueSchema, assignmentCollectionValueSchema, iDataObjectSchema }) {
|
|
14
|
+
|
|
15
|
+
// Parameters schema
|
|
16
|
+
const parametersSchema = z.object({
|
|
17
|
+
model: z.union([z.object({ __rl: z.literal(true), mode: z.union([z.literal('list'), z.literal('id')]), value: z.union([z.string(), z.number()]), cachedResultName: z.string().optional(), cachedResultUrl: z.string().optional() }), expressionSchema]).optional(),
|
|
18
|
+
options: z.object({ maxTokensToSample: numberOrExpression.optional(), temperature: numberOrExpression.optional(), topK: numberOrExpression.optional(), topP: numberOrExpression.optional(), thinking: booleanOrExpression.optional(), thinkingBudget: numberOrExpression.optional(), thinkingMode: z.union([z.literal('disabled'), z.literal('adaptive'), z.literal('manual'), expressionSchema]).optional(), effort: z.union([z.literal('low'), z.literal('medium'), z.literal('high'), z.literal('xhigh'), z.literal('max'), expressionSchema]).optional(), effort: z.union([z.literal('low'), z.literal('medium'), z.literal('high'), expressionSchema]).optional(), thinkingBudget: numberOrExpression.optional() }).optional(),
|
|
19
|
+
});
|
|
20
|
+
|
|
21
|
+
// Return combined config schema
|
|
22
|
+
return z.object({
|
|
23
|
+
parameters: parametersSchema.nullable().optional(),
|
|
24
|
+
});
|
|
25
|
+
};
|
|
@@ -0,0 +1,85 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* Anthropic Chat Model Node - Version 1.5
|
|
3
|
+
* Language Model Anthropic
|
|
4
|
+
*/
|
|
5
|
+
|
|
6
|
+
|
|
7
|
+
export interface LcLmChatAnthropicV15Params {
|
|
8
|
+
/**
|
|
9
|
+
* The model. Choose from the list, or specify an ID. <a href="https://docs.anthropic.com/claude/docs/models-overview">Learn more</a>.
|
|
10
|
+
* @builderHint Default to claude-sonnet-4-6 (latest Sonnet); use claude-opus-4-7 when the user needs the most capable model. Never use Claude Sonnet 4.5, Claude 3.x, Claude 2, or LEGACY options — those are superseded and are not valid choices. When extended thinking is needed on Opus 4.7+, set Thinking Mode to Adaptive and choose an Effort level. The legacy Manual thinking mode is rejected by Opus 4.7.
|
|
11
|
+
* @searchListMethod searchModels
|
|
12
|
+
* @default {"mode":"list","value":"claude-sonnet-4-6","cachedResultName":"Claude Sonnet 4.6"}
|
|
13
|
+
*/
|
|
14
|
+
model?: { __rl: true; mode: 'list' | 'id'; value: string; cachedResultName?: string };
|
|
15
|
+
/**
|
|
16
|
+
* Additional options to add
|
|
17
|
+
* @default {}
|
|
18
|
+
*/
|
|
19
|
+
options?: {
|
|
20
|
+
/** The maximum number of tokens to generate in the completion
|
|
21
|
+
* @default 4096
|
|
22
|
+
*/
|
|
23
|
+
maxTokensToSample?: number | Expression<number>;
|
|
24
|
+
/** Controls randomness: Lowering results in less random completions. As the temperature approaches zero, the model will become deterministic and repetitive.
|
|
25
|
+
* @displayOptions.hide { thinking: [true], thinkingMode: ["adaptive", "manual"] }
|
|
26
|
+
* @default 0.7
|
|
27
|
+
*/
|
|
28
|
+
temperature?: number | Expression<number>;
|
|
29
|
+
/** Used to remove "long tail" low probability responses. Defaults to -1, which disables it.
|
|
30
|
+
* @displayOptions.hide { thinking: [true], thinkingMode: ["adaptive", "manual"] }
|
|
31
|
+
* @default -1
|
|
32
|
+
*/
|
|
33
|
+
topK?: number | Expression<number>;
|
|
34
|
+
/** Controls diversity via nucleus sampling: 0.5 means half of all likelihood-weighted options are considered. We generally recommend altering this or temperature but not both.
|
|
35
|
+
* @displayOptions.hide { thinking: [true], thinkingMode: ["adaptive", "manual"] }
|
|
36
|
+
* @default 1
|
|
37
|
+
*/
|
|
38
|
+
topP?: number | Expression<number>;
|
|
39
|
+
/** Whether to enable thinking mode for the model
|
|
40
|
+
* @default false
|
|
41
|
+
*/
|
|
42
|
+
thinking?: boolean | Expression<boolean>;
|
|
43
|
+
/** The maximum number of tokens to use for thinking
|
|
44
|
+
* @displayOptions.show { thinking: [true] }
|
|
45
|
+
* @default 1024
|
|
46
|
+
*/
|
|
47
|
+
thinkingBudget?: number | Expression<number>;
|
|
48
|
+
/** How extended thinking should be configured for the model
|
|
49
|
+
* @default disabled
|
|
50
|
+
*/
|
|
51
|
+
thinkingMode?: 'disabled' | 'adaptive' | 'manual' | Expression<string>;
|
|
52
|
+
/** Effort level for adaptive thinking
|
|
53
|
+
* @displayOptions.show { thinkingMode: ["adaptive"], /model.value: [{"_cnd":{"includes":"opus"}}] }
|
|
54
|
+
* @default medium
|
|
55
|
+
*/
|
|
56
|
+
effort?: 'low' | 'medium' | 'high' | 'xhigh' | 'max' | Expression<string>;
|
|
57
|
+
/** Effort level for adaptive thinking
|
|
58
|
+
* @displayOptions.show { thinkingMode: ["adaptive"], /model.value: [{"_cnd":{"regex":"^(?!.*opus).*"}}] }
|
|
59
|
+
* @default medium
|
|
60
|
+
*/
|
|
61
|
+
effort?: 'low' | 'medium' | 'high' | Expression<string>;
|
|
62
|
+
/** Maximum tokens used for thinking. Manual mode is rejected by Opus 4.7+.
|
|
63
|
+
* @displayOptions.show { thinkingMode: ["manual"] }
|
|
64
|
+
* @default 1024
|
|
65
|
+
*/
|
|
66
|
+
thinkingBudget?: number | Expression<number>;
|
|
67
|
+
};
|
|
68
|
+
}
|
|
69
|
+
|
|
70
|
+
export interface LcLmChatAnthropicV15Credentials {
|
|
71
|
+
anthropicApi: CredentialReference;
|
|
72
|
+
}
|
|
73
|
+
|
|
74
|
+
interface LcLmChatAnthropicV15NodeBase {
|
|
75
|
+
type: '@n8n/n8n-nodes-langchain.lmChatAnthropic';
|
|
76
|
+
version: 1.5;
|
|
77
|
+
credentials?: LcLmChatAnthropicV15Credentials;
|
|
78
|
+
isTrigger: true;
|
|
79
|
+
}
|
|
80
|
+
|
|
81
|
+
export type LcLmChatAnthropicV15ParamsNode = LcLmChatAnthropicV15NodeBase & {
|
|
82
|
+
config: NodeConfig<LcLmChatAnthropicV15Params>;
|
|
83
|
+
};
|
|
84
|
+
|
|
85
|
+
export type LcLmChatAnthropicV15Node = LcLmChatAnthropicV15ParamsNode;
|
|
@@ -10,11 +10,11 @@
|
|
|
10
10
|
* @generated - CommonJS JavaScript for runtime loading
|
|
11
11
|
*/
|
|
12
12
|
|
|
13
|
-
module.exports = function getSchema({ parameters, z, expressionSchema, stringOrExpression, numberOrExpression, booleanOrExpression, resourceLocatorValueSchema, resourceMapperValueSchema, filterValueSchema, assignmentCollectionValueSchema, iDataObjectSchema, resolveSchema }) {
|
|
13
|
+
module.exports = function getSchema({ parameters, z, expressionSchema, stringOrExpression, numberOrExpression, booleanOrExpression, resourceLocatorValueSchema, resourceMapperValueSchema, filterValueSchema, assignmentCollectionValueSchema, iDataObjectSchema, resolveSchema, resolveOneOfSchemas }) {
|
|
14
14
|
|
|
15
15
|
// Parameters schema
|
|
16
16
|
const parametersSchema = z.object({
|
|
17
|
-
model:
|
|
17
|
+
model: resolveOneOfSchemas({ parameters, variants: [{ schema: stringOrExpression, required: false, displayOptions: {"hide":{"modelSource":["inferenceProfile"]}} }, { schema: stringOrExpression, required: false, displayOptions: {"show":{"modelSource":["inferenceProfile"]}} }] }),
|
|
18
18
|
options: z.object({ maxTokensToSample: numberOrExpression.optional(), temperature: numberOrExpression.optional() }).optional(),
|
|
19
19
|
});
|
|
20
20
|
|
|
@@ -8,7 +8,6 @@ export interface LcLmChatAwsBedrockV1Params {
|
|
|
8
8
|
/**
|
|
9
9
|
* The model which will generate the completion. <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/foundation-models.html">Learn more</a>.
|
|
10
10
|
* @builderHint Default to the latest Claude Sonnet on Bedrock (anthropic.claude-sonnet-4-6 family). For Claude Sonnet 4+, switch Model Source to Inference Profiles. Avoid claude-sonnet-4-5, claude-3.x, and non-Claude legacy models unless requested.
|
|
11
|
-
* @displayOptions.hide { modelSource: ["inferenceProfile"] }
|
|
12
11
|
*/
|
|
13
12
|
model?: string | Expression<string>;
|
|
14
13
|
/**
|
|
@@ -10,12 +10,12 @@
|
|
|
10
10
|
* @generated - CommonJS JavaScript for runtime loading
|
|
11
11
|
*/
|
|
12
12
|
|
|
13
|
-
module.exports = function getSchema({ parameters, z, expressionSchema, stringOrExpression, numberOrExpression, booleanOrExpression, resourceLocatorValueSchema, resourceMapperValueSchema, filterValueSchema, assignmentCollectionValueSchema, iDataObjectSchema, resolveSchema }) {
|
|
13
|
+
module.exports = function getSchema({ parameters, z, expressionSchema, stringOrExpression, numberOrExpression, booleanOrExpression, resourceLocatorValueSchema, resourceMapperValueSchema, filterValueSchema, assignmentCollectionValueSchema, iDataObjectSchema, resolveSchema, resolveOneOfSchemas }) {
|
|
14
14
|
|
|
15
15
|
// Parameters schema
|
|
16
16
|
const parametersSchema = z.object({
|
|
17
17
|
modelSource: z.union([z.literal('onDemand'), z.literal('inferenceProfile'), expressionSchema]).optional(),
|
|
18
|
-
model:
|
|
18
|
+
model: resolveOneOfSchemas({ parameters, variants: [{ schema: stringOrExpression, required: false, displayOptions: {"hide":{"modelSource":["inferenceProfile"]}}, defaults: {"modelSource":"onDemand"} }, { schema: stringOrExpression, required: false, displayOptions: {"show":{"modelSource":["inferenceProfile"]}}, defaults: {"modelSource":"onDemand"} }] }),
|
|
19
19
|
options: z.object({ maxTokensToSample: numberOrExpression.optional(), temperature: numberOrExpression.optional() }).optional(),
|
|
20
20
|
});
|
|
21
21
|
|
|
@@ -13,7 +13,6 @@ export interface LcLmChatAwsBedrockV11Params {
|
|
|
13
13
|
/**
|
|
14
14
|
* The model which will generate the completion. <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/foundation-models.html">Learn more</a>.
|
|
15
15
|
* @builderHint Default to the latest Claude Sonnet on Bedrock (anthropic.claude-sonnet-4-6 family). For Claude Sonnet 4+, switch Model Source to Inference Profiles. Avoid claude-sonnet-4-5, claude-3.x, and non-Claude legacy models unless requested.
|
|
16
|
-
* @displayOptions.hide { modelSource: ["inferenceProfile"] }
|
|
17
16
|
*/
|
|
18
17
|
model?: string | Expression<string>;
|
|
19
18
|
/**
|
|
@@ -15,7 +15,7 @@ module.exports = function getSchema({ parameters, z, expressionSchema, stringOrE
|
|
|
15
15
|
// Parameters schema
|
|
16
16
|
const parametersSchema = z.object({
|
|
17
17
|
model: stringOrExpression.optional(),
|
|
18
|
-
options: z.object({ temperature: numberOrExpression.optional(), topK: numberOrExpression.optional(), topP: numberOrExpression.optional(), frequencyPenalty: numberOrExpression.optional(), keepAlive: stringOrExpression.optional(), lowVram: booleanOrExpression.optional(), mainGpu: numberOrExpression.optional(), numBatch: numberOrExpression.optional(), numCtx: numberOrExpression.optional(), numGpu: numberOrExpression.optional(), numPredict: numberOrExpression.optional(), numThread: numberOrExpression.optional(), penalizeNewline: booleanOrExpression.optional(), presencePenalty: numberOrExpression.optional(), repeatPenalty: numberOrExpression.optional(), useMLock: booleanOrExpression.optional(), useMMap: booleanOrExpression.optional(), vocabOnly: booleanOrExpression.optional(), format: z.union([z.literal('default'), z.literal('json'), expressionSchema]).optional() }).optional(),
|
|
18
|
+
options: z.object({ think: booleanOrExpression.optional(), temperature: numberOrExpression.optional(), topK: numberOrExpression.optional(), topP: numberOrExpression.optional(), frequencyPenalty: numberOrExpression.optional(), keepAlive: stringOrExpression.optional(), lowVram: booleanOrExpression.optional(), mainGpu: numberOrExpression.optional(), numBatch: numberOrExpression.optional(), numCtx: numberOrExpression.optional(), numGpu: numberOrExpression.optional(), numPredict: numberOrExpression.optional(), numThread: numberOrExpression.optional(), penalizeNewline: booleanOrExpression.optional(), presencePenalty: numberOrExpression.optional(), repeatPenalty: numberOrExpression.optional(), useMLock: booleanOrExpression.optional(), useMMap: booleanOrExpression.optional(), vocabOnly: booleanOrExpression.optional(), format: z.union([z.literal('default'), z.literal('json'), expressionSchema]).optional() }).optional(),
|
|
19
19
|
});
|
|
20
20
|
|
|
21
21
|
// Return combined config schema
|
|
@@ -15,6 +15,10 @@ export interface LcLmChatOllamaV1Params {
|
|
|
15
15
|
* @default {}
|
|
16
16
|
*/
|
|
17
17
|
options?: {
|
|
18
|
+
/** Whether to enable (default) thinking mode for supported models. When enabled, the model's thinking process is separated from the output. When disabled, the model outputs content directly (only for supported models).
|
|
19
|
+
* @default true
|
|
20
|
+
*/
|
|
21
|
+
think?: boolean | Expression<boolean>;
|
|
18
22
|
/** Controls the randomness of the generated text. Lower values make the output more focused and deterministic, while higher values make it more diverse and random.
|
|
19
23
|
* @default 0.7
|
|
20
24
|
*/
|
|
@@ -15,7 +15,7 @@ module.exports = function getSchema({ parameters, z, expressionSchema, stringOrE
|
|
|
15
15
|
// Parameters schema
|
|
16
16
|
const parametersSchema = z.object({
|
|
17
17
|
model: stringOrExpression.optional(),
|
|
18
|
-
options: z.object({ temperature: numberOrExpression.optional(), topK: numberOrExpression.optional(), topP: numberOrExpression.optional(), frequencyPenalty: numberOrExpression.optional(), keepAlive: stringOrExpression.optional(), lowVram: booleanOrExpression.optional(), mainGpu: numberOrExpression.optional(), numBatch: numberOrExpression.optional(), numCtx: numberOrExpression.optional(), numGpu: numberOrExpression.optional(), numPredict: numberOrExpression.optional(), numThread: numberOrExpression.optional(), penalizeNewline: booleanOrExpression.optional(), presencePenalty: numberOrExpression.optional(), repeatPenalty: numberOrExpression.optional(), useMLock: booleanOrExpression.optional(), useMMap: booleanOrExpression.optional(), vocabOnly: booleanOrExpression.optional(), format: z.union([z.literal('default'), z.literal('json'), expressionSchema]).optional() }).optional(),
|
|
18
|
+
options: z.object({ think: booleanOrExpression.optional(), temperature: numberOrExpression.optional(), topK: numberOrExpression.optional(), topP: numberOrExpression.optional(), frequencyPenalty: numberOrExpression.optional(), keepAlive: stringOrExpression.optional(), lowVram: booleanOrExpression.optional(), mainGpu: numberOrExpression.optional(), numBatch: numberOrExpression.optional(), numCtx: numberOrExpression.optional(), numGpu: numberOrExpression.optional(), numPredict: numberOrExpression.optional(), numThread: numberOrExpression.optional(), penalizeNewline: booleanOrExpression.optional(), presencePenalty: numberOrExpression.optional(), repeatPenalty: numberOrExpression.optional(), useMLock: booleanOrExpression.optional(), useMMap: booleanOrExpression.optional(), vocabOnly: booleanOrExpression.optional(), format: z.union([z.literal('default'), z.literal('json'), expressionSchema]).optional() }).optional(),
|
|
19
19
|
});
|
|
20
20
|
|
|
21
21
|
// Return combined config schema
|
|
@@ -15,6 +15,10 @@ export interface LcLmOllamaV1Params {
|
|
|
15
15
|
* @default {}
|
|
16
16
|
*/
|
|
17
17
|
options?: {
|
|
18
|
+
/** Whether to enable (default) thinking mode for supported models. When enabled, the model's thinking process is separated from the output. When disabled, the model outputs content directly (only for supported models).
|
|
19
|
+
* @default true
|
|
20
|
+
*/
|
|
21
|
+
think?: boolean | Expression<boolean>;
|
|
18
22
|
/** Controls the randomness of the generated text. Lower values make the output more focused and deterministic, while higher values make it more diverse and random.
|
|
19
23
|
* @default 0.7
|
|
20
24
|
*/
|
|
@@ -7,6 +7,7 @@
|
|
|
7
7
|
export interface LcMcpTriggerV1Params {
|
|
8
8
|
/**
|
|
9
9
|
* The way to authenticate
|
|
10
|
+
* @builderHint Default to 'none'. n8n exposes inbound trigger URLs publicly by design. Only select an authentication method when the user explicitly asks to authenticate inbound traffic.
|
|
10
11
|
* @default none
|
|
11
12
|
*/
|
|
12
13
|
authentication?: 'none' | 'bearerAuth' | 'headerAuth' | Expression<string>;
|