@mcp-abap-adt/llm-agent 2.9.0 → 2.9.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -26,7 +26,7 @@ export class AnthropicAgent extends BaseAgent {
26
26
  const requestBody = {
27
27
  model,
28
28
  messages: formattedMessages,
29
- max_tokens: config.maxTokens || 2000,
29
+ max_tokens: config.maxTokens || 4096,
30
30
  temperature: config.temperature || 0.7,
31
31
  };
32
32
  if (systemMessage) {
@@ -92,7 +92,7 @@ export class AnthropicAgent extends BaseAgent {
92
92
  const requestBody = {
93
93
  model,
94
94
  messages: formattedMessages,
95
- max_tokens: options?.maxTokens ?? config.maxTokens ?? 2000,
95
+ max_tokens: options?.maxTokens ?? config.maxTokens ?? 4096,
96
96
  temperature: options?.temperature ?? config.temperature ?? 0.7,
97
97
  stream: true,
98
98
  };
@@ -26,7 +26,7 @@ export class DeepSeekAgent extends BaseAgent {
26
26
  tools: functions.length > 0 ? functions : undefined,
27
27
  tool_choice: functions.length > 0 ? 'auto' : undefined,
28
28
  temperature: config.temperature || 0.7,
29
- max_tokens: config.maxTokens || 2000,
29
+ max_tokens: config.maxTokens || 4096,
30
30
  });
31
31
  const choice = response.data.choices?.[0];
32
32
  const message = choice?.message ?? {};
@@ -84,7 +84,7 @@ export class DeepSeekAgent extends BaseAgent {
84
84
  tools: functions.length > 0 ? functions : undefined,
85
85
  tool_choice: functions.length > 0 ? 'auto' : undefined,
86
86
  temperature: config.temperature || 0.7,
87
- max_tokens: config.maxTokens || 2000,
87
+ max_tokens: config.maxTokens || 4096,
88
88
  stream: true,
89
89
  stream_options: { include_usage: true },
90
90
  });
@@ -27,7 +27,7 @@ export class OpenAIAgent extends BaseAgent {
27
27
  tools: functions.length > 0 ? functions : undefined,
28
28
  tool_choice: functions.length > 0 ? 'auto' : undefined,
29
29
  temperature: options?.temperature ?? config.temperature ?? 0.7,
30
- max_tokens: options?.maxTokens ?? config.maxTokens ?? 2000,
30
+ max_tokens: options?.maxTokens ?? config.maxTokens ?? 4096,
31
31
  top_p: options?.topP,
32
32
  stop: options?.stop,
33
33
  });
@@ -60,7 +60,7 @@ export class OpenAIAgent extends BaseAgent {
60
60
  tools: functions.length > 0 ? functions : undefined,
61
61
  tool_choice: functions.length > 0 ? 'auto' : undefined,
62
62
  temperature: options?.temperature ?? config.temperature ?? 0.7,
63
- max_tokens: options?.maxTokens ?? config.maxTokens ?? 2000,
63
+ max_tokens: options?.maxTokens ?? config.maxTokens ?? 4096,
64
64
  top_p: options?.topP,
65
65
  stop: options?.stop,
66
66
  stream: true,
@@ -27,7 +27,7 @@ export class AnthropicProvider extends BaseLLMProvider {
27
27
  const requestBody = {
28
28
  model: this.model,
29
29
  messages: this.formatMessages(conversationMessages),
30
- max_tokens: this.config.maxTokens || 2000,
30
+ max_tokens: this.config.maxTokens || 4096,
31
31
  temperature: this.config.temperature || 0.7,
32
32
  };
33
33
  if (systemMessage) {
@@ -26,7 +26,7 @@ export class DeepSeekProvider extends BaseLLMProvider {
26
26
  tools: tools && tools.length > 0 ? tools : undefined,
27
27
  tool_choice: tools && tools.length > 0 ? 'auto' : undefined,
28
28
  temperature: this.config.temperature || 0.7,
29
- max_tokens: this.config.maxTokens || 2000,
29
+ max_tokens: this.config.maxTokens || 4096,
30
30
  });
31
31
  const choice = response.data.choices[0];
32
32
  return {
@@ -53,7 +53,7 @@ export class DeepSeekProvider extends BaseLLMProvider {
53
53
  tools: tools && tools.length > 0 ? tools : undefined,
54
54
  tool_choice: tools && tools.length > 0 ? 'auto' : undefined,
55
55
  temperature: this.config.temperature || 0.7,
56
- max_tokens: this.config.maxTokens || 2000,
56
+ max_tokens: this.config.maxTokens || 4096,
57
57
  stream: true,
58
58
  }, { responseType: 'stream' });
59
59
  const stream = response.data;
@@ -35,7 +35,7 @@ export class OpenAIProvider extends BaseLLMProvider {
35
35
  tools: tools && tools.length > 0 ? tools : undefined,
36
36
  tool_choice: tools && tools.length > 0 ? 'auto' : undefined,
37
37
  temperature: this.config.temperature || 0.7,
38
- max_tokens: this.config.maxTokens || 2000,
38
+ max_tokens: this.config.maxTokens || 4096,
39
39
  });
40
40
  const choice = response.data.choices[0];
41
41
  return {
@@ -62,7 +62,7 @@ export class OpenAIProvider extends BaseLLMProvider {
62
62
  tools: tools && tools.length > 0 ? tools : undefined,
63
63
  tool_choice: tools && tools.length > 0 ? 'auto' : undefined,
64
64
  temperature: this.config.temperature || 0.7,
65
- max_tokens: this.config.maxTokens || 2000,
65
+ max_tokens: this.config.maxTokens || 4096,
66
66
  stream: true,
67
67
  }, { responseType: 'stream' });
68
68
  const stream = response.data;
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@mcp-abap-adt/llm-agent",
3
- "version": "2.9.0",
3
+ "version": "2.9.1",
4
4
  "description": "Core components for building Smart LLM agents, plus a default MCP-orchestrated OpenAI-compatible server implementation.",
5
5
  "type": "module",
6
6
  "main": "dist/index.js",