illuma-agents 1.0.2 → 1.0.4
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/LICENSE +25 -21
- package/dist/cjs/agents/AgentContext.cjs +222 -0
- package/dist/cjs/agents/AgentContext.cjs.map +1 -0
- package/dist/cjs/common/enum.cjs +5 -4
- package/dist/cjs/common/enum.cjs.map +1 -1
- package/dist/cjs/events.cjs +7 -5
- package/dist/cjs/events.cjs.map +1 -1
- package/dist/cjs/graphs/Graph.cjs +328 -207
- package/dist/cjs/graphs/Graph.cjs.map +1 -1
- package/dist/cjs/graphs/MultiAgentGraph.cjs +507 -0
- package/dist/cjs/graphs/MultiAgentGraph.cjs.map +1 -0
- package/dist/cjs/llm/anthropic/index.cjs.map +1 -1
- package/dist/cjs/llm/google/index.cjs.map +1 -1
- package/dist/cjs/llm/ollama/index.cjs.map +1 -1
- package/dist/cjs/llm/openai/index.cjs +35 -0
- package/dist/cjs/llm/openai/index.cjs.map +1 -1
- package/dist/cjs/llm/openai/utils/index.cjs +3 -1
- package/dist/cjs/llm/openai/utils/index.cjs.map +1 -1
- package/dist/cjs/llm/openrouter/index.cjs.map +1 -1
- package/dist/cjs/llm/providers.cjs +0 -2
- package/dist/cjs/llm/providers.cjs.map +1 -1
- package/dist/cjs/llm/vertexai/index.cjs.map +1 -1
- package/dist/cjs/main.cjs +12 -1
- package/dist/cjs/main.cjs.map +1 -1
- package/dist/cjs/messages/cache.cjs +123 -0
- package/dist/cjs/messages/cache.cjs.map +1 -0
- package/dist/cjs/messages/content.cjs +53 -0
- package/dist/cjs/messages/content.cjs.map +1 -0
- package/dist/cjs/messages/format.cjs +17 -29
- package/dist/cjs/messages/format.cjs.map +1 -1
- package/dist/cjs/run.cjs +119 -74
- package/dist/cjs/run.cjs.map +1 -1
- package/dist/cjs/stream.cjs +77 -73
- package/dist/cjs/stream.cjs.map +1 -1
- package/dist/cjs/tools/Calculator.cjs +45 -0
- package/dist/cjs/tools/Calculator.cjs.map +1 -0
- package/dist/cjs/tools/CodeExecutor.cjs +22 -22
- package/dist/cjs/tools/CodeExecutor.cjs.map +1 -1
- package/dist/cjs/tools/ToolNode.cjs +5 -3
- package/dist/cjs/tools/ToolNode.cjs.map +1 -1
- package/dist/cjs/tools/handlers.cjs +20 -20
- package/dist/cjs/tools/handlers.cjs.map +1 -1
- package/dist/cjs/utils/events.cjs +31 -0
- package/dist/cjs/utils/events.cjs.map +1 -0
- package/dist/cjs/utils/handlers.cjs +70 -0
- package/dist/cjs/utils/handlers.cjs.map +1 -0
- package/dist/cjs/utils/tokens.cjs +54 -7
- package/dist/cjs/utils/tokens.cjs.map +1 -1
- package/dist/esm/agents/AgentContext.mjs +220 -0
- package/dist/esm/agents/AgentContext.mjs.map +1 -0
- package/dist/esm/common/enum.mjs +5 -4
- package/dist/esm/common/enum.mjs.map +1 -1
- package/dist/esm/events.mjs +7 -5
- package/dist/esm/events.mjs.map +1 -1
- package/dist/esm/graphs/Graph.mjs +330 -209
- package/dist/esm/graphs/Graph.mjs.map +1 -1
- package/dist/esm/graphs/MultiAgentGraph.mjs +505 -0
- package/dist/esm/graphs/MultiAgentGraph.mjs.map +1 -0
- package/dist/esm/llm/anthropic/index.mjs.map +1 -1
- package/dist/esm/llm/google/index.mjs.map +1 -1
- package/dist/esm/llm/ollama/index.mjs.map +1 -1
- package/dist/esm/llm/openai/index.mjs +35 -0
- package/dist/esm/llm/openai/index.mjs.map +1 -1
- package/dist/esm/llm/openai/utils/index.mjs +3 -1
- package/dist/esm/llm/openai/utils/index.mjs.map +1 -1
- package/dist/esm/llm/openrouter/index.mjs.map +1 -1
- package/dist/esm/llm/providers.mjs +0 -2
- package/dist/esm/llm/providers.mjs.map +1 -1
- package/dist/esm/llm/vertexai/index.mjs.map +1 -1
- package/dist/esm/main.mjs +7 -2
- package/dist/esm/main.mjs.map +1 -1
- package/dist/esm/messages/cache.mjs +120 -0
- package/dist/esm/messages/cache.mjs.map +1 -0
- package/dist/esm/messages/content.mjs +51 -0
- package/dist/esm/messages/content.mjs.map +1 -0
- package/dist/esm/messages/format.mjs +18 -29
- package/dist/esm/messages/format.mjs.map +1 -1
- package/dist/esm/run.mjs +119 -74
- package/dist/esm/run.mjs.map +1 -1
- package/dist/esm/stream.mjs +77 -73
- package/dist/esm/stream.mjs.map +1 -1
- package/dist/esm/tools/Calculator.mjs +24 -0
- package/dist/esm/tools/Calculator.mjs.map +1 -0
- package/dist/esm/tools/CodeExecutor.mjs +22 -22
- package/dist/esm/tools/CodeExecutor.mjs.map +1 -1
- package/dist/esm/tools/ToolNode.mjs +5 -3
- package/dist/esm/tools/ToolNode.mjs.map +1 -1
- package/dist/esm/tools/handlers.mjs +20 -20
- package/dist/esm/tools/handlers.mjs.map +1 -1
- package/dist/esm/utils/events.mjs +29 -0
- package/dist/esm/utils/events.mjs.map +1 -0
- package/dist/esm/utils/handlers.mjs +68 -0
- package/dist/esm/utils/handlers.mjs.map +1 -0
- package/dist/esm/utils/tokens.mjs +54 -8
- package/dist/esm/utils/tokens.mjs.map +1 -1
- package/dist/types/agents/AgentContext.d.ts +94 -0
- package/dist/types/common/enum.d.ts +7 -5
- package/dist/types/events.d.ts +3 -3
- package/dist/types/graphs/Graph.d.ts +60 -66
- package/dist/types/graphs/MultiAgentGraph.d.ts +47 -0
- package/dist/types/graphs/index.d.ts +1 -0
- package/dist/types/index.d.ts +1 -0
- package/dist/types/llm/openai/index.d.ts +10 -0
- package/dist/types/messages/cache.d.ts +20 -0
- package/dist/types/messages/content.d.ts +7 -0
- package/dist/types/messages/format.d.ts +1 -7
- package/dist/types/messages/index.d.ts +2 -0
- package/dist/types/messages/reducer.d.ts +9 -0
- package/dist/types/run.d.ts +16 -10
- package/dist/types/stream.d.ts +4 -3
- package/dist/types/tools/Calculator.d.ts +8 -0
- package/dist/types/tools/ToolNode.d.ts +1 -1
- package/dist/types/tools/handlers.d.ts +9 -7
- package/dist/types/tools/search/tool.d.ts +4 -4
- package/dist/types/types/graph.d.ts +124 -11
- package/dist/types/types/llm.d.ts +13 -9
- package/dist/types/types/messages.d.ts +4 -0
- package/dist/types/types/run.d.ts +46 -8
- package/dist/types/types/stream.d.ts +3 -2
- package/dist/types/utils/events.d.ts +6 -0
- package/dist/types/utils/handlers.d.ts +34 -0
- package/dist/types/utils/index.d.ts +1 -0
- package/dist/types/utils/tokens.d.ts +24 -0
- package/package.json +162 -145
- package/src/agents/AgentContext.ts +323 -0
- package/src/common/enum.ts +177 -176
- package/src/events.ts +197 -191
- package/src/graphs/Graph.ts +1058 -846
- package/src/graphs/MultiAgentGraph.ts +598 -0
- package/src/graphs/index.ts +2 -1
- package/src/index.ts +25 -24
- package/src/llm/anthropic/index.ts +413 -413
- package/src/llm/google/index.ts +222 -222
- package/src/llm/google/utils/zod_to_genai_parameters.ts +86 -88
- package/src/llm/ollama/index.ts +92 -92
- package/src/llm/openai/index.ts +894 -853
- package/src/llm/openai/utils/index.ts +920 -918
- package/src/llm/openrouter/index.ts +60 -60
- package/src/llm/providers.ts +55 -57
- package/src/llm/vertexai/index.ts +360 -360
- package/src/messages/cache.test.ts +461 -0
- package/src/messages/cache.ts +151 -0
- package/src/messages/content.test.ts +362 -0
- package/src/messages/content.ts +63 -0
- package/src/messages/format.ts +611 -625
- package/src/messages/formatAgentMessages.test.ts +1144 -917
- package/src/messages/index.ts +6 -4
- package/src/messages/reducer.ts +80 -0
- package/src/run.ts +447 -381
- package/src/scripts/abort.ts +157 -138
- package/src/scripts/ant_web_search.ts +158 -158
- package/src/scripts/cli.ts +172 -167
- package/src/scripts/cli2.ts +133 -125
- package/src/scripts/cli3.ts +184 -178
- package/src/scripts/cli4.ts +191 -184
- package/src/scripts/cli5.ts +191 -184
- package/src/scripts/code_exec.ts +213 -214
- package/src/scripts/code_exec_simple.ts +147 -129
- package/src/scripts/content.ts +138 -120
- package/src/scripts/handoff-test.ts +135 -0
- package/src/scripts/multi-agent-chain.ts +278 -0
- package/src/scripts/multi-agent-conditional.ts +220 -0
- package/src/scripts/multi-agent-document-review-chain.ts +197 -0
- package/src/scripts/multi-agent-hybrid-flow.ts +310 -0
- package/src/scripts/multi-agent-parallel.ts +343 -0
- package/src/scripts/multi-agent-sequence.ts +212 -0
- package/src/scripts/multi-agent-supervisor.ts +364 -0
- package/src/scripts/multi-agent-test.ts +186 -0
- package/src/scripts/search.ts +146 -150
- package/src/scripts/simple.ts +225 -225
- package/src/scripts/stream.ts +140 -122
- package/src/scripts/test-custom-prompt-key.ts +145 -0
- package/src/scripts/test-handoff-input.ts +170 -0
- package/src/scripts/test-multi-agent-list-handoff.ts +261 -0
- package/src/scripts/test-tools-before-handoff.ts +222 -0
- package/src/scripts/tools.ts +153 -155
- package/src/specs/agent-handoffs.test.ts +889 -0
- package/src/specs/anthropic.simple.test.ts +320 -317
- package/src/specs/azure.simple.test.ts +325 -316
- package/src/specs/openai.simple.test.ts +311 -316
- package/src/specs/openrouter.simple.test.ts +107 -0
- package/src/specs/prune.test.ts +758 -763
- package/src/specs/reasoning.test.ts +201 -165
- package/src/specs/thinking-prune.test.ts +769 -703
- package/src/specs/token-memoization.test.ts +39 -0
- package/src/stream.ts +664 -651
- package/src/tools/Calculator.test.ts +278 -0
- package/src/tools/Calculator.ts +25 -0
- package/src/tools/CodeExecutor.ts +220 -220
- package/src/tools/ToolNode.ts +170 -170
- package/src/tools/handlers.ts +341 -336
- package/src/types/graph.ts +372 -185
- package/src/types/llm.ts +141 -140
- package/src/types/messages.ts +4 -0
- package/src/types/run.ts +128 -89
- package/src/types/stream.ts +401 -400
- package/src/utils/events.ts +32 -0
- package/src/utils/handlers.ts +107 -0
- package/src/utils/index.ts +6 -5
- package/src/utils/llmConfig.ts +183 -183
- package/src/utils/tokens.ts +129 -70
- package/dist/types/scripts/abort.d.ts +0 -1
- package/dist/types/scripts/ant_web_search.d.ts +0 -1
- package/dist/types/scripts/args.d.ts +0 -7
- package/dist/types/scripts/caching.d.ts +0 -1
- package/dist/types/scripts/cli.d.ts +0 -1
- package/dist/types/scripts/cli2.d.ts +0 -1
- package/dist/types/scripts/cli3.d.ts +0 -1
- package/dist/types/scripts/cli4.d.ts +0 -1
- package/dist/types/scripts/cli5.d.ts +0 -1
- package/dist/types/scripts/code_exec.d.ts +0 -1
- package/dist/types/scripts/code_exec_files.d.ts +0 -1
- package/dist/types/scripts/code_exec_simple.d.ts +0 -1
- package/dist/types/scripts/content.d.ts +0 -1
- package/dist/types/scripts/empty_input.d.ts +0 -1
- package/dist/types/scripts/image.d.ts +0 -1
- package/dist/types/scripts/memory.d.ts +0 -1
- package/dist/types/scripts/search.d.ts +0 -1
- package/dist/types/scripts/simple.d.ts +0 -1
- package/dist/types/scripts/stream.d.ts +0 -1
- package/dist/types/scripts/thinking.d.ts +0 -1
- package/dist/types/scripts/tools.d.ts +0 -1
- package/dist/types/specs/spec.utils.d.ts +0 -1
- package/dist/types/tools/example.d.ts +0 -78
- package/src/tools/example.ts +0 -129
package/src/llm/ollama/index.ts
CHANGED
|
@@ -1,92 +1,92 @@
|
|
|
1
|
-
import { AIMessageChunk } from '@langchain/core/messages';
|
|
2
|
-
import { ChatGenerationChunk } from '@langchain/core/outputs';
|
|
3
|
-
import { ChatOllama as BaseChatOllama } from '@langchain/ollama';
|
|
4
|
-
import { CallbackManagerForLLMRun } from '@langchain/core/callbacks/manager';
|
|
5
|
-
import type {
|
|
6
|
-
ChatResponse as OllamaChatResponse,
|
|
7
|
-
Message as OllamaMessage,
|
|
8
|
-
} from 'ollama';
|
|
9
|
-
import type { UsageMetadata, BaseMessage } from '@langchain/core/messages';
|
|
10
|
-
import {
|
|
11
|
-
convertOllamaMessagesToLangChain,
|
|
12
|
-
convertToOllamaMessages,
|
|
13
|
-
} from './utils';
|
|
14
|
-
|
|
15
|
-
export class ChatOllama extends BaseChatOllama {
|
|
16
|
-
static lc_name(): 'IllumaOllama' {
|
|
17
|
-
return 'IllumaOllama';
|
|
18
|
-
}
|
|
19
|
-
async *_streamResponseChunks(
|
|
20
|
-
messages: BaseMessage[],
|
|
21
|
-
options: this['ParsedCallOptions'],
|
|
22
|
-
runManager?: CallbackManagerForLLMRun
|
|
23
|
-
): AsyncGenerator<ChatGenerationChunk> {
|
|
24
|
-
if (this.checkOrPullModel) {
|
|
25
|
-
// eslint-disable-next-line @typescript-eslint/ban-ts-comment
|
|
26
|
-
// @ts-ignore
|
|
27
|
-
if (!((await this.checkModelExistsOnMachine(this.model)) as boolean)) {
|
|
28
|
-
await this.pull(this.model, {
|
|
29
|
-
logProgress: true,
|
|
30
|
-
});
|
|
31
|
-
}
|
|
32
|
-
}
|
|
33
|
-
|
|
34
|
-
const params = this.invocationParams(options);
|
|
35
|
-
// TODO: remove cast after SDK adds support for tool calls
|
|
36
|
-
const ollamaMessages = convertToOllamaMessages(messages) as OllamaMessage[];
|
|
37
|
-
|
|
38
|
-
const usageMetadata: UsageMetadata = {
|
|
39
|
-
input_tokens: 0,
|
|
40
|
-
output_tokens: 0,
|
|
41
|
-
total_tokens: 0,
|
|
42
|
-
};
|
|
43
|
-
|
|
44
|
-
const stream = await this.client.chat({
|
|
45
|
-
...params,
|
|
46
|
-
messages: ollamaMessages,
|
|
47
|
-
stream: true,
|
|
48
|
-
});
|
|
49
|
-
|
|
50
|
-
let lastMetadata: Omit<OllamaChatResponse, 'message'> | undefined;
|
|
51
|
-
|
|
52
|
-
for await (const chunk of stream) {
|
|
53
|
-
if (options.signal?.aborted === true) {
|
|
54
|
-
this.client.abort();
|
|
55
|
-
}
|
|
56
|
-
const { message: responseMessage, ...rest } =
|
|
57
|
-
chunk as Partial<OllamaChatResponse>;
|
|
58
|
-
usageMetadata.input_tokens += rest.prompt_eval_count ?? 0;
|
|
59
|
-
usageMetadata.output_tokens += rest.eval_count ?? 0;
|
|
60
|
-
usageMetadata.total_tokens =
|
|
61
|
-
usageMetadata.input_tokens + usageMetadata.output_tokens;
|
|
62
|
-
lastMetadata = rest as Omit<OllamaChatResponse, 'message'>;
|
|
63
|
-
if (!responseMessage) {
|
|
64
|
-
continue;
|
|
65
|
-
}
|
|
66
|
-
const message = convertOllamaMessagesToLangChain(responseMessage);
|
|
67
|
-
const generationChunk = new ChatGenerationChunk({
|
|
68
|
-
text: responseMessage.content || '',
|
|
69
|
-
message,
|
|
70
|
-
});
|
|
71
|
-
yield generationChunk;
|
|
72
|
-
await runManager?.handleLLMNewToken(
|
|
73
|
-
responseMessage.content || '',
|
|
74
|
-
undefined,
|
|
75
|
-
undefined,
|
|
76
|
-
undefined,
|
|
77
|
-
undefined,
|
|
78
|
-
{ chunk: generationChunk }
|
|
79
|
-
);
|
|
80
|
-
}
|
|
81
|
-
|
|
82
|
-
// Yield the `response_metadata` as the final chunk.
|
|
83
|
-
yield new ChatGenerationChunk({
|
|
84
|
-
text: '',
|
|
85
|
-
message: new AIMessageChunk({
|
|
86
|
-
content: '',
|
|
87
|
-
response_metadata: lastMetadata,
|
|
88
|
-
usage_metadata: usageMetadata,
|
|
89
|
-
}),
|
|
90
|
-
});
|
|
91
|
-
}
|
|
92
|
-
}
|
|
1
|
+
import { AIMessageChunk } from '@langchain/core/messages';
|
|
2
|
+
import { ChatGenerationChunk } from '@langchain/core/outputs';
|
|
3
|
+
import { ChatOllama as BaseChatOllama } from '@langchain/ollama';
|
|
4
|
+
import { CallbackManagerForLLMRun } from '@langchain/core/callbacks/manager';
|
|
5
|
+
import type {
|
|
6
|
+
ChatResponse as OllamaChatResponse,
|
|
7
|
+
Message as OllamaMessage,
|
|
8
|
+
} from 'ollama';
|
|
9
|
+
import type { UsageMetadata, BaseMessage } from '@langchain/core/messages';
|
|
10
|
+
import {
|
|
11
|
+
convertOllamaMessagesToLangChain,
|
|
12
|
+
convertToOllamaMessages,
|
|
13
|
+
} from './utils';
|
|
14
|
+
|
|
15
|
+
export class ChatOllama extends BaseChatOllama {
|
|
16
|
+
static lc_name(): 'IllumaOllama' {
|
|
17
|
+
return 'IllumaOllama';
|
|
18
|
+
}
|
|
19
|
+
async *_streamResponseChunks(
|
|
20
|
+
messages: BaseMessage[],
|
|
21
|
+
options: this['ParsedCallOptions'],
|
|
22
|
+
runManager?: CallbackManagerForLLMRun
|
|
23
|
+
): AsyncGenerator<ChatGenerationChunk> {
|
|
24
|
+
if (this.checkOrPullModel) {
|
|
25
|
+
// eslint-disable-next-line @typescript-eslint/ban-ts-comment
|
|
26
|
+
// @ts-ignore
|
|
27
|
+
if (!((await this.checkModelExistsOnMachine(this.model)) as boolean)) {
|
|
28
|
+
await this.pull(this.model, {
|
|
29
|
+
logProgress: true,
|
|
30
|
+
});
|
|
31
|
+
}
|
|
32
|
+
}
|
|
33
|
+
|
|
34
|
+
const params = this.invocationParams(options);
|
|
35
|
+
// TODO: remove cast after SDK adds support for tool calls
|
|
36
|
+
const ollamaMessages = convertToOllamaMessages(messages) as OllamaMessage[];
|
|
37
|
+
|
|
38
|
+
const usageMetadata: UsageMetadata = {
|
|
39
|
+
input_tokens: 0,
|
|
40
|
+
output_tokens: 0,
|
|
41
|
+
total_tokens: 0,
|
|
42
|
+
};
|
|
43
|
+
|
|
44
|
+
const stream = await this.client.chat({
|
|
45
|
+
...params,
|
|
46
|
+
messages: ollamaMessages,
|
|
47
|
+
stream: true,
|
|
48
|
+
});
|
|
49
|
+
|
|
50
|
+
let lastMetadata: Omit<OllamaChatResponse, 'message'> | undefined;
|
|
51
|
+
|
|
52
|
+
for await (const chunk of stream) {
|
|
53
|
+
if (options.signal?.aborted === true) {
|
|
54
|
+
this.client.abort();
|
|
55
|
+
}
|
|
56
|
+
const { message: responseMessage, ...rest } =
|
|
57
|
+
chunk as Partial<OllamaChatResponse>;
|
|
58
|
+
usageMetadata.input_tokens += rest.prompt_eval_count ?? 0;
|
|
59
|
+
usageMetadata.output_tokens += rest.eval_count ?? 0;
|
|
60
|
+
usageMetadata.total_tokens =
|
|
61
|
+
usageMetadata.input_tokens + usageMetadata.output_tokens;
|
|
62
|
+
lastMetadata = rest as Omit<OllamaChatResponse, 'message'>;
|
|
63
|
+
if (!responseMessage) {
|
|
64
|
+
continue;
|
|
65
|
+
}
|
|
66
|
+
const message = convertOllamaMessagesToLangChain(responseMessage);
|
|
67
|
+
const generationChunk = new ChatGenerationChunk({
|
|
68
|
+
text: responseMessage.content || '',
|
|
69
|
+
message,
|
|
70
|
+
});
|
|
71
|
+
yield generationChunk;
|
|
72
|
+
await runManager?.handleLLMNewToken(
|
|
73
|
+
responseMessage.content || '',
|
|
74
|
+
undefined,
|
|
75
|
+
undefined,
|
|
76
|
+
undefined,
|
|
77
|
+
undefined,
|
|
78
|
+
{ chunk: generationChunk }
|
|
79
|
+
);
|
|
80
|
+
}
|
|
81
|
+
|
|
82
|
+
// Yield the `response_metadata` as the final chunk.
|
|
83
|
+
yield new ChatGenerationChunk({
|
|
84
|
+
text: '',
|
|
85
|
+
message: new AIMessageChunk({
|
|
86
|
+
content: '',
|
|
87
|
+
response_metadata: lastMetadata,
|
|
88
|
+
usage_metadata: usageMetadata,
|
|
89
|
+
}),
|
|
90
|
+
});
|
|
91
|
+
}
|
|
92
|
+
}
|