illuma-agents 1.0.2 → 1.0.3

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (225) hide show
  1. package/LICENSE +25 -21
  2. package/dist/cjs/agents/AgentContext.cjs +222 -0
  3. package/dist/cjs/agents/AgentContext.cjs.map +1 -0
  4. package/dist/cjs/common/enum.cjs +7 -6
  5. package/dist/cjs/common/enum.cjs.map +1 -1
  6. package/dist/cjs/events.cjs +7 -5
  7. package/dist/cjs/events.cjs.map +1 -1
  8. package/dist/cjs/graphs/Graph.cjs +328 -207
  9. package/dist/cjs/graphs/Graph.cjs.map +1 -1
  10. package/dist/cjs/graphs/MultiAgentGraph.cjs +507 -0
  11. package/dist/cjs/graphs/MultiAgentGraph.cjs.map +1 -0
  12. package/dist/cjs/llm/anthropic/index.cjs.map +1 -1
  13. package/dist/cjs/llm/google/index.cjs.map +1 -1
  14. package/dist/cjs/llm/ollama/index.cjs.map +1 -1
  15. package/dist/cjs/llm/openai/index.cjs +35 -0
  16. package/dist/cjs/llm/openai/index.cjs.map +1 -1
  17. package/dist/cjs/llm/openai/utils/index.cjs +3 -1
  18. package/dist/cjs/llm/openai/utils/index.cjs.map +1 -1
  19. package/dist/cjs/llm/openrouter/index.cjs.map +1 -1
  20. package/dist/cjs/llm/providers.cjs +0 -2
  21. package/dist/cjs/llm/providers.cjs.map +1 -1
  22. package/dist/cjs/llm/vertexai/index.cjs.map +1 -1
  23. package/dist/cjs/main.cjs +12 -1
  24. package/dist/cjs/main.cjs.map +1 -1
  25. package/dist/cjs/messages/cache.cjs +123 -0
  26. package/dist/cjs/messages/cache.cjs.map +1 -0
  27. package/dist/cjs/messages/content.cjs +53 -0
  28. package/dist/cjs/messages/content.cjs.map +1 -0
  29. package/dist/cjs/messages/format.cjs +17 -29
  30. package/dist/cjs/messages/format.cjs.map +1 -1
  31. package/dist/cjs/run.cjs +119 -74
  32. package/dist/cjs/run.cjs.map +1 -1
  33. package/dist/cjs/stream.cjs +77 -73
  34. package/dist/cjs/stream.cjs.map +1 -1
  35. package/dist/cjs/tools/Calculator.cjs +45 -0
  36. package/dist/cjs/tools/Calculator.cjs.map +1 -0
  37. package/dist/cjs/tools/CodeExecutor.cjs +22 -22
  38. package/dist/cjs/tools/CodeExecutor.cjs.map +1 -1
  39. package/dist/cjs/tools/ToolNode.cjs +5 -3
  40. package/dist/cjs/tools/ToolNode.cjs.map +1 -1
  41. package/dist/cjs/tools/handlers.cjs +20 -20
  42. package/dist/cjs/tools/handlers.cjs.map +1 -1
  43. package/dist/cjs/utils/events.cjs +31 -0
  44. package/dist/cjs/utils/events.cjs.map +1 -0
  45. package/dist/cjs/utils/handlers.cjs +70 -0
  46. package/dist/cjs/utils/handlers.cjs.map +1 -0
  47. package/dist/cjs/utils/tokens.cjs +54 -7
  48. package/dist/cjs/utils/tokens.cjs.map +1 -1
  49. package/dist/esm/agents/AgentContext.mjs +220 -0
  50. package/dist/esm/agents/AgentContext.mjs.map +1 -0
  51. package/dist/esm/common/enum.mjs +7 -6
  52. package/dist/esm/common/enum.mjs.map +1 -1
  53. package/dist/esm/events.mjs +7 -5
  54. package/dist/esm/events.mjs.map +1 -1
  55. package/dist/esm/graphs/Graph.mjs +330 -209
  56. package/dist/esm/graphs/Graph.mjs.map +1 -1
  57. package/dist/esm/graphs/MultiAgentGraph.mjs +505 -0
  58. package/dist/esm/graphs/MultiAgentGraph.mjs.map +1 -0
  59. package/dist/esm/llm/anthropic/index.mjs.map +1 -1
  60. package/dist/esm/llm/google/index.mjs.map +1 -1
  61. package/dist/esm/llm/ollama/index.mjs.map +1 -1
  62. package/dist/esm/llm/openai/index.mjs +35 -0
  63. package/dist/esm/llm/openai/index.mjs.map +1 -1
  64. package/dist/esm/llm/openai/utils/index.mjs +3 -1
  65. package/dist/esm/llm/openai/utils/index.mjs.map +1 -1
  66. package/dist/esm/llm/openrouter/index.mjs.map +1 -1
  67. package/dist/esm/llm/providers.mjs +0 -2
  68. package/dist/esm/llm/providers.mjs.map +1 -1
  69. package/dist/esm/llm/vertexai/index.mjs.map +1 -1
  70. package/dist/esm/main.mjs +7 -2
  71. package/dist/esm/main.mjs.map +1 -1
  72. package/dist/esm/messages/cache.mjs +120 -0
  73. package/dist/esm/messages/cache.mjs.map +1 -0
  74. package/dist/esm/messages/content.mjs +51 -0
  75. package/dist/esm/messages/content.mjs.map +1 -0
  76. package/dist/esm/messages/format.mjs +18 -29
  77. package/dist/esm/messages/format.mjs.map +1 -1
  78. package/dist/esm/run.mjs +119 -74
  79. package/dist/esm/run.mjs.map +1 -1
  80. package/dist/esm/stream.mjs +77 -73
  81. package/dist/esm/stream.mjs.map +1 -1
  82. package/dist/esm/tools/Calculator.mjs +24 -0
  83. package/dist/esm/tools/Calculator.mjs.map +1 -0
  84. package/dist/esm/tools/CodeExecutor.mjs +22 -22
  85. package/dist/esm/tools/CodeExecutor.mjs.map +1 -1
  86. package/dist/esm/tools/ToolNode.mjs +5 -3
  87. package/dist/esm/tools/ToolNode.mjs.map +1 -1
  88. package/dist/esm/tools/handlers.mjs +20 -20
  89. package/dist/esm/tools/handlers.mjs.map +1 -1
  90. package/dist/esm/utils/events.mjs +29 -0
  91. package/dist/esm/utils/events.mjs.map +1 -0
  92. package/dist/esm/utils/handlers.mjs +68 -0
  93. package/dist/esm/utils/handlers.mjs.map +1 -0
  94. package/dist/esm/utils/tokens.mjs +54 -8
  95. package/dist/esm/utils/tokens.mjs.map +1 -1
  96. package/dist/types/agents/AgentContext.d.ts +94 -0
  97. package/dist/types/common/enum.d.ts +9 -7
  98. package/dist/types/events.d.ts +3 -3
  99. package/dist/types/graphs/Graph.d.ts +60 -66
  100. package/dist/types/graphs/MultiAgentGraph.d.ts +47 -0
  101. package/dist/types/graphs/index.d.ts +1 -0
  102. package/dist/types/index.d.ts +1 -0
  103. package/dist/types/llm/openai/index.d.ts +10 -0
  104. package/dist/types/messages/cache.d.ts +20 -0
  105. package/dist/types/messages/content.d.ts +7 -0
  106. package/dist/types/messages/format.d.ts +1 -7
  107. package/dist/types/messages/index.d.ts +2 -0
  108. package/dist/types/messages/reducer.d.ts +9 -0
  109. package/dist/types/run.d.ts +16 -10
  110. package/dist/types/stream.d.ts +4 -3
  111. package/dist/types/tools/Calculator.d.ts +8 -0
  112. package/dist/types/tools/ToolNode.d.ts +1 -1
  113. package/dist/types/tools/handlers.d.ts +9 -7
  114. package/dist/types/tools/search/tool.d.ts +4 -4
  115. package/dist/types/types/graph.d.ts +124 -11
  116. package/dist/types/types/llm.d.ts +13 -9
  117. package/dist/types/types/messages.d.ts +4 -0
  118. package/dist/types/types/run.d.ts +46 -8
  119. package/dist/types/types/stream.d.ts +3 -2
  120. package/dist/types/utils/events.d.ts +6 -0
  121. package/dist/types/utils/handlers.d.ts +34 -0
  122. package/dist/types/utils/index.d.ts +1 -0
  123. package/dist/types/utils/tokens.d.ts +24 -0
  124. package/package.json +162 -145
  125. package/src/agents/AgentContext.ts +323 -0
  126. package/src/common/enum.ts +177 -176
  127. package/src/events.ts +197 -191
  128. package/src/graphs/Graph.ts +1058 -846
  129. package/src/graphs/MultiAgentGraph.ts +598 -0
  130. package/src/graphs/index.ts +2 -1
  131. package/src/index.ts +25 -24
  132. package/src/llm/anthropic/index.ts +413 -413
  133. package/src/llm/google/index.ts +222 -222
  134. package/src/llm/google/utils/zod_to_genai_parameters.ts +86 -88
  135. package/src/llm/ollama/index.ts +92 -92
  136. package/src/llm/openai/index.ts +894 -853
  137. package/src/llm/openai/utils/index.ts +920 -918
  138. package/src/llm/openrouter/index.ts +60 -60
  139. package/src/llm/providers.ts +55 -57
  140. package/src/llm/vertexai/index.ts +360 -360
  141. package/src/messages/cache.test.ts +461 -0
  142. package/src/messages/cache.ts +151 -0
  143. package/src/messages/content.test.ts +362 -0
  144. package/src/messages/content.ts +63 -0
  145. package/src/messages/format.ts +611 -625
  146. package/src/messages/formatAgentMessages.test.ts +1144 -917
  147. package/src/messages/index.ts +6 -4
  148. package/src/messages/reducer.ts +80 -0
  149. package/src/run.ts +447 -381
  150. package/src/scripts/abort.ts +157 -138
  151. package/src/scripts/ant_web_search.ts +158 -158
  152. package/src/scripts/cli.ts +172 -167
  153. package/src/scripts/cli2.ts +133 -125
  154. package/src/scripts/cli3.ts +184 -178
  155. package/src/scripts/cli4.ts +191 -184
  156. package/src/scripts/cli5.ts +191 -184
  157. package/src/scripts/code_exec.ts +213 -214
  158. package/src/scripts/code_exec_simple.ts +147 -129
  159. package/src/scripts/content.ts +138 -120
  160. package/src/scripts/handoff-test.ts +135 -0
  161. package/src/scripts/multi-agent-chain.ts +278 -0
  162. package/src/scripts/multi-agent-conditional.ts +220 -0
  163. package/src/scripts/multi-agent-document-review-chain.ts +197 -0
  164. package/src/scripts/multi-agent-hybrid-flow.ts +310 -0
  165. package/src/scripts/multi-agent-parallel.ts +343 -0
  166. package/src/scripts/multi-agent-sequence.ts +212 -0
  167. package/src/scripts/multi-agent-supervisor.ts +364 -0
  168. package/src/scripts/multi-agent-test.ts +186 -0
  169. package/src/scripts/search.ts +146 -150
  170. package/src/scripts/simple.ts +225 -225
  171. package/src/scripts/stream.ts +140 -122
  172. package/src/scripts/test-custom-prompt-key.ts +145 -0
  173. package/src/scripts/test-handoff-input.ts +170 -0
  174. package/src/scripts/test-multi-agent-list-handoff.ts +261 -0
  175. package/src/scripts/test-tools-before-handoff.ts +222 -0
  176. package/src/scripts/tools.ts +153 -155
  177. package/src/specs/agent-handoffs.test.ts +889 -0
  178. package/src/specs/anthropic.simple.test.ts +320 -317
  179. package/src/specs/azure.simple.test.ts +325 -316
  180. package/src/specs/openai.simple.test.ts +311 -316
  181. package/src/specs/openrouter.simple.test.ts +107 -0
  182. package/src/specs/prune.test.ts +758 -763
  183. package/src/specs/reasoning.test.ts +201 -165
  184. package/src/specs/thinking-prune.test.ts +769 -703
  185. package/src/specs/token-memoization.test.ts +39 -0
  186. package/src/stream.ts +664 -651
  187. package/src/tools/Calculator.test.ts +278 -0
  188. package/src/tools/Calculator.ts +25 -0
  189. package/src/tools/CodeExecutor.ts +220 -220
  190. package/src/tools/ToolNode.ts +170 -170
  191. package/src/tools/handlers.ts +341 -336
  192. package/src/types/graph.ts +372 -185
  193. package/src/types/llm.ts +141 -140
  194. package/src/types/messages.ts +4 -0
  195. package/src/types/run.ts +128 -89
  196. package/src/types/stream.ts +401 -400
  197. package/src/utils/events.ts +32 -0
  198. package/src/utils/handlers.ts +107 -0
  199. package/src/utils/index.ts +6 -5
  200. package/src/utils/llmConfig.ts +183 -183
  201. package/src/utils/tokens.ts +129 -70
  202. package/dist/types/scripts/abort.d.ts +0 -1
  203. package/dist/types/scripts/ant_web_search.d.ts +0 -1
  204. package/dist/types/scripts/args.d.ts +0 -7
  205. package/dist/types/scripts/caching.d.ts +0 -1
  206. package/dist/types/scripts/cli.d.ts +0 -1
  207. package/dist/types/scripts/cli2.d.ts +0 -1
  208. package/dist/types/scripts/cli3.d.ts +0 -1
  209. package/dist/types/scripts/cli4.d.ts +0 -1
  210. package/dist/types/scripts/cli5.d.ts +0 -1
  211. package/dist/types/scripts/code_exec.d.ts +0 -1
  212. package/dist/types/scripts/code_exec_files.d.ts +0 -1
  213. package/dist/types/scripts/code_exec_simple.d.ts +0 -1
  214. package/dist/types/scripts/content.d.ts +0 -1
  215. package/dist/types/scripts/empty_input.d.ts +0 -1
  216. package/dist/types/scripts/image.d.ts +0 -1
  217. package/dist/types/scripts/memory.d.ts +0 -1
  218. package/dist/types/scripts/search.d.ts +0 -1
  219. package/dist/types/scripts/simple.d.ts +0 -1
  220. package/dist/types/scripts/stream.d.ts +0 -1
  221. package/dist/types/scripts/thinking.d.ts +0 -1
  222. package/dist/types/scripts/tools.d.ts +0 -1
  223. package/dist/types/specs/spec.utils.d.ts +0 -1
  224. package/dist/types/tools/example.d.ts +0 -78
  225. package/src/tools/example.ts +0 -129
@@ -1,129 +1,147 @@
1
- // src/scripts/cli.ts
2
- import { config } from 'dotenv';
3
- config();
4
- import { HumanMessage, BaseMessage } from '@langchain/core/messages';
5
- import { TavilySearchResults } from '@langchain/community/tools/tavily_search';
6
- import type * as t from '@/types';
7
- import { ChatModelStreamHandler, createContentAggregator } from '@/stream';
8
- import { createCodeExecutionTool } from '@/tools/CodeExecutor';
9
- import { ToolEndHandler, ModelEndHandler } from '@/events';
10
- import { createTokenCounter } from '@/utils/tokens';
11
- import { getLLMConfig } from '@/utils/llmConfig';
12
- import { getArgs } from '@/scripts/args';
13
- import { GraphEvents } from '@/common';
14
- import { Run } from '@/run';
15
-
16
- const conversationHistory: BaseMessage[] = [];
17
-
18
- async function testCodeExecution(): Promise<void> {
19
- const { userName, location, provider, currentDate } = await getArgs();
20
- const { contentParts, aggregateContent } = createContentAggregator();
21
- const customHandlers = {
22
- [GraphEvents.TOOL_END]: new ToolEndHandler(),
23
- [GraphEvents.CHAT_MODEL_END]: new ModelEndHandler(),
24
- [GraphEvents.CHAT_MODEL_STREAM]: new ChatModelStreamHandler(),
25
- [GraphEvents.ON_RUN_STEP_COMPLETED]: {
26
- handle: (event: GraphEvents.ON_RUN_STEP_COMPLETED, data: t.StreamEventData): void => {
27
- console.log('====== ON_RUN_STEP_COMPLETED ======');
28
- console.dir(data, { depth: null });
29
- aggregateContent({ event, data: data as unknown as { result: t.ToolEndEvent } });
30
- }
31
- },
32
- [GraphEvents.ON_RUN_STEP]: {
33
- handle: (event: GraphEvents.ON_RUN_STEP, data: t.StreamEventData): void => {
34
- console.log('====== ON_RUN_STEP ======');
35
- console.dir(data, { depth: null });
36
- aggregateContent({ event, data: data as t.RunStep });
37
- }
38
- },
39
- [GraphEvents.ON_RUN_STEP_DELTA]: {
40
- handle: (event: GraphEvents.ON_RUN_STEP_DELTA, data: t.StreamEventData): void => {
41
- console.log('====== ON_RUN_STEP_DELTA ======');
42
- console.dir(data, { depth: null });
43
- aggregateContent({ event, data: data as t.RunStepDeltaEvent });
44
- }
45
- },
46
- [GraphEvents.ON_MESSAGE_DELTA]: {
47
- handle: (event: GraphEvents.ON_MESSAGE_DELTA, data: t.StreamEventData): void => {
48
- console.log('====== ON_MESSAGE_DELTA ======');
49
- console.dir(data, { depth: null });
50
- aggregateContent({ event, data: data as t.MessageDeltaEvent });
51
- }
52
- },
53
- [GraphEvents.TOOL_START]: {
54
- handle: (_event: string, data: t.StreamEventData, metadata?: Record<string, unknown>): void => {
55
- console.log('====== TOOL_START ======');
56
- console.dir(data, { depth: null });
57
- }
58
- },
59
- };
60
-
61
- const llmConfig = getLLMConfig(provider);
62
- const instructions = 'You are a friendly AI assistant with coding capabilities. Always address the user by their name.';
63
- const additional_instructions = `The user's name is ${userName} and they are located in ${location}.`;
64
-
65
- const runConfig: t.RunConfig = {
66
- runId: 'message-num-1',
67
- graphConfig: {
68
- type: 'standard',
69
- llmConfig,
70
- tools: [new TavilySearchResults(), createCodeExecutionTool()],
71
- instructions,
72
- additional_instructions,
73
- },
74
- returnContent: true,
75
- customHandlers,
76
- };
77
- const run = await Run.create<t.IState>(runConfig);
78
-
79
- const config = {
80
- configurable: {
81
- provider,
82
- thread_id: 'conversation-num-1',
83
- },
84
- streamMode: 'values',
85
- version: 'v2' as const,
86
- };
87
-
88
- console.log('Test 1: Simple Code Execution');
89
-
90
- // const userMessage1 = `how much memory is this (its in bytes) in MB? 31192000`;
91
- // const userMessage1 = `can you show me a good use case for rscript by running some code`;
92
- const userMessage1 = `Run hello world in french and in english, using python. please run 2 parallel code executions.`;
93
- const humanMessage = new HumanMessage(userMessage1);
94
- const tokenCounter = await createTokenCounter();
95
- const indexTokenCountMap = {
96
- 0: tokenCounter(humanMessage),
97
- };
98
-
99
- conversationHistory.push(humanMessage);
100
-
101
- let inputs = {
102
- messages: conversationHistory,
103
- };
104
- const finalContentParts1 = await run.processStream(inputs, config, {
105
- maxContextTokens: 8000,
106
- indexTokenCountMap,
107
- tokenCounter,
108
- });
109
- const finalMessages1 = run.getRunMessages();
110
- if (finalMessages1) {
111
- conversationHistory.push(...finalMessages1);
112
- }
113
- console.log('\n\n====================\n\n');
114
- console.dir(contentParts, { depth: null });
115
- }
116
-
117
- process.on('unhandledRejection', (reason, promise) => {
118
- console.error('Unhandled Rejection at:', promise, 'reason:', reason);
119
- console.log('Conversation history:');
120
- console.dir(conversationHistory, { depth: null });
121
- process.exit(1);
122
- });
123
-
124
- testCodeExecution().catch((err) => {
125
- console.error(err);
126
- console.log('Conversation history:');
127
- console.dir(conversationHistory, { depth: null });
128
- process.exit(1);
129
- });
1
+ // src/scripts/cli.ts
2
+ import { config } from 'dotenv';
3
+ config();
4
+ import { HumanMessage, BaseMessage } from '@langchain/core/messages';
5
+ import type * as t from '@/types';
6
+ import { ChatModelStreamHandler, createContentAggregator } from '@/stream';
7
+ import { createCodeExecutionTool } from '@/tools/CodeExecutor';
8
+ import { ToolEndHandler, ModelEndHandler } from '@/events';
9
+ import { createTokenCounter } from '@/utils/tokens';
10
+ import { getLLMConfig } from '@/utils/llmConfig';
11
+ import { getArgs } from '@/scripts/args';
12
+ import { GraphEvents } from '@/common';
13
+ import { Run } from '@/run';
14
+
15
+ const conversationHistory: BaseMessage[] = [];
16
+
17
+ async function testCodeExecution(): Promise<void> {
18
+ const { userName, location, provider, currentDate } = await getArgs();
19
+ const { contentParts, aggregateContent } = createContentAggregator();
20
+ const customHandlers = {
21
+ [GraphEvents.TOOL_END]: new ToolEndHandler(),
22
+ [GraphEvents.CHAT_MODEL_END]: new ModelEndHandler(),
23
+ [GraphEvents.CHAT_MODEL_STREAM]: new ChatModelStreamHandler(),
24
+ [GraphEvents.ON_RUN_STEP_COMPLETED]: {
25
+ handle: (
26
+ event: GraphEvents.ON_RUN_STEP_COMPLETED,
27
+ data: t.StreamEventData
28
+ ): void => {
29
+ console.log('====== ON_RUN_STEP_COMPLETED ======');
30
+ console.dir(data, { depth: null });
31
+ aggregateContent({
32
+ event,
33
+ data: data as unknown as { result: t.ToolEndEvent },
34
+ });
35
+ },
36
+ },
37
+ [GraphEvents.ON_RUN_STEP]: {
38
+ handle: (
39
+ event: GraphEvents.ON_RUN_STEP,
40
+ data: t.StreamEventData
41
+ ): void => {
42
+ console.log('====== ON_RUN_STEP ======');
43
+ console.dir(data, { depth: null });
44
+ aggregateContent({ event, data: data as t.RunStep });
45
+ },
46
+ },
47
+ [GraphEvents.ON_RUN_STEP_DELTA]: {
48
+ handle: (
49
+ event: GraphEvents.ON_RUN_STEP_DELTA,
50
+ data: t.StreamEventData
51
+ ): void => {
52
+ console.log('====== ON_RUN_STEP_DELTA ======');
53
+ console.dir(data, { depth: null });
54
+ aggregateContent({ event, data: data as t.RunStepDeltaEvent });
55
+ },
56
+ },
57
+ [GraphEvents.ON_MESSAGE_DELTA]: {
58
+ handle: (
59
+ event: GraphEvents.ON_MESSAGE_DELTA,
60
+ data: t.StreamEventData
61
+ ): void => {
62
+ console.log('====== ON_MESSAGE_DELTA ======');
63
+ console.dir(data, { depth: null });
64
+ aggregateContent({ event, data: data as t.MessageDeltaEvent });
65
+ },
66
+ },
67
+ [GraphEvents.TOOL_START]: {
68
+ handle: (
69
+ _event: string,
70
+ data: t.StreamEventData,
71
+ metadata?: Record<string, unknown>
72
+ ): void => {
73
+ console.log('====== TOOL_START ======');
74
+ console.dir(data, { depth: null });
75
+ },
76
+ },
77
+ };
78
+
79
+ const llmConfig = getLLMConfig(provider);
80
+ const instructions =
81
+ 'You are a friendly AI assistant with coding capabilities. Always address the user by their name.';
82
+ const additional_instructions = `The user's name is ${userName} and they are located in ${location}.`;
83
+
84
+ // const userMessage1 = `how much memory is this (its in bytes) in MB? 31192000`;
85
+ // const userMessage1 = `can you show me a good use case for rscript by running some code`;
86
+ const userMessage1 = `Run hello world in french and in english, using python. please run 2 parallel code executions.`;
87
+ const humanMessage = new HumanMessage(userMessage1);
88
+ const tokenCounter = await createTokenCounter();
89
+ const indexTokenCountMap = {
90
+ 0: tokenCounter(humanMessage),
91
+ };
92
+
93
+ const runConfig: t.RunConfig = {
94
+ runId: 'message-num-1',
95
+ graphConfig: {
96
+ type: 'standard',
97
+ llmConfig,
98
+ tools: [createCodeExecutionTool()],
99
+ instructions,
100
+ additional_instructions,
101
+ maxContextTokens: 8000,
102
+ },
103
+ returnContent: true,
104
+ customHandlers,
105
+ indexTokenCountMap,
106
+ tokenCounter,
107
+ };
108
+ const run = await Run.create<t.IState>(runConfig);
109
+
110
+ const config = {
111
+ configurable: {
112
+ provider,
113
+ thread_id: 'conversation-num-1',
114
+ },
115
+ streamMode: 'values',
116
+ version: 'v2' as const,
117
+ };
118
+
119
+ console.log('Test 1: Simple Code Execution');
120
+
121
+ conversationHistory.push(humanMessage);
122
+
123
+ let inputs = {
124
+ messages: conversationHistory,
125
+ };
126
+ const finalContentParts1 = await run.processStream(inputs, config);
127
+ const finalMessages1 = run.getRunMessages();
128
+ if (finalMessages1) {
129
+ conversationHistory.push(...finalMessages1);
130
+ }
131
+ console.log('\n\n====================\n\n');
132
+ console.dir(contentParts, { depth: null });
133
+ }
134
+
135
+ process.on('unhandledRejection', (reason, promise) => {
136
+ console.error('Unhandled Rejection at:', promise, 'reason:', reason);
137
+ console.log('Conversation history:');
138
+ console.dir(conversationHistory, { depth: null });
139
+ process.exit(1);
140
+ });
141
+
142
+ testCodeExecution().catch((err) => {
143
+ console.error(err);
144
+ console.log('Conversation history:');
145
+ console.dir(conversationHistory, { depth: null });
146
+ process.exit(1);
147
+ });
@@ -1,120 +1,138 @@
1
- /* eslint-disable no-console */
2
- // src/scripts/cli.ts
3
- import { config } from 'dotenv';
4
- config();
5
- import { HumanMessage, BaseMessage } from '@langchain/core/messages';
6
- import { TavilySearchResults } from '@langchain/community/tools/tavily_search';
7
- import type * as t from '@/types';
8
- import { ChatModelStreamHandler, createContentAggregator } from '@/stream';
9
- import { ToolEndHandler } from '@/events';
10
-
11
-
12
- import { getArgs } from '@/scripts/args';
13
- import { Run } from '@/run';
14
- import { GraphEvents, Callback } from '@/common';
15
- import { getLLMConfig } from '@/utils/llmConfig';
16
-
17
- const conversationHistory: BaseMessage[] = [];
18
- async function testStandardStreaming(): Promise<void> {
19
- const { userName, location, provider, currentDate } = await getArgs();
20
- const { contentParts, aggregateContent } = createContentAggregator();
21
- const customHandlers = {
22
- [GraphEvents.TOOL_END]: new ToolEndHandler(),
23
- // [GraphEvents.CHAT_MODEL_END]: new ModelEndHandler(),
24
- [GraphEvents.CHAT_MODEL_STREAM]: new ChatModelStreamHandler(),
25
- [GraphEvents.ON_RUN_STEP_COMPLETED]: {
26
- handle: (event: GraphEvents.ON_RUN_STEP_COMPLETED, data: t.StreamEventData): void => {
27
- console.log('====== ON_RUN_STEP_COMPLETED ======');
28
- // console.dir(data, { depth: null });
29
- aggregateContent({ event, data: data as unknown as { result: t.ToolEndEvent } });
30
- }
31
- },
32
- [GraphEvents.ON_RUN_STEP]: {
33
- handle: (event: GraphEvents.ON_RUN_STEP, data: t.StreamEventData): void => {
34
- console.log('====== ON_RUN_STEP ======');
35
- console.dir(data, { depth: null });
36
- aggregateContent({ event, data: data as t.RunStep });
37
- }
38
- },
39
- [GraphEvents.ON_RUN_STEP_DELTA]: {
40
- handle: (event: GraphEvents.ON_RUN_STEP_DELTA, data: t.StreamEventData): void => {
41
- console.log('====== ON_RUN_STEP_DELTA ======');
42
- console.dir(data, { depth: null });
43
- aggregateContent({ event, data: data as t.RunStepDeltaEvent });
44
- }
45
- },
46
- [GraphEvents.ON_MESSAGE_DELTA]: {
47
- handle: (event: GraphEvents.ON_MESSAGE_DELTA, data: t.StreamEventData): void => {
48
- console.log('====== ON_MESSAGE_DELTA ======');
49
- console.dir(data, { depth: null });
50
- aggregateContent({ event, data: data as t.MessageDeltaEvent });
51
- }
52
- },
53
- [GraphEvents.TOOL_START]: {
54
- handle: (_event: string, data: t.StreamEventData, metadata?: Record<string, unknown>): void => {
55
- console.log('====== TOOL_START ======');
56
- // console.dir(data, { depth: null });
57
- }
58
- },
59
- };
60
-
61
- const llmConfig = getLLMConfig(provider);
62
-
63
- const run = await Run.create<t.IState>({
64
- runId: 'test-run-id',
65
- graphConfig: {
66
- type: 'standard',
67
- llmConfig,
68
- tools: [new TavilySearchResults()],
69
- instructions: 'You are a friendly AI assistant. Always address the user by their name.',
70
- additional_instructions: `The user's name is ${userName} and they are located in ${location}.`,
71
- },
72
- returnContent: true,
73
- customHandlers,
74
- });
75
-
76
- const config = {
77
- configurable: {
78
- provider,
79
- thread_id: 'conversation-num-1',
80
- },
81
- streamMode: 'values',
82
- version: 'v2' as const,
83
- };
84
-
85
- console.log('Test 1: Weather query (content parts test)');
86
-
87
- const userMessage = `
88
- Make a search for the weather in ${location} today, which is ${currentDate}.
89
- Make sure to always refer to me by name, which is ${userName}.
90
- After giving me a thorough summary, tell me a joke about the weather forecast we went over.
91
- `;
92
-
93
- conversationHistory.push(new HumanMessage(userMessage));
94
-
95
- const inputs = {
96
- messages: conversationHistory,
97
- };
98
- const finalContentParts = await run.processStream(inputs, config);
99
- const finalMessages = run.getRunMessages();
100
- if (finalMessages) {
101
- conversationHistory.push(...finalMessages);
102
- console.dir(conversationHistory, { depth: null });
103
- }
104
- // console.dir(finalContentParts, { depth: null });
105
- console.log('\n\n====================\n\n');
106
- // console.dir(contentParts, { depth: null });
107
- }
108
-
109
- process.on('unhandledRejection', (reason, promise) => {
110
- console.error('Unhandled Rejection at:', promise, 'reason:', reason);
111
- console.log('Conversation history:');
112
- process.exit(1);
113
- });
114
-
115
- testStandardStreaming().catch((err) => {
116
- console.error(err);
117
- console.log('Conversation history:');
118
- console.dir(conversationHistory, { depth: null });
119
- process.exit(1);
120
- });
1
+ /* eslint-disable no-console */
2
+ // src/scripts/cli.ts
3
+ import { config } from 'dotenv';
4
+ config();
5
+ import { HumanMessage, BaseMessage } from '@langchain/core/messages';
6
+ import type * as t from '@/types';
7
+ import { ChatModelStreamHandler, createContentAggregator } from '@/stream';
8
+ import { ToolEndHandler } from '@/events';
9
+
10
+ import { getArgs } from '@/scripts/args';
11
+ import { Run } from '@/run';
12
+ import { GraphEvents, Callback } from '@/common';
13
+ import { getLLMConfig } from '@/utils/llmConfig';
14
+
15
+ const conversationHistory: BaseMessage[] = [];
16
+ async function testStandardStreaming(): Promise<void> {
17
+ const { userName, location, provider, currentDate } = await getArgs();
18
+ const { contentParts, aggregateContent } = createContentAggregator();
19
+ const customHandlers = {
20
+ [GraphEvents.TOOL_END]: new ToolEndHandler(),
21
+ // [GraphEvents.CHAT_MODEL_END]: new ModelEndHandler(),
22
+ [GraphEvents.CHAT_MODEL_STREAM]: new ChatModelStreamHandler(),
23
+ [GraphEvents.ON_RUN_STEP_COMPLETED]: {
24
+ handle: (
25
+ event: GraphEvents.ON_RUN_STEP_COMPLETED,
26
+ data: t.StreamEventData
27
+ ): void => {
28
+ console.log('====== ON_RUN_STEP_COMPLETED ======');
29
+ // console.dir(data, { depth: null });
30
+ aggregateContent({
31
+ event,
32
+ data: data as unknown as { result: t.ToolEndEvent },
33
+ });
34
+ },
35
+ },
36
+ [GraphEvents.ON_RUN_STEP]: {
37
+ handle: (
38
+ event: GraphEvents.ON_RUN_STEP,
39
+ data: t.StreamEventData
40
+ ): void => {
41
+ console.log('====== ON_RUN_STEP ======');
42
+ console.dir(data, { depth: null });
43
+ aggregateContent({ event, data: data as t.RunStep });
44
+ },
45
+ },
46
+ [GraphEvents.ON_RUN_STEP_DELTA]: {
47
+ handle: (
48
+ event: GraphEvents.ON_RUN_STEP_DELTA,
49
+ data: t.StreamEventData
50
+ ): void => {
51
+ console.log('====== ON_RUN_STEP_DELTA ======');
52
+ console.dir(data, { depth: null });
53
+ aggregateContent({ event, data: data as t.RunStepDeltaEvent });
54
+ },
55
+ },
56
+ [GraphEvents.ON_MESSAGE_DELTA]: {
57
+ handle: (
58
+ event: GraphEvents.ON_MESSAGE_DELTA,
59
+ data: t.StreamEventData
60
+ ): void => {
61
+ console.log('====== ON_MESSAGE_DELTA ======');
62
+ console.dir(data, { depth: null });
63
+ aggregateContent({ event, data: data as t.MessageDeltaEvent });
64
+ },
65
+ },
66
+ [GraphEvents.TOOL_START]: {
67
+ handle: (
68
+ _event: string,
69
+ data: t.StreamEventData,
70
+ metadata?: Record<string, unknown>
71
+ ): void => {
72
+ console.log('====== TOOL_START ======');
73
+ // console.dir(data, { depth: null });
74
+ },
75
+ },
76
+ };
77
+
78
+ const llmConfig = getLLMConfig(provider);
79
+
80
+ const run = await Run.create<t.IState>({
81
+ runId: 'test-run-id',
82
+ graphConfig: {
83
+ type: 'standard',
84
+ llmConfig,
85
+ tools: [],
86
+ instructions:
87
+ 'You are a friendly AI assistant. Always address the user by their name.',
88
+ additional_instructions: `The user's name is ${userName} and they are located in ${location}.`,
89
+ },
90
+ returnContent: true,
91
+ customHandlers,
92
+ });
93
+
94
+ const config = {
95
+ configurable: {
96
+ provider,
97
+ thread_id: 'conversation-num-1',
98
+ },
99
+ streamMode: 'values',
100
+ version: 'v2' as const,
101
+ };
102
+
103
+ console.log('Test 1: Weather query (content parts test)');
104
+
105
+ const userMessage = `
106
+ Make a search for the weather in ${location} today, which is ${currentDate}.
107
+ Make sure to always refer to me by name, which is ${userName}.
108
+ After giving me a thorough summary, tell me a joke about the weather forecast we went over.
109
+ `;
110
+
111
+ conversationHistory.push(new HumanMessage(userMessage));
112
+
113
+ const inputs = {
114
+ messages: conversationHistory,
115
+ };
116
+ const finalContentParts = await run.processStream(inputs, config);
117
+ const finalMessages = run.getRunMessages();
118
+ if (finalMessages) {
119
+ conversationHistory.push(...finalMessages);
120
+ console.dir(conversationHistory, { depth: null });
121
+ }
122
+ // console.dir(finalContentParts, { depth: null });
123
+ console.log('\n\n====================\n\n');
124
+ // console.dir(contentParts, { depth: null });
125
+ }
126
+
127
+ process.on('unhandledRejection', (reason, promise) => {
128
+ console.error('Unhandled Rejection at:', promise, 'reason:', reason);
129
+ console.log('Conversation history:');
130
+ process.exit(1);
131
+ });
132
+
133
+ testStandardStreaming().catch((err) => {
134
+ console.error(err);
135
+ console.log('Conversation history:');
136
+ console.dir(conversationHistory, { depth: null });
137
+ process.exit(1);
138
+ });