beeai-framework 0.1.19 → 0.1.21

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (375) hide show
  1. package/dist/adapters/a2a/agents/agent.cjs +4 -4
  2. package/dist/adapters/a2a/agents/agent.cjs.map +1 -1
  3. package/dist/adapters/a2a/agents/agent.d.cts +2 -2
  4. package/dist/adapters/a2a/agents/agent.d.ts +2 -2
  5. package/dist/adapters/a2a/agents/agent.js +4 -4
  6. package/dist/adapters/a2a/agents/agent.js.map +1 -1
  7. package/dist/adapters/a2a/agents/events.d.cts +1 -1
  8. package/dist/adapters/a2a/agents/events.d.ts +1 -1
  9. package/dist/adapters/a2a/agents/types.d.cts +2 -2
  10. package/dist/adapters/a2a/agents/types.d.ts +2 -2
  11. package/dist/adapters/a2a/agents/utils.cjs +2 -2
  12. package/dist/adapters/a2a/agents/utils.cjs.map +1 -1
  13. package/dist/adapters/a2a/agents/utils.js +2 -2
  14. package/dist/adapters/a2a/agents/utils.js.map +1 -1
  15. package/dist/adapters/a2a/serve/agent_executor.cjs +1 -3
  16. package/dist/adapters/a2a/serve/agent_executor.cjs.map +1 -1
  17. package/dist/adapters/a2a/serve/agent_executor.d.cts +4 -3
  18. package/dist/adapters/a2a/serve/agent_executor.d.ts +4 -3
  19. package/dist/adapters/a2a/serve/agent_executor.js +1 -3
  20. package/dist/adapters/a2a/serve/agent_executor.js.map +1 -1
  21. package/dist/adapters/a2a/serve/server.cjs +15 -0
  22. package/dist/adapters/a2a/serve/server.cjs.map +1 -1
  23. package/dist/adapters/a2a/serve/server.d.cts +4 -2
  24. package/dist/adapters/a2a/serve/server.d.ts +4 -2
  25. package/dist/adapters/a2a/serve/server.js +15 -0
  26. package/dist/adapters/a2a/serve/server.js.map +1 -1
  27. package/dist/adapters/acp/agents/agent.d.cts +2 -2
  28. package/dist/adapters/acp/agents/agent.d.ts +2 -2
  29. package/dist/adapters/acp/agents/events.d.cts +1 -1
  30. package/dist/adapters/acp/agents/events.d.ts +1 -1
  31. package/dist/adapters/{beeai_platform → agentstack}/agents/agent.cjs +4 -4
  32. package/dist/adapters/agentstack/agents/agent.cjs.map +1 -0
  33. package/dist/adapters/{beeai_platform → agentstack}/agents/agent.d.cts +12 -12
  34. package/dist/adapters/{beeai_platform → agentstack}/agents/agent.d.ts +12 -12
  35. package/dist/adapters/{beeai_platform → agentstack}/agents/agent.js +4 -4
  36. package/dist/adapters/agentstack/agents/agent.js.map +1 -0
  37. package/dist/adapters/agentstack/agents/events.d.cts +23 -0
  38. package/dist/adapters/agentstack/agents/events.d.ts +23 -0
  39. package/dist/adapters/{beeai_platform → agentstack}/agents/types.d.cts +4 -4
  40. package/dist/adapters/{beeai_platform → agentstack}/agents/types.d.ts +4 -4
  41. package/dist/adapters/amazon-bedrock/backend/chat.cjs +3 -2
  42. package/dist/adapters/amazon-bedrock/backend/chat.cjs.map +1 -1
  43. package/dist/adapters/amazon-bedrock/backend/chat.d.cts +5 -6
  44. package/dist/adapters/amazon-bedrock/backend/chat.d.ts +5 -6
  45. package/dist/adapters/amazon-bedrock/backend/chat.js +3 -2
  46. package/dist/adapters/amazon-bedrock/backend/chat.js.map +1 -1
  47. package/dist/adapters/amazon-bedrock/backend/embedding.cjs +2 -2
  48. package/dist/adapters/amazon-bedrock/backend/embedding.cjs.map +1 -1
  49. package/dist/adapters/amazon-bedrock/backend/embedding.d.cts +4 -4
  50. package/dist/adapters/amazon-bedrock/backend/embedding.d.ts +4 -4
  51. package/dist/adapters/amazon-bedrock/backend/embedding.js +2 -2
  52. package/dist/adapters/amazon-bedrock/backend/embedding.js.map +1 -1
  53. package/dist/adapters/anthropic/backend/chat.cjs +3 -2
  54. package/dist/adapters/anthropic/backend/chat.cjs.map +1 -1
  55. package/dist/adapters/anthropic/backend/chat.d.cts +5 -6
  56. package/dist/adapters/anthropic/backend/chat.d.ts +5 -6
  57. package/dist/adapters/anthropic/backend/chat.js +3 -2
  58. package/dist/adapters/anthropic/backend/chat.js.map +1 -1
  59. package/dist/adapters/anthropic/backend/embedding.d.cts +2 -2
  60. package/dist/adapters/anthropic/backend/embedding.d.ts +2 -2
  61. package/dist/adapters/azure-openai/backend/chat.cjs +3 -2
  62. package/dist/adapters/azure-openai/backend/chat.cjs.map +1 -1
  63. package/dist/adapters/azure-openai/backend/chat.d.cts +5 -6
  64. package/dist/adapters/azure-openai/backend/chat.d.ts +5 -6
  65. package/dist/adapters/azure-openai/backend/chat.js +3 -2
  66. package/dist/adapters/azure-openai/backend/chat.js.map +1 -1
  67. package/dist/adapters/azure-openai/backend/embedding.cjs +2 -2
  68. package/dist/adapters/azure-openai/backend/embedding.cjs.map +1 -1
  69. package/dist/adapters/azure-openai/backend/embedding.d.cts +3 -3
  70. package/dist/adapters/azure-openai/backend/embedding.d.ts +3 -3
  71. package/dist/adapters/azure-openai/backend/embedding.js +2 -2
  72. package/dist/adapters/azure-openai/backend/embedding.js.map +1 -1
  73. package/dist/adapters/dummy/backend/chat.d.cts +2 -2
  74. package/dist/adapters/dummy/backend/chat.d.ts +2 -2
  75. package/dist/adapters/dummy/backend/embedding.d.cts +2 -2
  76. package/dist/adapters/dummy/backend/embedding.d.ts +2 -2
  77. package/dist/adapters/google-vertex/backend/chat.cjs +3 -2
  78. package/dist/adapters/google-vertex/backend/chat.cjs.map +1 -1
  79. package/dist/adapters/google-vertex/backend/chat.d.cts +5 -6
  80. package/dist/adapters/google-vertex/backend/chat.d.ts +5 -6
  81. package/dist/adapters/google-vertex/backend/chat.js +3 -2
  82. package/dist/adapters/google-vertex/backend/chat.js.map +1 -1
  83. package/dist/adapters/google-vertex/backend/embedding.d.cts +2 -2
  84. package/dist/adapters/google-vertex/backend/embedding.d.ts +2 -2
  85. package/dist/adapters/groq/backend/chat.cjs +3 -2
  86. package/dist/adapters/groq/backend/chat.cjs.map +1 -1
  87. package/dist/adapters/groq/backend/chat.d.cts +5 -6
  88. package/dist/adapters/groq/backend/chat.d.ts +5 -6
  89. package/dist/adapters/groq/backend/chat.js +3 -2
  90. package/dist/adapters/groq/backend/chat.js.map +1 -1
  91. package/dist/adapters/groq/backend/embedding.d.cts +2 -2
  92. package/dist/adapters/groq/backend/embedding.d.ts +2 -2
  93. package/dist/adapters/langchain/backend/chat.cjs +6 -5
  94. package/dist/adapters/langchain/backend/chat.cjs.map +1 -1
  95. package/dist/adapters/langchain/backend/chat.d.cts +2 -2
  96. package/dist/adapters/langchain/backend/chat.d.ts +2 -2
  97. package/dist/adapters/langchain/backend/chat.js +6 -5
  98. package/dist/adapters/langchain/backend/chat.js.map +1 -1
  99. package/dist/adapters/langchain/backend/embedding.d.cts +3 -3
  100. package/dist/adapters/langchain/backend/embedding.d.ts +3 -3
  101. package/dist/adapters/langchain/tools.d.cts +1 -1
  102. package/dist/adapters/langchain/tools.d.ts +1 -1
  103. package/dist/adapters/mcp/serve/server.cjs +12 -4
  104. package/dist/adapters/mcp/serve/server.cjs.map +1 -1
  105. package/dist/adapters/mcp/serve/server.d.cts +1 -1
  106. package/dist/adapters/mcp/serve/server.d.ts +1 -1
  107. package/dist/adapters/mcp/serve/server.js +12 -4
  108. package/dist/adapters/mcp/serve/server.js.map +1 -1
  109. package/dist/adapters/ollama/backend/chat.cjs +1 -4
  110. package/dist/adapters/ollama/backend/chat.cjs.map +1 -1
  111. package/dist/adapters/ollama/backend/chat.d.cts +4 -4
  112. package/dist/adapters/ollama/backend/chat.d.ts +4 -4
  113. package/dist/adapters/ollama/backend/chat.js +1 -4
  114. package/dist/adapters/ollama/backend/chat.js.map +1 -1
  115. package/dist/adapters/ollama/backend/client.cjs +2 -2
  116. package/dist/adapters/ollama/backend/client.cjs.map +1 -1
  117. package/dist/adapters/ollama/backend/client.d.cts +1 -1
  118. package/dist/adapters/ollama/backend/client.d.ts +1 -1
  119. package/dist/adapters/ollama/backend/client.js +1 -1
  120. package/dist/adapters/ollama/backend/client.js.map +1 -1
  121. package/dist/adapters/ollama/backend/embedding.cjs.map +1 -1
  122. package/dist/adapters/ollama/backend/embedding.d.cts +3 -3
  123. package/dist/adapters/ollama/backend/embedding.d.ts +3 -3
  124. package/dist/adapters/ollama/backend/embedding.js.map +1 -1
  125. package/dist/adapters/openai/backend/chat.cjs +3 -2
  126. package/dist/adapters/openai/backend/chat.cjs.map +1 -1
  127. package/dist/adapters/openai/backend/chat.d.cts +5 -6
  128. package/dist/adapters/openai/backend/chat.d.ts +5 -6
  129. package/dist/adapters/openai/backend/chat.js +3 -2
  130. package/dist/adapters/openai/backend/chat.js.map +1 -1
  131. package/dist/adapters/openai/backend/client.cjs +2 -11
  132. package/dist/adapters/openai/backend/client.cjs.map +1 -1
  133. package/dist/adapters/openai/backend/client.js +3 -12
  134. package/dist/adapters/openai/backend/client.js.map +1 -1
  135. package/dist/adapters/openai/backend/embedding.cjs +2 -2
  136. package/dist/adapters/openai/backend/embedding.cjs.map +1 -1
  137. package/dist/adapters/openai/backend/embedding.d.cts +4 -4
  138. package/dist/adapters/openai/backend/embedding.d.ts +4 -4
  139. package/dist/adapters/openai/backend/embedding.js +2 -2
  140. package/dist/adapters/openai/backend/embedding.js.map +1 -1
  141. package/dist/adapters/vercel/backend/chat.cjs +26 -21
  142. package/dist/adapters/vercel/backend/chat.cjs.map +1 -1
  143. package/dist/adapters/vercel/backend/chat.d.cts +6 -4
  144. package/dist/adapters/vercel/backend/chat.d.ts +6 -4
  145. package/dist/adapters/vercel/backend/chat.js +27 -22
  146. package/dist/adapters/vercel/backend/chat.js.map +1 -1
  147. package/dist/adapters/vercel/backend/embedding.cjs.map +1 -1
  148. package/dist/adapters/vercel/backend/embedding.d.cts +3 -3
  149. package/dist/adapters/vercel/backend/embedding.d.ts +3 -3
  150. package/dist/adapters/vercel/backend/embedding.js.map +1 -1
  151. package/dist/adapters/vercel/backend/utils.cjs +11 -0
  152. package/dist/adapters/vercel/backend/utils.cjs.map +1 -1
  153. package/dist/adapters/vercel/backend/utils.d.cts +19 -2
  154. package/dist/adapters/vercel/backend/utils.d.ts +19 -2
  155. package/dist/adapters/vercel/backend/utils.js +11 -1
  156. package/dist/adapters/vercel/backend/utils.js.map +1 -1
  157. package/dist/adapters/watsonx/backend/chat.cjs +10 -4
  158. package/dist/adapters/watsonx/backend/chat.cjs.map +1 -1
  159. package/dist/adapters/watsonx/backend/chat.d.cts +2 -2
  160. package/dist/adapters/watsonx/backend/chat.d.ts +2 -2
  161. package/dist/adapters/watsonx/backend/chat.js +10 -4
  162. package/dist/adapters/watsonx/backend/chat.js.map +1 -1
  163. package/dist/adapters/watsonx/backend/embedding.d.cts +2 -2
  164. package/dist/adapters/watsonx/backend/embedding.d.ts +2 -2
  165. package/dist/adapters/xai/backend/chat.cjs +3 -2
  166. package/dist/adapters/xai/backend/chat.cjs.map +1 -1
  167. package/dist/adapters/xai/backend/chat.d.cts +5 -6
  168. package/dist/adapters/xai/backend/chat.d.ts +5 -6
  169. package/dist/adapters/xai/backend/chat.js +3 -2
  170. package/dist/adapters/xai/backend/chat.js.map +1 -1
  171. package/dist/{agent-7NzfPC6d.d.ts → agent-CHebzsQB.d.ts} +4 -4
  172. package/dist/{agent-B6pOfx0K.d.cts → agent-DvvVGvLr.d.cts} +4 -4
  173. package/dist/agents/base.d.cts +2 -2
  174. package/dist/agents/base.d.ts +2 -2
  175. package/dist/agents/experimental/replan/agent.d.cts +3 -3
  176. package/dist/agents/experimental/replan/agent.d.ts +3 -3
  177. package/dist/agents/experimental/replan/prompts.d.cts +1 -1
  178. package/dist/agents/experimental/replan/prompts.d.ts +1 -1
  179. package/dist/agents/experimental/streamlit/agent.d.cts +3 -3
  180. package/dist/agents/experimental/streamlit/agent.d.ts +3 -3
  181. package/dist/agents/react/agent.d.cts +5 -4
  182. package/dist/agents/react/agent.d.ts +5 -4
  183. package/dist/agents/react/prompts.d.cts +1 -1
  184. package/dist/agents/react/prompts.d.ts +1 -1
  185. package/dist/agents/react/runners/base.cjs.map +1 -1
  186. package/dist/agents/react/runners/base.d.cts +5 -4
  187. package/dist/agents/react/runners/base.d.ts +5 -4
  188. package/dist/agents/react/runners/base.js.map +1 -1
  189. package/dist/agents/react/runners/deep-think/prompts.d.cts +1 -1
  190. package/dist/agents/react/runners/deep-think/prompts.d.ts +1 -1
  191. package/dist/agents/react/runners/deep-think/runner.d.cts +6 -5
  192. package/dist/agents/react/runners/deep-think/runner.d.ts +6 -5
  193. package/dist/agents/react/runners/default/runner.cjs.map +1 -1
  194. package/dist/agents/react/runners/default/runner.d.cts +6 -5
  195. package/dist/agents/react/runners/default/runner.d.ts +6 -5
  196. package/dist/agents/react/runners/default/runner.js.map +1 -1
  197. package/dist/agents/react/runners/granite/prompts.d.cts +1 -1
  198. package/dist/agents/react/runners/granite/prompts.d.ts +1 -1
  199. package/dist/agents/react/runners/granite/runner.cjs +4 -2
  200. package/dist/agents/react/runners/granite/runner.cjs.map +1 -1
  201. package/dist/agents/react/runners/granite/runner.d.cts +6 -5
  202. package/dist/agents/react/runners/granite/runner.d.ts +6 -5
  203. package/dist/agents/react/runners/granite/runner.js +4 -2
  204. package/dist/agents/react/runners/granite/runner.js.map +1 -1
  205. package/dist/agents/react/types.d.cts +3 -2
  206. package/dist/agents/react/types.d.ts +3 -2
  207. package/dist/agents/toolCalling/agent.cjs +9 -5
  208. package/dist/agents/toolCalling/agent.cjs.map +1 -1
  209. package/dist/agents/toolCalling/agent.d.cts +3 -3
  210. package/dist/agents/toolCalling/agent.d.ts +3 -3
  211. package/dist/agents/toolCalling/agent.js +9 -5
  212. package/dist/agents/toolCalling/agent.js.map +1 -1
  213. package/dist/agents/toolCalling/types.d.cts +1 -1
  214. package/dist/agents/toolCalling/types.d.ts +1 -1
  215. package/dist/agents/types.d.cts +2 -2
  216. package/dist/agents/types.d.ts +2 -2
  217. package/dist/backend/backend.d.cts +2 -2
  218. package/dist/backend/backend.d.ts +2 -2
  219. package/dist/backend/chat.cjs +1 -1
  220. package/dist/backend/chat.cjs.map +1 -1
  221. package/dist/backend/chat.d.cts +2 -2
  222. package/dist/backend/chat.d.ts +2 -2
  223. package/dist/backend/chat.js +1 -1
  224. package/dist/backend/chat.js.map +1 -1
  225. package/dist/backend/core.d.cts +2 -2
  226. package/dist/backend/core.d.ts +2 -2
  227. package/dist/backend/embedding.d.cts +2 -2
  228. package/dist/backend/embedding.d.ts +2 -2
  229. package/dist/backend/message.cjs +59 -3
  230. package/dist/backend/message.cjs.map +1 -1
  231. package/dist/backend/message.d.cts +3 -0
  232. package/dist/backend/message.d.ts +3 -0
  233. package/dist/backend/message.js +59 -3
  234. package/dist/backend/message.js.map +1 -1
  235. package/dist/backend/utils.d.cts +2 -2
  236. package/dist/backend/utils.d.ts +2 -2
  237. package/dist/{base-Bguef5Dg.d.ts → base-BQgDlcGa.d.ts} +1 -1
  238. package/dist/{base-C1hO5EJe.d.cts → base-CwcxwybB.d.cts} +1 -1
  239. package/dist/cache/fileCache.cjs.map +1 -1
  240. package/dist/cache/fileCache.js.map +1 -1
  241. package/dist/{chat-BBoOSvzm.d.ts → chat-C0s-o6ll.d.ts} +4 -2
  242. package/dist/{chat-BZ55YQab.d.cts → chat-CRb3vUVg.d.cts} +4 -2
  243. package/dist/context.d.cts +1 -1
  244. package/dist/context.d.ts +1 -1
  245. package/dist/emitter/emitter.d.cts +1 -1
  246. package/dist/emitter/emitter.d.ts +1 -1
  247. package/dist/emitter/types.d.cts +1 -1
  248. package/dist/emitter/types.d.ts +1 -1
  249. package/dist/{emitter-BqpLJQVb.d.ts → emitter-CZFbzlUi.d.ts} +3 -3
  250. package/dist/{emitter-pJzHC_AM.d.cts → emitter-DpqUYjXH.d.cts} +3 -3
  251. package/dist/index.d.cts +3 -3
  252. package/dist/index.d.ts +3 -3
  253. package/dist/internals/fetcher.d.cts +1 -1
  254. package/dist/internals/fetcher.d.ts +1 -1
  255. package/dist/internals/helpers/array.cjs +35 -0
  256. package/dist/internals/helpers/array.cjs.map +1 -1
  257. package/dist/internals/helpers/array.d.cts +6 -1
  258. package/dist/internals/helpers/array.d.ts +6 -1
  259. package/dist/internals/helpers/array.js +35 -1
  260. package/dist/internals/helpers/array.js.map +1 -1
  261. package/dist/internals/helpers/object.cjs +30 -5
  262. package/dist/internals/helpers/object.cjs.map +1 -1
  263. package/dist/internals/helpers/object.d.cts +3 -2
  264. package/dist/internals/helpers/object.d.ts +3 -2
  265. package/dist/internals/helpers/object.js +29 -5
  266. package/dist/internals/helpers/object.js.map +1 -1
  267. package/dist/logger/logger.cjs +8 -4
  268. package/dist/logger/logger.cjs.map +1 -1
  269. package/dist/logger/logger.d.cts +21 -15
  270. package/dist/logger/logger.d.ts +21 -15
  271. package/dist/logger/logger.js +6 -3
  272. package/dist/logger/logger.js.map +1 -1
  273. package/dist/logger/pretty.cjs +2 -2
  274. package/dist/logger/pretty.cjs.map +1 -1
  275. package/dist/logger/pretty.js +2 -2
  276. package/dist/logger/pretty.js.map +1 -1
  277. package/dist/memory/summarizeMemory.d.cts +2 -2
  278. package/dist/memory/summarizeMemory.d.ts +2 -2
  279. package/dist/parsers/linePrefix.cjs.map +1 -1
  280. package/dist/parsers/linePrefix.d.cts +4 -8
  281. package/dist/parsers/linePrefix.d.ts +4 -8
  282. package/dist/parsers/linePrefix.js.map +1 -1
  283. package/dist/serializer/utils.cjs.map +1 -1
  284. package/dist/serializer/utils.js.map +1 -1
  285. package/dist/tools/arxiv.cjs.map +1 -1
  286. package/dist/tools/arxiv.d.cts +1 -1
  287. package/dist/tools/arxiv.d.ts +1 -1
  288. package/dist/tools/arxiv.js.map +1 -1
  289. package/dist/tools/base.cjs.map +1 -1
  290. package/dist/tools/base.d.cts +1 -1
  291. package/dist/tools/base.d.ts +1 -1
  292. package/dist/tools/base.js.map +1 -1
  293. package/dist/tools/calculator.d.cts +1 -1
  294. package/dist/tools/calculator.d.ts +1 -1
  295. package/dist/tools/custom.d.cts +2 -2
  296. package/dist/tools/custom.d.ts +2 -2
  297. package/dist/tools/database/elasticsearch.cjs.map +1 -1
  298. package/dist/tools/database/elasticsearch.d.cts +1 -1
  299. package/dist/tools/database/elasticsearch.d.ts +1 -1
  300. package/dist/tools/database/elasticsearch.js.map +1 -1
  301. package/dist/tools/database/milvus.cjs +3 -3
  302. package/dist/tools/database/milvus.cjs.map +1 -1
  303. package/dist/tools/database/milvus.d.cts +1 -1
  304. package/dist/tools/database/milvus.d.ts +1 -1
  305. package/dist/tools/database/milvus.js +3 -3
  306. package/dist/tools/database/milvus.js.map +1 -1
  307. package/dist/tools/database/qdrant.cjs +2 -2
  308. package/dist/tools/database/qdrant.cjs.map +1 -1
  309. package/dist/tools/database/qdrant.d.cts +1 -1
  310. package/dist/tools/database/qdrant.d.ts +1 -1
  311. package/dist/tools/database/qdrant.js +2 -2
  312. package/dist/tools/database/qdrant.js.map +1 -1
  313. package/dist/tools/database/sql.cjs.map +1 -1
  314. package/dist/tools/database/sql.d.cts +1 -1
  315. package/dist/tools/database/sql.d.ts +1 -1
  316. package/dist/tools/database/sql.js.map +1 -1
  317. package/dist/tools/llm.d.cts +2 -2
  318. package/dist/tools/llm.d.ts +2 -2
  319. package/dist/tools/mcp.d.cts +1 -1
  320. package/dist/tools/mcp.d.ts +1 -1
  321. package/dist/tools/openapi.d.cts +1 -1
  322. package/dist/tools/openapi.d.ts +1 -1
  323. package/dist/tools/python/output.d.cts +1 -1
  324. package/dist/tools/python/output.d.ts +1 -1
  325. package/dist/tools/python/python.d.cts +2 -2
  326. package/dist/tools/python/python.d.ts +2 -2
  327. package/dist/tools/search/base.cjs.map +1 -1
  328. package/dist/tools/search/base.d.cts +1 -1
  329. package/dist/tools/search/base.d.ts +1 -1
  330. package/dist/tools/search/base.js.map +1 -1
  331. package/dist/tools/search/duckDuckGoSearch.cjs.map +1 -1
  332. package/dist/tools/search/duckDuckGoSearch.d.cts +1 -1
  333. package/dist/tools/search/duckDuckGoSearch.d.ts +1 -1
  334. package/dist/tools/search/duckDuckGoSearch.js.map +1 -1
  335. package/dist/tools/search/googleSearch.cjs.map +1 -1
  336. package/dist/tools/search/googleSearch.d.cts +1 -1
  337. package/dist/tools/search/googleSearch.d.ts +1 -1
  338. package/dist/tools/search/googleSearch.js.map +1 -1
  339. package/dist/tools/search/searXNGSearch.d.cts +1 -1
  340. package/dist/tools/search/searXNGSearch.d.ts +1 -1
  341. package/dist/tools/search/wikipedia.cjs +25 -0
  342. package/dist/tools/search/wikipedia.cjs.map +1 -1
  343. package/dist/tools/search/wikipedia.d.cts +2 -3
  344. package/dist/tools/search/wikipedia.d.ts +2 -3
  345. package/dist/tools/search/wikipedia.js +23 -0
  346. package/dist/tools/search/wikipedia.js.map +1 -1
  347. package/dist/tools/similarity.d.cts +1 -1
  348. package/dist/tools/similarity.d.ts +1 -1
  349. package/dist/tools/weather/openMeteo.cjs.map +1 -1
  350. package/dist/tools/weather/openMeteo.d.cts +1 -1
  351. package/dist/tools/weather/openMeteo.d.ts +1 -1
  352. package/dist/tools/weather/openMeteo.js.map +1 -1
  353. package/dist/tools/web/webCrawler.cjs.map +1 -1
  354. package/dist/tools/web/webCrawler.d.cts +1 -1
  355. package/dist/tools/web/webCrawler.d.ts +1 -1
  356. package/dist/tools/web/webCrawler.js.map +1 -1
  357. package/dist/version.cjs +1 -1
  358. package/dist/version.js +1 -1
  359. package/dist/workflows/agent.d.cts +3 -3
  360. package/dist/workflows/agent.d.ts +3 -3
  361. package/dist/workflows/workflow.d.cts +1 -1
  362. package/dist/workflows/workflow.d.ts +1 -1
  363. package/package.json +68 -67
  364. package/dist/adapters/beeai_platform/agents/agent.cjs.map +0 -1
  365. package/dist/adapters/beeai_platform/agents/agent.js.map +0 -1
  366. package/dist/adapters/beeai_platform/agents/events.d.cts +0 -23
  367. package/dist/adapters/beeai_platform/agents/events.d.ts +0 -23
  368. /package/dist/adapters/{beeai_platform → agentstack}/agents/events.cjs +0 -0
  369. /package/dist/adapters/{beeai_platform → agentstack}/agents/events.cjs.map +0 -0
  370. /package/dist/adapters/{beeai_platform → agentstack}/agents/events.js +0 -0
  371. /package/dist/adapters/{beeai_platform → agentstack}/agents/events.js.map +0 -0
  372. /package/dist/adapters/{beeai_platform → agentstack}/agents/types.cjs +0 -0
  373. /package/dist/adapters/{beeai_platform → agentstack}/agents/types.cjs.map +0 -0
  374. /package/dist/adapters/{beeai_platform → agentstack}/agents/types.js +0 -0
  375. /package/dist/adapters/{beeai_platform → agentstack}/agents/types.js.map +0 -0
@@ -1 +1 @@
1
- {"version":3,"sources":["../../../../src/adapters/azure-openai/backend/embedding.ts"],"names":["AzureOpenAIEmbeddingModel","VercelEmbeddingModel","modelId","getEnv","settings","client","model","AzureOpenAIClient","ensure","instance","textEmbeddingModel"],"mappings":";;;;;;AAiBO,MAAMA,kCAAkCC,oBAAAA,CAAAA;EAjB/C;;;EAkBE,WAAA,CACEC,OAAAA,GAAuCC,OACrC,8BAAA,EACA,wBAAA,GAEFC,QAAAA,GAA8C,IAC9CC,MAAAA,EACA;AACA,IAAA,MAAMC,KAAAA,GAAQC,kBAAkBC,MAAAA,CAAOH,MAAAA,EAAQI,QAAAA,CAASC,kBAAAA,CAAmBR,SAASE,QAAAA,CAAAA;AACpF,IAAA,KAAA,CAAME,KAAAA,CAAAA;AACR;AACF","file":"embedding.js","sourcesContent":["/**\n * Copyright 2025 © BeeAI a Series of LF Projects, LLC\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport { VercelEmbeddingModel } from \"@/adapters/vercel/backend/embedding.js\";\nimport {\n AzureOpenAIClient,\n AzureOpenAIClientSettings,\n} from \"@/adapters/azure-openai/backend/client.js\";\nimport { getEnv } from \"@/internals/env.js\";\nimport { AzureOpenAIProvider as VercelAzureOpenAIProviderSettings } from \"@ai-sdk/azure\";\n\ntype AzureOpenAIParameters = Parameters<VercelAzureOpenAIProviderSettings[\"textEmbeddingModel\"]>;\nexport type AzureOpenAIEmbeddingModelId = NonNullable<AzureOpenAIParameters[0]>;\nexport type AzureOpenAIEmbeddingModelSettings = Record<string, any>;\n\nexport class AzureOpenAIEmbeddingModel extends VercelEmbeddingModel {\n constructor(\n modelId: AzureOpenAIEmbeddingModelId = getEnv(\n \"AZURE_OPENAI_EMBEDDING_MODEL\",\n \"text-embedding-3-small\",\n ),\n settings: AzureOpenAIEmbeddingModelSettings = {},\n client?: AzureOpenAIClient | AzureOpenAIClientSettings,\n ) {\n const model = AzureOpenAIClient.ensure(client).instance.textEmbeddingModel(modelId, settings);\n super(model);\n }\n}\n"]}
1
+ {"version":3,"sources":["../../../../src/adapters/azure-openai/backend/embedding.ts"],"names":["AzureOpenAIEmbeddingModel","VercelEmbeddingModel","modelId","getEnv","_settings","client","model","AzureOpenAIClient","ensure","instance","textEmbeddingModel"],"mappings":";;;;;;AAiBO,MAAMA,kCAAkCC,oBAAAA,CAAAA;EAjB/C;;;EAkBE,WAAA,CACEC,OAAAA,GAAuCC,OACrC,8BAAA,EACA,wBAAA,GAEFC,SAAAA,GAA+C,IAC/CC,MAAAA,EACA;AACA,IAAA,MAAMC,QAAQC,iBAAAA,CAAkBC,MAAAA,CAAOH,MAAAA,CAAAA,CAAQI,QAAAA,CAASC,mBAAmBR,OAAAA,CAAAA;AAC3E,IAAA,KAAA,CAAMI,KAAAA,CAAAA;AACR;AACF","file":"embedding.js","sourcesContent":["/**\n * Copyright 2025 © BeeAI a Series of LF Projects, LLC\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport { VercelEmbeddingModel } from \"@/adapters/vercel/backend/embedding.js\";\nimport {\n AzureOpenAIClient,\n AzureOpenAIClientSettings,\n} from \"@/adapters/azure-openai/backend/client.js\";\nimport { getEnv } from \"@/internals/env.js\";\nimport { AzureOpenAIProvider as VercelAzureOpenAIProviderSettings } from \"@ai-sdk/azure\";\n\ntype AzureOpenAIParameters = Parameters<VercelAzureOpenAIProviderSettings[\"textEmbeddingModel\"]>;\nexport type AzureOpenAIEmbeddingModelId = NonNullable<AzureOpenAIParameters[0]>;\nexport type AzureOpenAIEmbeddingModelSettings = Record<string, any>;\n\nexport class AzureOpenAIEmbeddingModel extends VercelEmbeddingModel {\n constructor(\n modelId: AzureOpenAIEmbeddingModelId = getEnv(\n \"AZURE_OPENAI_EMBEDDING_MODEL\",\n \"text-embedding-3-small\",\n ),\n _settings: AzureOpenAIEmbeddingModelSettings = {},\n client?: AzureOpenAIClient | AzureOpenAIClientSettings,\n ) {\n const model = AzureOpenAIClient.ensure(client).instance.textEmbeddingModel(modelId);\n super(model);\n }\n}\n"]}
@@ -1,7 +1,7 @@
1
1
  import { Logger } from '../../../logger/logger.cjs';
2
- import { m as ChatModel, C as ChatModelParameters, g as ChatModelEvents, d as ChatModelInput, n as ChatModelOutput, i as ChatModelCache, l as ChatModelToolChoiceSupport } from '../../../chat-BZ55YQab.cjs';
2
+ import { m as ChatModel, C as ChatModelParameters, g as ChatModelEvents, d as ChatModelInput, n as ChatModelOutput, i as ChatModelCache, l as ChatModelToolChoiceSupport } from '../../../chat-CRb3vUVg.cjs';
3
3
  import { GetRunContext } from '../../../context.cjs';
4
- import { E as Emitter } from '../../../emitter-pJzHC_AM.cjs';
4
+ import { E as Emitter } from '../../../emitter-DpqUYjXH.cjs';
5
5
  import 'pino';
6
6
  import '../../../errors.cjs';
7
7
  import '../../../internals/types.cjs';
@@ -1,7 +1,7 @@
1
1
  import { Logger } from '../../../logger/logger.js';
2
- import { m as ChatModel, C as ChatModelParameters, g as ChatModelEvents, d as ChatModelInput, n as ChatModelOutput, i as ChatModelCache, l as ChatModelToolChoiceSupport } from '../../../chat-BBoOSvzm.js';
2
+ import { m as ChatModel, C as ChatModelParameters, g as ChatModelEvents, d as ChatModelInput, n as ChatModelOutput, i as ChatModelCache, l as ChatModelToolChoiceSupport } from '../../../chat-C0s-o6ll.js';
3
3
  import { GetRunContext } from '../../../context.js';
4
- import { E as Emitter } from '../../../emitter-BqpLJQVb.js';
4
+ import { E as Emitter } from '../../../emitter-CZFbzlUi.js';
5
5
  import 'pino';
6
6
  import '../../../errors.js';
7
7
  import '../../../internals/types.js';
@@ -1,12 +1,12 @@
1
1
  import { GetRunContext } from '../../../context.cjs';
2
- import { E as Emitter } from '../../../emitter-pJzHC_AM.cjs';
2
+ import { E as Emitter } from '../../../emitter-DpqUYjXH.cjs';
3
3
  import { EmbeddingModel, EmbeddingModelEvents, EmbeddingModelInput, EmbeddingModelOutput } from '../../../backend/embedding.cjs';
4
4
  import '../../../internals/serializable.cjs';
5
5
  import '../../../internals/types.cjs';
6
6
  import '../../../internals/helpers/guards.cjs';
7
7
  import '../../../internals/helpers/promise.cjs';
8
8
  import '../../../errors.cjs';
9
- import '../../../chat-BZ55YQab.cjs';
9
+ import '../../../chat-CRb3vUVg.cjs';
10
10
  import '../../../backend/message.cjs';
11
11
  import 'ai';
12
12
  import 'promise-based-task';
@@ -1,12 +1,12 @@
1
1
  import { GetRunContext } from '../../../context.js';
2
- import { E as Emitter } from '../../../emitter-BqpLJQVb.js';
2
+ import { E as Emitter } from '../../../emitter-CZFbzlUi.js';
3
3
  import { EmbeddingModel, EmbeddingModelEvents, EmbeddingModelInput, EmbeddingModelOutput } from '../../../backend/embedding.js';
4
4
  import '../../../internals/serializable.js';
5
5
  import '../../../internals/types.js';
6
6
  import '../../../internals/helpers/guards.js';
7
7
  import '../../../internals/helpers/promise.js';
8
8
  import '../../../errors.js';
9
- import '../../../chat-BBoOSvzm.js';
9
+ import '../../../chat-C0s-o6ll.js';
10
10
  import '../../../backend/message.js';
11
11
  import 'ai';
12
12
  import 'promise-based-task';
@@ -10,9 +10,10 @@ class GoogleVertexChatModel extends chat_cjs.VercelChatModel {
10
10
  static {
11
11
  __name(this, "GoogleVertexChatModel");
12
12
  }
13
- constructor(modelId = env_cjs.getEnv("GOOGLE_VERTEX_CHAT_MODEL", "gemini-1.5-pro"), settings = {}, client) {
14
- const model = client_cjs.GoogleVertexClient.ensure(client).instance.languageModel(modelId, settings);
13
+ constructor(modelId = env_cjs.getEnv("GOOGLE_VERTEX_CHAT_MODEL", "gemini-1.5-pro"), parameters = {}, client) {
14
+ const model = client_cjs.GoogleVertexClient.ensure(client).instance.languageModel(modelId);
15
15
  super(model);
16
+ Object.assign(this.parameters, parameters ?? {});
16
17
  }
17
18
  }
18
19
 
@@ -1 +1 @@
1
- {"version":3,"sources":["../../../../src/adapters/google-vertex/backend/chat.ts"],"names":["GoogleVertexChatModel","VercelChatModel","modelId","getEnv","settings","client","model","GoogleVertexClient","ensure","instance","languageModel"],"mappings":";;;;;;;;AAiBO,MAAMA,8BAA8BC,wBAAAA,CAAAA;EAjB3C;;;EAkBE,WAAA,CACEC,OAAAA,GAAmCC,eAAO,0BAAA,EAA4B,gBAAA,GACtEC,QAAAA,GAA0C,IAC1CC,MAAAA,EACA;AACA,IAAA,MAAMC,KAAAA,GAAQC,8BAAmBC,MAAAA,CAAOH,MAAAA,EAAQI,QAAAA,CAASC,aAAAA,CAAcR,SAASE,QAAAA,CAAAA;AAChF,IAAA,KAAA,CAAME,KAAAA,CAAAA;AACR;AACF","file":"chat.cjs","sourcesContent":["/**\n * Copyright 2025 © BeeAI a Series of LF Projects, LLC\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport { GoogleVertexProvider } from \"@ai-sdk/google-vertex\";\nimport { VercelChatModel } from \"@/adapters/vercel/backend/chat.js\";\nimport {\n GoogleVertexClient,\n GoogleVertexClientSettings,\n} from \"@/adapters/google-vertex/backend/client.js\";\nimport { getEnv } from \"@/internals/env.js\";\n\ntype GoogleVertexParameters = Parameters<GoogleVertexProvider[\"languageModel\"]>;\nexport type GoogleVertexChatModelId = NonNullable<GoogleVertexParameters[0]>;\nexport type GoogleVertexChatModelSettings = NonNullable<GoogleVertexParameters[1]>;\n\nexport class GoogleVertexChatModel extends VercelChatModel {\n constructor(\n modelId: GoogleVertexChatModelId = getEnv(\"GOOGLE_VERTEX_CHAT_MODEL\", \"gemini-1.5-pro\"),\n settings: GoogleVertexChatModelSettings = {},\n client?: GoogleVertexClientSettings | GoogleVertexClient,\n ) {\n const model = GoogleVertexClient.ensure(client).instance.languageModel(modelId, settings);\n super(model);\n }\n}\n"]}
1
+ {"version":3,"sources":["../../../../src/adapters/google-vertex/backend/chat.ts"],"names":["GoogleVertexChatModel","VercelChatModel","modelId","getEnv","parameters","client","model","GoogleVertexClient","ensure","instance","languageModel","Object","assign"],"mappings":";;;;;;;;AAiBO,MAAMA,8BAA8BC,wBAAAA,CAAAA;EAjB3C;;;EAkBE,WAAA,CACEC,OAAAA,GAAmCC,eAAO,0BAAA,EAA4B,gBAAA,GACtEC,UAAAA,GAAkC,IAClCC,MAAAA,EACA;AACA,IAAA,MAAMC,QAAQC,6BAAAA,CAAmBC,MAAAA,CAAOH,MAAAA,CAAAA,CAAQI,QAAAA,CAASC,cAAcR,OAAAA,CAAAA;AACvE,IAAA,KAAA,CAAMI,KAAAA,CAAAA;AACNK,IAAAA,MAAAA,CAAOC,MAAAA,CAAO,IAAA,CAAKR,UAAAA,EAAYA,UAAAA,IAAc,EAAC,CAAA;AAChD;AACF","file":"chat.cjs","sourcesContent":["/**\n * Copyright 2025 © BeeAI a Series of LF Projects, LLC\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport { GoogleVertexProvider } from \"@ai-sdk/google-vertex\";\nimport { VercelChatModel } from \"@/adapters/vercel/backend/chat.js\";\nimport {\n GoogleVertexClient,\n GoogleVertexClientSettings,\n} from \"@/adapters/google-vertex/backend/client.js\";\nimport { getEnv } from \"@/internals/env.js\";\nimport { ChatModelParameters } from \"@/backend/chat.js\";\n\ntype GoogleVertexParameters = Parameters<GoogleVertexProvider[\"languageModel\"]>;\nexport type GoogleVertexChatModelId = NonNullable<GoogleVertexParameters[0]>;\n\nexport class GoogleVertexChatModel extends VercelChatModel {\n constructor(\n modelId: GoogleVertexChatModelId = getEnv(\"GOOGLE_VERTEX_CHAT_MODEL\", \"gemini-1.5-pro\"),\n parameters: ChatModelParameters = {},\n client?: GoogleVertexClientSettings | GoogleVertexClient,\n ) {\n const model = GoogleVertexClient.ensure(client).instance.languageModel(modelId);\n super(model);\n Object.assign(this.parameters, parameters ?? {});\n }\n}\n"]}
@@ -1,17 +1,17 @@
1
1
  import { GoogleVertexProvider } from '@ai-sdk/google-vertex';
2
2
  import { VercelChatModel } from '../../vercel/backend/chat.cjs';
3
3
  import { GoogleVertexClientSettings, GoogleVertexClient } from './client.cjs';
4
+ import { C as ChatModelParameters } from '../../../chat-CRb3vUVg.cjs';
4
5
  import '../../../logger/logger.cjs';
5
6
  import 'pino';
6
7
  import '../../../errors.cjs';
7
8
  import '../../../internals/types.cjs';
8
9
  import '../../../internals/helpers/guards.cjs';
9
10
  import '../../../internals/serializable.cjs';
10
- import '../../../chat-BZ55YQab.cjs';
11
- import '../../../backend/message.cjs';
12
11
  import 'ai';
12
+ import '../../../emitter-DpqUYjXH.cjs';
13
+ import '../../../backend/message.cjs';
13
14
  import '../../../context.cjs';
14
- import '../../../emitter-pJzHC_AM.cjs';
15
15
  import '../../../internals/helpers/promise.cjs';
16
16
  import 'promise-based-task';
17
17
  import '../../../cache/base.cjs';
@@ -31,9 +31,8 @@ import '../../../backend/client.cjs';
31
31
 
32
32
  type GoogleVertexParameters = Parameters<GoogleVertexProvider["languageModel"]>;
33
33
  type GoogleVertexChatModelId = NonNullable<GoogleVertexParameters[0]>;
34
- type GoogleVertexChatModelSettings = NonNullable<GoogleVertexParameters[1]>;
35
34
  declare class GoogleVertexChatModel extends VercelChatModel {
36
- constructor(modelId?: GoogleVertexChatModelId, settings?: GoogleVertexChatModelSettings, client?: GoogleVertexClientSettings | GoogleVertexClient);
35
+ constructor(modelId?: GoogleVertexChatModelId, parameters?: ChatModelParameters, client?: GoogleVertexClientSettings | GoogleVertexClient);
37
36
  }
38
37
 
39
- export { GoogleVertexChatModel, type GoogleVertexChatModelId, type GoogleVertexChatModelSettings };
38
+ export { GoogleVertexChatModel, type GoogleVertexChatModelId };
@@ -1,17 +1,17 @@
1
1
  import { GoogleVertexProvider } from '@ai-sdk/google-vertex';
2
2
  import { VercelChatModel } from '../../vercel/backend/chat.js';
3
3
  import { GoogleVertexClientSettings, GoogleVertexClient } from './client.js';
4
+ import { C as ChatModelParameters } from '../../../chat-C0s-o6ll.js';
4
5
  import '../../../logger/logger.js';
5
6
  import 'pino';
6
7
  import '../../../errors.js';
7
8
  import '../../../internals/types.js';
8
9
  import '../../../internals/helpers/guards.js';
9
10
  import '../../../internals/serializable.js';
10
- import '../../../chat-BBoOSvzm.js';
11
- import '../../../backend/message.js';
12
11
  import 'ai';
12
+ import '../../../emitter-CZFbzlUi.js';
13
+ import '../../../backend/message.js';
13
14
  import '../../../context.js';
14
- import '../../../emitter-BqpLJQVb.js';
15
15
  import '../../../internals/helpers/promise.js';
16
16
  import 'promise-based-task';
17
17
  import '../../../cache/base.js';
@@ -31,9 +31,8 @@ import '../../../backend/client.js';
31
31
 
32
32
  type GoogleVertexParameters = Parameters<GoogleVertexProvider["languageModel"]>;
33
33
  type GoogleVertexChatModelId = NonNullable<GoogleVertexParameters[0]>;
34
- type GoogleVertexChatModelSettings = NonNullable<GoogleVertexParameters[1]>;
35
34
  declare class GoogleVertexChatModel extends VercelChatModel {
36
- constructor(modelId?: GoogleVertexChatModelId, settings?: GoogleVertexChatModelSettings, client?: GoogleVertexClientSettings | GoogleVertexClient);
35
+ constructor(modelId?: GoogleVertexChatModelId, parameters?: ChatModelParameters, client?: GoogleVertexClientSettings | GoogleVertexClient);
37
36
  }
38
37
 
39
- export { GoogleVertexChatModel, type GoogleVertexChatModelId, type GoogleVertexChatModelSettings };
38
+ export { GoogleVertexChatModel, type GoogleVertexChatModelId };
@@ -8,9 +8,10 @@ class GoogleVertexChatModel extends VercelChatModel {
8
8
  static {
9
9
  __name(this, "GoogleVertexChatModel");
10
10
  }
11
- constructor(modelId = getEnv("GOOGLE_VERTEX_CHAT_MODEL", "gemini-1.5-pro"), settings = {}, client) {
12
- const model = GoogleVertexClient.ensure(client).instance.languageModel(modelId, settings);
11
+ constructor(modelId = getEnv("GOOGLE_VERTEX_CHAT_MODEL", "gemini-1.5-pro"), parameters = {}, client) {
12
+ const model = GoogleVertexClient.ensure(client).instance.languageModel(modelId);
13
13
  super(model);
14
+ Object.assign(this.parameters, parameters ?? {});
14
15
  }
15
16
  }
16
17
 
@@ -1 +1 @@
1
- {"version":3,"sources":["../../../../src/adapters/google-vertex/backend/chat.ts"],"names":["GoogleVertexChatModel","VercelChatModel","modelId","getEnv","settings","client","model","GoogleVertexClient","ensure","instance","languageModel"],"mappings":";;;;;;AAiBO,MAAMA,8BAA8BC,eAAAA,CAAAA;EAjB3C;;;EAkBE,WAAA,CACEC,OAAAA,GAAmCC,OAAO,0BAAA,EAA4B,gBAAA,GACtEC,QAAAA,GAA0C,IAC1CC,MAAAA,EACA;AACA,IAAA,MAAMC,KAAAA,GAAQC,mBAAmBC,MAAAA,CAAOH,MAAAA,EAAQI,QAAAA,CAASC,aAAAA,CAAcR,SAASE,QAAAA,CAAAA;AAChF,IAAA,KAAA,CAAME,KAAAA,CAAAA;AACR;AACF","file":"chat.js","sourcesContent":["/**\n * Copyright 2025 © BeeAI a Series of LF Projects, LLC\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport { GoogleVertexProvider } from \"@ai-sdk/google-vertex\";\nimport { VercelChatModel } from \"@/adapters/vercel/backend/chat.js\";\nimport {\n GoogleVertexClient,\n GoogleVertexClientSettings,\n} from \"@/adapters/google-vertex/backend/client.js\";\nimport { getEnv } from \"@/internals/env.js\";\n\ntype GoogleVertexParameters = Parameters<GoogleVertexProvider[\"languageModel\"]>;\nexport type GoogleVertexChatModelId = NonNullable<GoogleVertexParameters[0]>;\nexport type GoogleVertexChatModelSettings = NonNullable<GoogleVertexParameters[1]>;\n\nexport class GoogleVertexChatModel extends VercelChatModel {\n constructor(\n modelId: GoogleVertexChatModelId = getEnv(\"GOOGLE_VERTEX_CHAT_MODEL\", \"gemini-1.5-pro\"),\n settings: GoogleVertexChatModelSettings = {},\n client?: GoogleVertexClientSettings | GoogleVertexClient,\n ) {\n const model = GoogleVertexClient.ensure(client).instance.languageModel(modelId, settings);\n super(model);\n }\n}\n"]}
1
+ {"version":3,"sources":["../../../../src/adapters/google-vertex/backend/chat.ts"],"names":["GoogleVertexChatModel","VercelChatModel","modelId","getEnv","parameters","client","model","GoogleVertexClient","ensure","instance","languageModel","Object","assign"],"mappings":";;;;;;AAiBO,MAAMA,8BAA8BC,eAAAA,CAAAA;EAjB3C;;;EAkBE,WAAA,CACEC,OAAAA,GAAmCC,OAAO,0BAAA,EAA4B,gBAAA,GACtEC,UAAAA,GAAkC,IAClCC,MAAAA,EACA;AACA,IAAA,MAAMC,QAAQC,kBAAAA,CAAmBC,MAAAA,CAAOH,MAAAA,CAAAA,CAAQI,QAAAA,CAASC,cAAcR,OAAAA,CAAAA;AACvE,IAAA,KAAA,CAAMI,KAAAA,CAAAA;AACNK,IAAAA,MAAAA,CAAOC,MAAAA,CAAO,IAAA,CAAKR,UAAAA,EAAYA,UAAAA,IAAc,EAAC,CAAA;AAChD;AACF","file":"chat.js","sourcesContent":["/**\n * Copyright 2025 © BeeAI a Series of LF Projects, LLC\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport { GoogleVertexProvider } from \"@ai-sdk/google-vertex\";\nimport { VercelChatModel } from \"@/adapters/vercel/backend/chat.js\";\nimport {\n GoogleVertexClient,\n GoogleVertexClientSettings,\n} from \"@/adapters/google-vertex/backend/client.js\";\nimport { getEnv } from \"@/internals/env.js\";\nimport { ChatModelParameters } from \"@/backend/chat.js\";\n\ntype GoogleVertexParameters = Parameters<GoogleVertexProvider[\"languageModel\"]>;\nexport type GoogleVertexChatModelId = NonNullable<GoogleVertexParameters[0]>;\n\nexport class GoogleVertexChatModel extends VercelChatModel {\n constructor(\n modelId: GoogleVertexChatModelId = getEnv(\"GOOGLE_VERTEX_CHAT_MODEL\", \"gemini-1.5-pro\"),\n parameters: ChatModelParameters = {},\n client?: GoogleVertexClientSettings | GoogleVertexClient,\n ) {\n const model = GoogleVertexClient.ensure(client).instance.languageModel(modelId);\n super(model);\n Object.assign(this.parameters, parameters ?? {});\n }\n}\n"]}
@@ -7,10 +7,10 @@ import '../../../internals/types.cjs';
7
7
  import '../../../internals/helpers/guards.cjs';
8
8
  import '../../../backend/embedding.cjs';
9
9
  import '../../../context.cjs';
10
- import '../../../emitter-pJzHC_AM.cjs';
10
+ import '../../../emitter-DpqUYjXH.cjs';
11
11
  import '../../../internals/helpers/promise.cjs';
12
12
  import '../../../errors.cjs';
13
- import '../../../chat-BZ55YQab.cjs';
13
+ import '../../../chat-CRb3vUVg.cjs';
14
14
  import '../../../backend/message.cjs';
15
15
  import 'ai';
16
16
  import 'promise-based-task';
@@ -7,10 +7,10 @@ import '../../../internals/types.js';
7
7
  import '../../../internals/helpers/guards.js';
8
8
  import '../../../backend/embedding.js';
9
9
  import '../../../context.js';
10
- import '../../../emitter-BqpLJQVb.js';
10
+ import '../../../emitter-CZFbzlUi.js';
11
11
  import '../../../internals/helpers/promise.js';
12
12
  import '../../../errors.js';
13
- import '../../../chat-BBoOSvzm.js';
13
+ import '../../../chat-C0s-o6ll.js';
14
14
  import '../../../backend/message.js';
15
15
  import 'ai';
16
16
  import 'promise-based-task';
@@ -10,9 +10,10 @@ class GroqChatModel extends chat_cjs.VercelChatModel {
10
10
  static {
11
11
  __name(this, "GroqChatModel");
12
12
  }
13
- constructor(modelId = env_cjs.getEnv("GROQ_CHAT_MODEL", "gemma2-9b-it"), settings = {}, client) {
14
- const model = client_cjs.GroqClient.ensure(client).instance.languageModel(modelId, settings);
13
+ constructor(modelId = env_cjs.getEnv("GROQ_CHAT_MODEL", "gemma2-9b-it"), parameters = {}, client) {
14
+ const model = client_cjs.GroqClient.ensure(client).instance.languageModel(modelId);
15
15
  super(model);
16
+ Object.assign(this.parameters, parameters ?? {});
16
17
  }
17
18
  static {
18
19
  this.register();
@@ -1 +1 @@
1
- {"version":3,"sources":["../../../../src/adapters/groq/backend/chat.ts"],"names":["GroqChatModel","VercelChatModel","modelId","getEnv","settings","client","model","GroqClient","ensure","instance","languageModel","register"],"mappings":";;;;;;;;AAcO,MAAMA,sBAAsBC,wBAAAA,CAAAA;EAdnC;;;EAeE,WAAA,CACEC,OAAAA,GAA2BC,eAAO,iBAAA,EAAmB,cAAA,GACrDC,QAAAA,GAAkC,IAClCC,MAAAA,EACA;AACA,IAAA,MAAMC,KAAAA,GAAQC,sBAAWC,MAAAA,CAAOH,MAAAA,EAAQI,QAAAA,CAASC,aAAAA,CAAcR,SAASE,QAAAA,CAAAA;AACxE,IAAA,KAAA,CAAME,KAAAA,CAAAA;AACR;EAEA;AACE,IAAA,IAAA,CAAKK,QAAAA,EAAQ;AACf;AACF","file":"chat.cjs","sourcesContent":["/**\n * Copyright 2025 © BeeAI a Series of LF Projects, LLC\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport { VercelChatModel } from \"@/adapters/vercel/backend/chat.js\";\nimport { GroqClient, GroqClientSettings } from \"@/adapters/groq/backend/client.js\";\nimport { getEnv } from \"@/internals/env.js\";\nimport { GroqProvider } from \"@ai-sdk/groq\";\n\ntype GroqParameters = Parameters<GroqProvider[\"languageModel\"]>;\nexport type GroqChatModelId = NonNullable<GroqParameters[0]>;\nexport type GroqChatModelSettings = NonNullable<GroqParameters[1]>;\n\nexport class GroqChatModel extends VercelChatModel {\n constructor(\n modelId: GroqChatModelId = getEnv(\"GROQ_CHAT_MODEL\", \"gemma2-9b-it\"),\n settings: GroqChatModelSettings = {},\n client?: GroqClientSettings | GroqClient,\n ) {\n const model = GroqClient.ensure(client).instance.languageModel(modelId, settings);\n super(model);\n }\n\n static {\n this.register();\n }\n}\n"]}
1
+ {"version":3,"sources":["../../../../src/adapters/groq/backend/chat.ts"],"names":["GroqChatModel","VercelChatModel","modelId","getEnv","parameters","client","model","GroqClient","ensure","instance","languageModel","Object","assign","register"],"mappings":";;;;;;;;AAcO,MAAMA,sBAAsBC,wBAAAA,CAAAA;EAdnC;;;EAeE,WAAA,CACEC,OAAAA,GAA2BC,eAAO,iBAAA,EAAmB,cAAA,GACrDC,UAAAA,GAAkC,IAClCC,MAAAA,EACA;AACA,IAAA,MAAMC,QAAQC,qBAAAA,CAAWC,MAAAA,CAAOH,MAAAA,CAAAA,CAAQI,QAAAA,CAASC,cAAcR,OAAAA,CAAAA;AAC/D,IAAA,KAAA,CAAMI,KAAAA,CAAAA;AACNK,IAAAA,MAAAA,CAAOC,MAAAA,CAAO,IAAA,CAAKR,UAAAA,EAAYA,UAAAA,IAAc,EAAC,CAAA;AAChD;EAEA;AACE,IAAA,IAAA,CAAKS,QAAAA,EAAQ;AACf;AACF","file":"chat.cjs","sourcesContent":["/**\n * Copyright 2025 © BeeAI a Series of LF Projects, LLC\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport { VercelChatModel } from \"@/adapters/vercel/backend/chat.js\";\nimport { GroqClient, GroqClientSettings } from \"@/adapters/groq/backend/client.js\";\nimport { getEnv } from \"@/internals/env.js\";\nimport { GroqProvider } from \"@ai-sdk/groq\";\nimport { ChatModelParameters } from \"@/backend/chat.js\";\n\ntype GroqParameters = Parameters<GroqProvider[\"languageModel\"]>;\nexport type GroqChatModelId = NonNullable<GroqParameters[0]>;\n\nexport class GroqChatModel extends VercelChatModel {\n constructor(\n modelId: GroqChatModelId = getEnv(\"GROQ_CHAT_MODEL\", \"gemma2-9b-it\"),\n parameters: ChatModelParameters = {},\n client?: GroqClientSettings | GroqClient,\n ) {\n const model = GroqClient.ensure(client).instance.languageModel(modelId);\n super(model);\n Object.assign(this.parameters, parameters ?? {});\n }\n\n static {\n this.register();\n }\n}\n"]}
@@ -1,17 +1,17 @@
1
1
  import { VercelChatModel } from '../../vercel/backend/chat.cjs';
2
2
  import { GroqClientSettings, GroqClient } from './client.cjs';
3
3
  import { GroqProvider } from '@ai-sdk/groq';
4
+ import { C as ChatModelParameters } from '../../../chat-CRb3vUVg.cjs';
4
5
  import '../../../logger/logger.cjs';
5
6
  import 'pino';
6
7
  import '../../../errors.cjs';
7
8
  import '../../../internals/types.cjs';
8
9
  import '../../../internals/helpers/guards.cjs';
9
10
  import '../../../internals/serializable.cjs';
10
- import '../../../chat-BZ55YQab.cjs';
11
- import '../../../backend/message.cjs';
12
11
  import 'ai';
12
+ import '../../../emitter-DpqUYjXH.cjs';
13
+ import '../../../backend/message.cjs';
13
14
  import '../../../context.cjs';
14
- import '../../../emitter-pJzHC_AM.cjs';
15
15
  import '../../../internals/helpers/promise.cjs';
16
16
  import 'promise-based-task';
17
17
  import '../../../cache/base.cjs';
@@ -31,9 +31,8 @@ import '../../../backend/client.cjs';
31
31
 
32
32
  type GroqParameters = Parameters<GroqProvider["languageModel"]>;
33
33
  type GroqChatModelId = NonNullable<GroqParameters[0]>;
34
- type GroqChatModelSettings = NonNullable<GroqParameters[1]>;
35
34
  declare class GroqChatModel extends VercelChatModel {
36
- constructor(modelId?: GroqChatModelId, settings?: GroqChatModelSettings, client?: GroqClientSettings | GroqClient);
35
+ constructor(modelId?: GroqChatModelId, parameters?: ChatModelParameters, client?: GroqClientSettings | GroqClient);
37
36
  }
38
37
 
39
- export { GroqChatModel, type GroqChatModelId, type GroqChatModelSettings };
38
+ export { GroqChatModel, type GroqChatModelId };
@@ -1,17 +1,17 @@
1
1
  import { VercelChatModel } from '../../vercel/backend/chat.js';
2
2
  import { GroqClientSettings, GroqClient } from './client.js';
3
3
  import { GroqProvider } from '@ai-sdk/groq';
4
+ import { C as ChatModelParameters } from '../../../chat-C0s-o6ll.js';
4
5
  import '../../../logger/logger.js';
5
6
  import 'pino';
6
7
  import '../../../errors.js';
7
8
  import '../../../internals/types.js';
8
9
  import '../../../internals/helpers/guards.js';
9
10
  import '../../../internals/serializable.js';
10
- import '../../../chat-BBoOSvzm.js';
11
- import '../../../backend/message.js';
12
11
  import 'ai';
12
+ import '../../../emitter-CZFbzlUi.js';
13
+ import '../../../backend/message.js';
13
14
  import '../../../context.js';
14
- import '../../../emitter-BqpLJQVb.js';
15
15
  import '../../../internals/helpers/promise.js';
16
16
  import 'promise-based-task';
17
17
  import '../../../cache/base.js';
@@ -31,9 +31,8 @@ import '../../../backend/client.js';
31
31
 
32
32
  type GroqParameters = Parameters<GroqProvider["languageModel"]>;
33
33
  type GroqChatModelId = NonNullable<GroqParameters[0]>;
34
- type GroqChatModelSettings = NonNullable<GroqParameters[1]>;
35
34
  declare class GroqChatModel extends VercelChatModel {
36
- constructor(modelId?: GroqChatModelId, settings?: GroqChatModelSettings, client?: GroqClientSettings | GroqClient);
35
+ constructor(modelId?: GroqChatModelId, parameters?: ChatModelParameters, client?: GroqClientSettings | GroqClient);
37
36
  }
38
37
 
39
- export { GroqChatModel, type GroqChatModelId, type GroqChatModelSettings };
38
+ export { GroqChatModel, type GroqChatModelId };
@@ -8,9 +8,10 @@ class GroqChatModel extends VercelChatModel {
8
8
  static {
9
9
  __name(this, "GroqChatModel");
10
10
  }
11
- constructor(modelId = getEnv("GROQ_CHAT_MODEL", "gemma2-9b-it"), settings = {}, client) {
12
- const model = GroqClient.ensure(client).instance.languageModel(modelId, settings);
11
+ constructor(modelId = getEnv("GROQ_CHAT_MODEL", "gemma2-9b-it"), parameters = {}, client) {
12
+ const model = GroqClient.ensure(client).instance.languageModel(modelId);
13
13
  super(model);
14
+ Object.assign(this.parameters, parameters ?? {});
14
15
  }
15
16
  static {
16
17
  this.register();
@@ -1 +1 @@
1
- {"version":3,"sources":["../../../../src/adapters/groq/backend/chat.ts"],"names":["GroqChatModel","VercelChatModel","modelId","getEnv","settings","client","model","GroqClient","ensure","instance","languageModel","register"],"mappings":";;;;;;AAcO,MAAMA,sBAAsBC,eAAAA,CAAAA;EAdnC;;;EAeE,WAAA,CACEC,OAAAA,GAA2BC,OAAO,iBAAA,EAAmB,cAAA,GACrDC,QAAAA,GAAkC,IAClCC,MAAAA,EACA;AACA,IAAA,MAAMC,KAAAA,GAAQC,WAAWC,MAAAA,CAAOH,MAAAA,EAAQI,QAAAA,CAASC,aAAAA,CAAcR,SAASE,QAAAA,CAAAA;AACxE,IAAA,KAAA,CAAME,KAAAA,CAAAA;AACR;EAEA;AACE,IAAA,IAAA,CAAKK,QAAAA,EAAQ;AACf;AACF","file":"chat.js","sourcesContent":["/**\n * Copyright 2025 © BeeAI a Series of LF Projects, LLC\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport { VercelChatModel } from \"@/adapters/vercel/backend/chat.js\";\nimport { GroqClient, GroqClientSettings } from \"@/adapters/groq/backend/client.js\";\nimport { getEnv } from \"@/internals/env.js\";\nimport { GroqProvider } from \"@ai-sdk/groq\";\n\ntype GroqParameters = Parameters<GroqProvider[\"languageModel\"]>;\nexport type GroqChatModelId = NonNullable<GroqParameters[0]>;\nexport type GroqChatModelSettings = NonNullable<GroqParameters[1]>;\n\nexport class GroqChatModel extends VercelChatModel {\n constructor(\n modelId: GroqChatModelId = getEnv(\"GROQ_CHAT_MODEL\", \"gemma2-9b-it\"),\n settings: GroqChatModelSettings = {},\n client?: GroqClientSettings | GroqClient,\n ) {\n const model = GroqClient.ensure(client).instance.languageModel(modelId, settings);\n super(model);\n }\n\n static {\n this.register();\n }\n}\n"]}
1
+ {"version":3,"sources":["../../../../src/adapters/groq/backend/chat.ts"],"names":["GroqChatModel","VercelChatModel","modelId","getEnv","parameters","client","model","GroqClient","ensure","instance","languageModel","Object","assign","register"],"mappings":";;;;;;AAcO,MAAMA,sBAAsBC,eAAAA,CAAAA;EAdnC;;;EAeE,WAAA,CACEC,OAAAA,GAA2BC,OAAO,iBAAA,EAAmB,cAAA,GACrDC,UAAAA,GAAkC,IAClCC,MAAAA,EACA;AACA,IAAA,MAAMC,QAAQC,UAAAA,CAAWC,MAAAA,CAAOH,MAAAA,CAAAA,CAAQI,QAAAA,CAASC,cAAcR,OAAAA,CAAAA;AAC/D,IAAA,KAAA,CAAMI,KAAAA,CAAAA;AACNK,IAAAA,MAAAA,CAAOC,MAAAA,CAAO,IAAA,CAAKR,UAAAA,EAAYA,UAAAA,IAAc,EAAC,CAAA;AAChD;EAEA;AACE,IAAA,IAAA,CAAKS,QAAAA,EAAQ;AACf;AACF","file":"chat.js","sourcesContent":["/**\n * Copyright 2025 © BeeAI a Series of LF Projects, LLC\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport { VercelChatModel } from \"@/adapters/vercel/backend/chat.js\";\nimport { GroqClient, GroqClientSettings } from \"@/adapters/groq/backend/client.js\";\nimport { getEnv } from \"@/internals/env.js\";\nimport { GroqProvider } from \"@ai-sdk/groq\";\nimport { ChatModelParameters } from \"@/backend/chat.js\";\n\ntype GroqParameters = Parameters<GroqProvider[\"languageModel\"]>;\nexport type GroqChatModelId = NonNullable<GroqParameters[0]>;\n\nexport class GroqChatModel extends VercelChatModel {\n constructor(\n modelId: GroqChatModelId = getEnv(\"GROQ_CHAT_MODEL\", \"gemma2-9b-it\"),\n parameters: ChatModelParameters = {},\n client?: GroqClientSettings | GroqClient,\n ) {\n const model = GroqClient.ensure(client).instance.languageModel(modelId);\n super(model);\n Object.assign(this.parameters, parameters ?? {});\n }\n\n static {\n this.register();\n }\n}\n"]}
@@ -3,13 +3,13 @@ import { GroqClientSettings, GroqClient } from './client.cjs';
3
3
  import { GroqProvider } from '@ai-sdk/groq';
4
4
  import '../../../backend/embedding.cjs';
5
5
  import '../../../context.cjs';
6
- import '../../../emitter-pJzHC_AM.cjs';
6
+ import '../../../emitter-DpqUYjXH.cjs';
7
7
  import '../../../internals/types.cjs';
8
8
  import '../../../internals/helpers/guards.cjs';
9
9
  import '../../../internals/serializable.cjs';
10
10
  import '../../../internals/helpers/promise.cjs';
11
11
  import '../../../errors.cjs';
12
- import '../../../chat-BZ55YQab.cjs';
12
+ import '../../../chat-CRb3vUVg.cjs';
13
13
  import '../../../backend/message.cjs';
14
14
  import 'ai';
15
15
  import 'promise-based-task';
@@ -3,13 +3,13 @@ import { GroqClientSettings, GroqClient } from './client.js';
3
3
  import { GroqProvider } from '@ai-sdk/groq';
4
4
  import '../../../backend/embedding.js';
5
5
  import '../../../context.js';
6
- import '../../../emitter-BqpLJQVb.js';
6
+ import '../../../emitter-CZFbzlUi.js';
7
7
  import '../../../internals/types.js';
8
8
  import '../../../internals/helpers/guards.js';
9
9
  import '../../../internals/serializable.js';
10
10
  import '../../../internals/helpers/promise.js';
11
11
  import '../../../errors.js';
12
- import '../../../chat-BBoOSvzm.js';
12
+ import '../../../chat-C0s-o6ll.js';
13
13
  import '../../../backend/message.js';
14
14
  import 'ai';
15
15
  import 'promise-based-task';
@@ -82,13 +82,14 @@ class LangChainChatModel extends chat_cjs.ChatModel {
82
82
  }
83
83
  })));
84
84
  }
85
- const usage = {
85
+ const stop = output.response_metadata.stop_sequence || "stop";
86
+ return new chat_cjs.ChatModelOutput(messages, {
86
87
  totalTokens: output.usage_metadata?.total_tokens ?? 0,
87
88
  promptTokens: output.usage_metadata?.input_tokens ?? 0,
88
- completionTokens: output.usage_metadata?.output_tokens ?? 0
89
- };
90
- const stop = output.response_metadata.stop_sequence || "stop";
91
- return new chat_cjs.ChatModelOutput(messages, usage, stop);
89
+ completionTokens: output.usage_metadata?.output_tokens ?? 0,
90
+ reasoningTokens: output.usage_metadata?.output_token_details?.reasoning,
91
+ cachedPromptTokens: output.usage_metadata?.input_token_details?.cache_read
92
+ }, stop);
92
93
  }
93
94
  async _createStructure(input, run) {
94
95
  const { messages, options } = this.prepareInput(input, run);
@@ -1 +1 @@
1
- {"version":3,"sources":["../../../../src/adapters/langchain/backend/chat.ts"],"names":["LangChainChatModel","ChatModel","emitter","lcLLM","parameters","Emitter","root","child","namespace","creator","modelId","_modelType","providerId","_create","input","run","preparedInput","prepareInput","response","bindTools","tools","invoke","messages","options","prepareOutput","_createStream","stream","chunk","map","msg","role","content","type","runId","stop","stopSequences","signal","tool_choice","toolChoice","output","push","AssistantMessage","message","text","image_url","toString","ValueError","usage","totalTokens","usage_metadata","total_tokens","promptTokens","input_tokens","completionTokens","output_tokens","response_metadata","stop_sequence","ChatModelOutput","_createStructure","raw","parsed","withStructuredOutput","schema","method","strict","includeRaw","object","createSnapshot","loadSnapshot","snapshot","Object","assign"],"mappings":";;;;;;;;;AA0BO,MAAMA,2BAA2BC,kBAAAA,CAAAA;EA1BxC;;;;;AA2BkBC,EAAAA,OAAAA;EAEhB,WAAA,CACqBC,KAAAA,EACHC,UAAAA,GAAkC,EAAC,EACnD;AACA,IAAA,KAAA,EAAK,EAAA,IAAA,CAHcD,KAAAA,GAAAA,KAAAA,EAAAA,KACHC,UAAAA,GAAAA,UAAAA;AAGhB,IAAA,IAAA,CAAKF,OAAAA,GAAUG,mBAAAA,CAAQC,IAAAA,CAAKC,KAAAA,CAAM;MAChCC,SAAAA,EAAW;AAAC,QAAA,SAAA;AAAW,QAAA,WAAA;AAAa,QAAA;;MACpCC,OAAAA,EAAS;KACX,CAAA;AACF;AAEA,EAAA,IAAIC,OAAAA,GAAkB;AACpB,IAAA,OAAO,IAAA,CAAKP,MAAMQ,UAAAA,EAAU;AAC9B;AAEA,EAAA,IAAIC,UAAAA,GAAa;AACf,IAAA,OAAO,WAAA;AACT;EAEA,MAAgBC,OAAAA,CAAQC,OAAuBC,GAAAA,EAAiD;AAC9F,IAAA,MAAMC,aAAAA,GAAgB,IAAA,CAAKC,YAAAA,CAAaH,KAAAA,EAAOC,GAAAA,CAAAA;AAC/C,IAAA,MAAMG,QAAAA,GAAW,IAAA,CAAKf,KAAAA,CAAMgB,SAAAA,GACxB,MAAM,IAAA,CAAKhB,KAAAA,CACRgB,SAAAA,CAAUL,KAAAA,CAAMM,KAAAA,IAAS,EAAE,CAAA,CAC3BC,OAAOL,aAAAA,CAAcM,QAAAA,EAAUN,aAAAA,CAAcO,OAAO,CAAA,GACvD,MAAM,IAAA,CAAKpB,KAAAA,CAAMkB,MAAAA,CAAOL,aAAAA,CAAcM,QAAAA,EAAUN,aAAAA,CAAcO,OAAO,CAAA;AAEzE,IAAA,OAAO,IAAA,CAAKC,cAAcN,QAAAA,CAAAA;AAC5B;EAEA,OAAiBO,aAAAA,CACfX,OACAC,GAAAA,EACiC;AACjC,IAAA,MAAMC,aAAAA,GAAgB,IAAA,CAAKC,YAAAA,CAAaH,KAAAA,EAAOC,GAAAA,CAAAA;AAE/C,IAAA,MAAMW,MAAAA,GAAS,IAAA,CAAKvB,KAAAA,CAAMgB,SAAAA,GACtB,MAAM,IAAA,CAAKhB,KAAAA,CACRgB,SAAAA,CAAUL,KAAAA,CAAMM,KAAAA,IAAS,EAAE,CAAA,CAC3BM,OAAOV,aAAAA,CAAcM,QAAAA,EAAUN,aAAAA,CAAcO,OAAO,CAAA,GACvD,MAAM,IAAA,CAAKpB,KAAAA,CAAMuB,MAAAA,CAAOV,aAAAA,CAAcM,QAAAA,EAAUN,aAAAA,CAAcO,OAAO,CAAA;AAEzE,IAAA,WAAA,MAAiBL,YAAYQ,MAAAA,EAAQ;AACnC,MAAA,MAAMC,KAAAA,GAAQ,IAAA,CAAKH,aAAAA,CAAcN,QAAAA,CAAAA;AACjC,MAAA,MAAMS,KAAAA;AACR;AACF;AAEUV,EAAAA,YAAAA,CAAaH,OAAuBC,GAAAA,EAAuB;AACnE,IAAA,MAAMO,QAAAA,GAA8BR,KAAAA,CAAMQ,QAAAA,CAASM,GAAAA,CAAI,CAACC,GAAAA,MAAS;AAC/DC,MAAAA,IAAAA,EAAMD,GAAAA,CAAIC,IAAAA;AACVC,MAAAA,OAAAA,EAASF,GAAAA,CAAIE,OAAAA;AACbC,MAAAA,IAAAA,EAAMH,GAAAA,CAAIC;KAEZ,CAAA,CAAA;AAEA,IAAA,MAAMP,OAAAA,GAAoC;AACxCU,MAAAA,KAAAA,EAAOlB,GAAAA,CAAIkB,KAAAA;AACXC,MAAAA,IAAAA,EAAMpB,KAAAA,CAAMqB,aAAAA;AACZC,MAAAA,MAAAA,EAAQrB,GAAAA,CAAIqB,MAAAA;AACZC,MAAAA,WAAAA,EAAavB,KAAAA,CAAMwB;AACrB,KAAA;AAEA,IAAA,OAAO;AAAEhB,MAAAA,QAAAA;AAAUC,MAAAA;AAAQ,KAAA;AAC7B;AAEUC,EAAAA,aAAAA,CAAce,MAAAA,EAAwB;AAC9C,IAAA,MAAMjB,WAAsB,EAAA;AAC5B,IAAA,IAAI,OAAOiB,MAAAA,CAAOR,OAAAA,KAAY,QAAA,EAAU;AACtCT,MAAAA,QAAAA,CAASkB,IAAAA,CAAK,IAAIC,4BAAAA,CAAiBF,MAAAA,CAAOR,OAAO,CAAA,CAAA;KACnD,MAAO;AACLT,MAAAA,QAAAA,CAASkB,KACP,IAAIC,4BAAAA,CACFF,OAAOR,OAAAA,CAAQH,GAAAA,CAAI,CAACc,OAAAA,KAAAA;AAClB,QAAA,IAAIA,OAAAA,CAAQV,SAAS,MAAA,EAAQ;AAC3B,UAAA,OAAO;YAAEA,IAAAA,EAAM,MAAA;AAAQW,YAAAA,IAAAA,EAAMD,OAAAA,CAAQC;AAAK,WAAA;SAC5C,MAAA,IAAWD,OAAAA,CAAQV,SAAS,WAAA,EAAa;AACvC,UAAA,OAAO;YAAEA,IAAAA,EAAM,MAAA;YAAQW,IAAAA,EAAMD,OAAAA,CAAQE,UAAUC,QAAAA;AAAW,WAAA;SAC5D,MAAO;AACL,UAAA,MAAM,IAAIC,qBAAAA,CAAW,CAAA,sBAAA,EAAyBJ,OAAAA,CAAQV,IAAI,CAAA,CAAA,CAAG,CAAA;AAC/D;AACF,OAAA,CAAA,CAAA,CAAA;AAGN;AAEA,IAAA,MAAMe,KAAAA,GAAwB;MAC5BC,WAAAA,EAAaT,MAAAA,CAAOU,gBAAgBC,YAAAA,IAAgB,CAAA;MACpDC,YAAAA,EAAcZ,MAAAA,CAAOU,gBAAgBG,YAAAA,IAAgB,CAAA;MACrDC,gBAAAA,EAAkBd,MAAAA,CAAOU,gBAAgBK,aAAAA,IAAiB;AAC5D,KAAA;AAEA,IAAA,MAAMpB,IAAAA,GAA8BK,MAAAA,CAAOgB,iBAAAA,CAAkBC,aAAAA,IAAiB,MAAA;AAE9E,IAAA,OAAO,IAAIC,wBAAAA,CAAgBnC,QAAAA,EAAUyB,KAAAA,EAAOb,IAAAA,CAAAA;AAC9C;EAEA,MAAgBwB,gBAAAA,CACd5C,OACAC,GAAAA,EACmC;AACnC,IAAA,MAAM,EAAEO,QAAAA,EAAUC,OAAAA,KAAY,IAAA,CAAKN,YAAAA,CAAaH,OAAOC,GAAAA,CAAAA;AACvD,IAAA,MAAM,EAAE4C,KAAKC,MAAAA,EAAM,GAAK,MAAM,IAAA,CAAKzD,KAAAA,CAChC0D,oBAAAA,CAA0B/C,KAAAA,CAAMgD,MAAAA,EAAQ;MACvCC,MAAAA,EAAQ,YAAA;MACRC,MAAAA,EAAQ,KAAA;MACRC,UAAAA,EAAY;KACd,CAAA,CACC5C,MAAAA,CAAOC,QAAAA,EAAUC,OAAAA,CAAAA;AAEpB,IAAA,OAAO;MAAE2C,MAAAA,EAAQN,MAAAA;MAAarB,MAAAA,EAAQ,IAAA,CAAKf,cAAcmC,GAAAA;AAAuB,KAAA;AAClF;EAEAQ,cAAAA,GAAiB;AACf,IAAA,OAAO;AAAE,MAAA,GAAG,MAAMA,cAAAA,EAAAA;AAAkBjE,MAAAA,OAAAA,EAAS,IAAA,CAAKA,OAAAA;AAASC,MAAAA,KAAAA,EAAO,IAAA,CAAKA;AAAM,KAAA;AAC/E;AAEAiE,EAAAA,YAAAA,CAAaC,QAAAA,EAAwD;AACnEC,IAAAA,MAAAA,CAAOC,MAAAA,CAAO,MAAMF,QAAAA,CAAAA;AACtB;AACF","file":"chat.cjs","sourcesContent":["/**\n * Copyright 2025 © BeeAI a Series of LF Projects, LLC\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport {\n ChatModel,\n ChatModelEmitter,\n ChatModelFinishReason,\n ChatModelInput,\n ChatModelObjectInput,\n ChatModelObjectOutput,\n ChatModelOutput,\n ChatModelParameters,\n ChatModelUsage,\n} from \"@/backend/chat.js\";\nimport { RunContext } from \"@/context.js\";\nimport { Emitter } from \"@/emitter/emitter.js\";\nimport {\n BaseChatModel,\n BaseChatModelCallOptions,\n} from \"@langchain/core/language_models/chat_models\";\nimport { AIMessageChunk, BaseMessageLike } from \"@langchain/core/messages\";\nimport { AssistantMessage, Message } from \"@/backend/message.js\";\nimport { ValueError } from \"@/errors.js\";\n\nexport class LangChainChatModel extends ChatModel {\n public readonly emitter: ChatModelEmitter;\n\n constructor(\n protected readonly lcLLM: BaseChatModel,\n public readonly parameters: ChatModelParameters = {},\n ) {\n super();\n this.emitter = Emitter.root.child({\n namespace: [\"backend\", \"langchain\", \"chat\"],\n creator: this,\n });\n }\n\n get modelId(): string {\n return this.lcLLM._modelType();\n }\n\n get providerId() {\n return \"langchain\";\n }\n\n protected async _create(input: ChatModelInput, run: RunContext<this>): Promise<ChatModelOutput> {\n const preparedInput = this.prepareInput(input, run);\n const response = this.lcLLM.bindTools\n ? await this.lcLLM\n .bindTools(input.tools ?? [])\n .invoke(preparedInput.messages, preparedInput.options)\n : await this.lcLLM.invoke(preparedInput.messages, preparedInput.options);\n\n return this.prepareOutput(response);\n }\n\n protected async *_createStream(\n input: ChatModelInput,\n run: RunContext<this>,\n ): AsyncGenerator<ChatModelOutput> {\n const preparedInput = this.prepareInput(input, run);\n\n const stream = this.lcLLM.bindTools\n ? await this.lcLLM\n .bindTools(input.tools ?? [])\n .stream(preparedInput.messages, preparedInput.options)\n : await this.lcLLM.stream(preparedInput.messages, preparedInput.options);\n\n for await (const response of stream) {\n const chunk = this.prepareOutput(response);\n yield chunk;\n }\n }\n\n protected prepareInput(input: ChatModelInput, run: RunContext<this>) {\n const messages: BaseMessageLike[] = input.messages.map((msg) => ({\n role: msg.role,\n content: msg.content,\n type: msg.role,\n // TODO\n }));\n\n const options: BaseChatModelCallOptions = {\n runId: run.runId,\n stop: input.stopSequences,\n signal: run.signal,\n tool_choice: input.toolChoice,\n };\n\n return { messages, options };\n }\n\n protected prepareOutput(output: AIMessageChunk) {\n const messages: Message[] = [];\n if (typeof output.content === \"string\") {\n messages.push(new AssistantMessage(output.content));\n } else {\n messages.push(\n new AssistantMessage(\n output.content.map((message) => {\n if (message.type === \"text\") {\n return { type: \"text\", text: message.text };\n } else if (message.type === \"image_url\") {\n return { type: \"text\", text: message.image_url.toString() };\n } else {\n throw new ValueError(`Unknown message type \"${message.type}\"`);\n }\n }),\n ),\n );\n }\n\n const usage: ChatModelUsage = {\n totalTokens: output.usage_metadata?.total_tokens ?? 0,\n promptTokens: output.usage_metadata?.input_tokens ?? 0,\n completionTokens: output.usage_metadata?.output_tokens ?? 0,\n };\n\n const stop: ChatModelFinishReason = output.response_metadata.stop_sequence || \"stop\";\n\n return new ChatModelOutput(messages, usage, stop);\n }\n\n protected async _createStructure<T>(\n input: ChatModelObjectInput<T>,\n run: RunContext<this>,\n ): Promise<ChatModelObjectOutput<T>> {\n const { messages, options } = this.prepareInput(input, run);\n const { raw, parsed } = await this.lcLLM\n .withStructuredOutput<any>(input.schema, {\n method: \"jsonSchema\",\n strict: false,\n includeRaw: true,\n })\n .invoke(messages, options);\n\n return { object: parsed as T, output: this.prepareOutput(raw as AIMessageChunk) };\n }\n\n createSnapshot() {\n return { ...super.createSnapshot(), emitter: this.emitter, lcLLM: this.lcLLM };\n }\n\n loadSnapshot(snapshot: ReturnType<typeof this.createSnapshot>): void {\n Object.assign(this, snapshot);\n }\n}\n"]}
1
+ {"version":3,"sources":["../../../../src/adapters/langchain/backend/chat.ts"],"names":["LangChainChatModel","ChatModel","emitter","lcLLM","parameters","Emitter","root","child","namespace","creator","modelId","_modelType","providerId","_create","input","run","preparedInput","prepareInput","response","bindTools","tools","invoke","messages","options","prepareOutput","_createStream","stream","chunk","map","msg","role","content","type","runId","stop","stopSequences","signal","tool_choice","toolChoice","output","push","AssistantMessage","message","text","image_url","toString","ValueError","response_metadata","stop_sequence","ChatModelOutput","totalTokens","usage_metadata","total_tokens","promptTokens","input_tokens","completionTokens","output_tokens","reasoningTokens","output_token_details","reasoning","cachedPromptTokens","input_token_details","cache_read","_createStructure","raw","parsed","withStructuredOutput","schema","method","strict","includeRaw","object","createSnapshot","loadSnapshot","snapshot","Object","assign"],"mappings":";;;;;;;;;AAyBO,MAAMA,2BAA2BC,kBAAAA,CAAAA;EAzBxC;;;;;AA0BkBC,EAAAA,OAAAA;EAEhB,WAAA,CACqBC,KAAAA,EACHC,UAAAA,GAAkC,EAAC,EACnD;AACA,IAAA,KAAA,EAAK,EAAA,IAAA,CAHcD,KAAAA,GAAAA,KAAAA,EAAAA,KACHC,UAAAA,GAAAA,UAAAA;AAGhB,IAAA,IAAA,CAAKF,OAAAA,GAAUG,mBAAAA,CAAQC,IAAAA,CAAKC,KAAAA,CAAM;MAChCC,SAAAA,EAAW;AAAC,QAAA,SAAA;AAAW,QAAA,WAAA;AAAa,QAAA;;MACpCC,OAAAA,EAAS;KACX,CAAA;AACF;AAEA,EAAA,IAAIC,OAAAA,GAAkB;AACpB,IAAA,OAAO,IAAA,CAAKP,MAAMQ,UAAAA,EAAU;AAC9B;AAEA,EAAA,IAAIC,UAAAA,GAAa;AACf,IAAA,OAAO,WAAA;AACT;EAEA,MAAgBC,OAAAA,CAAQC,OAAuBC,GAAAA,EAAiD;AAC9F,IAAA,MAAMC,aAAAA,GAAgB,IAAA,CAAKC,YAAAA,CAAaH,KAAAA,EAAOC,GAAAA,CAAAA;AAC/C,IAAA,MAAMG,QAAAA,GAAW,IAAA,CAAKf,KAAAA,CAAMgB,SAAAA,GACxB,MAAM,IAAA,CAAKhB,KAAAA,CACRgB,SAAAA,CAAUL,KAAAA,CAAMM,KAAAA,IAAS,EAAE,CAAA,CAC3BC,OAAOL,aAAAA,CAAcM,QAAAA,EAAUN,aAAAA,CAAcO,OAAO,CAAA,GACvD,MAAM,IAAA,CAAKpB,KAAAA,CAAMkB,MAAAA,CAAOL,aAAAA,CAAcM,QAAAA,EAAUN,aAAAA,CAAcO,OAAO,CAAA;AAEzE,IAAA,OAAO,IAAA,CAAKC,cAAcN,QAAAA,CAAAA;AAC5B;EAEA,OAAiBO,aAAAA,CACfX,OACAC,GAAAA,EACiC;AACjC,IAAA,MAAMC,aAAAA,GAAgB,IAAA,CAAKC,YAAAA,CAAaH,KAAAA,EAAOC,GAAAA,CAAAA;AAE/C,IAAA,MAAMW,MAAAA,GAAS,IAAA,CAAKvB,KAAAA,CAAMgB,SAAAA,GACtB,MAAM,IAAA,CAAKhB,KAAAA,CACRgB,SAAAA,CAAUL,KAAAA,CAAMM,KAAAA,IAAS,EAAE,CAAA,CAC3BM,OAAOV,aAAAA,CAAcM,QAAAA,EAAUN,aAAAA,CAAcO,OAAO,CAAA,GACvD,MAAM,IAAA,CAAKpB,KAAAA,CAAMuB,MAAAA,CAAOV,aAAAA,CAAcM,QAAAA,EAAUN,aAAAA,CAAcO,OAAO,CAAA;AAEzE,IAAA,WAAA,MAAiBL,YAAYQ,MAAAA,EAAQ;AACnC,MAAA,MAAMC,KAAAA,GAAQ,IAAA,CAAKH,aAAAA,CAAcN,QAAAA,CAAAA;AACjC,MAAA,MAAMS,KAAAA;AACR;AACF;AAEUV,EAAAA,YAAAA,CAAaH,OAAuBC,GAAAA,EAAuB;AACnE,IAAA,MAAMO,QAAAA,GAA8BR,KAAAA,CAAMQ,QAAAA,CAASM,GAAAA,CAAI,CAACC,GAAAA,MAAS;AAC/DC,MAAAA,IAAAA,EAAMD,GAAAA,CAAIC,IAAAA;AACVC,MAAAA,OAAAA,EAASF,GAAAA,CAAIE,OAAAA;AACbC,MAAAA,IAAAA,EAAMH,GAAAA,CAAIC;KAEZ,CAAA,CAAA;AAEA,IAAA,MAAMP,OAAAA,GAAoC;AACxCU,MAAAA,KAAAA,EAAOlB,GAAAA,CAAIkB,KAAAA;AACXC,MAAAA,IAAAA,EAAMpB,KAAAA,CAAMqB,aAAAA;AACZC,MAAAA,MAAAA,EAAQrB,GAAAA,CAAIqB,MAAAA;AACZC,MAAAA,WAAAA,EAAavB,KAAAA,CAAMwB;AACrB,KAAA;AAEA,IAAA,OAAO;AAAEhB,MAAAA,QAAAA;AAAUC,MAAAA;AAAQ,KAAA;AAC7B;AAEUC,EAAAA,aAAAA,CAAce,MAAAA,EAAwB;AAC9C,IAAA,MAAMjB,WAAsB,EAAA;AAC5B,IAAA,IAAI,OAAOiB,MAAAA,CAAOR,OAAAA,KAAY,QAAA,EAAU;AACtCT,MAAAA,QAAAA,CAASkB,IAAAA,CAAK,IAAIC,4BAAAA,CAAiBF,MAAAA,CAAOR,OAAO,CAAA,CAAA;KACnD,MAAO;AACLT,MAAAA,QAAAA,CAASkB,KACP,IAAIC,4BAAAA,CACFF,OAAOR,OAAAA,CAAQH,GAAAA,CAAI,CAACc,OAAAA,KAAAA;AAClB,QAAA,IAAIA,OAAAA,CAAQV,SAAS,MAAA,EAAQ;AAC3B,UAAA,OAAO;YAAEA,IAAAA,EAAM,MAAA;AAAQW,YAAAA,IAAAA,EAAMD,OAAAA,CAAQC;AAAK,WAAA;SAC5C,MAAA,IAAWD,OAAAA,CAAQV,SAAS,WAAA,EAAa;AACvC,UAAA,OAAO;YAAEA,IAAAA,EAAM,MAAA;YAAQW,IAAAA,EAAMD,OAAAA,CAAQE,UAAUC,QAAAA;AAAW,WAAA;SAC5D,MAAO;AACL,UAAA,MAAM,IAAIC,qBAAAA,CAAW,CAAA,sBAAA,EAAyBJ,OAAAA,CAAQV,IAAI,CAAA,CAAA,CAAG,CAAA;AAC/D;AACF,OAAA,CAAA,CAAA,CAAA;AAGN;AAEA,IAAA,MAAME,IAAAA,GAA8BK,MAAAA,CAAOQ,iBAAAA,CAAkBC,aAAAA,IAAiB,MAAA;AAC9E,IAAA,OAAO,IAAIC,yBACT3B,QAAAA,EACA;MACE4B,WAAAA,EAAaX,MAAAA,CAAOY,gBAAgBC,YAAAA,IAAgB,CAAA;MACpDC,YAAAA,EAAcd,MAAAA,CAAOY,gBAAgBG,YAAAA,IAAgB,CAAA;MACrDC,gBAAAA,EAAkBhB,MAAAA,CAAOY,gBAAgBK,aAAAA,IAAiB,CAAA;MAC1DC,eAAAA,EAAiBlB,MAAAA,CAAOY,gBAAgBO,oBAAAA,EAAsBC,SAAAA;MAC9DC,kBAAAA,EAAoBrB,MAAAA,CAAOY,gBAAgBU,mBAAAA,EAAqBC;AAClE,KAAA,EACA5B,IAAAA,CAAAA;AAEJ;EAEA,MAAgB6B,gBAAAA,CACdjD,OACAC,GAAAA,EACmC;AACnC,IAAA,MAAM,EAAEO,QAAAA,EAAUC,OAAAA,KAAY,IAAA,CAAKN,YAAAA,CAAaH,OAAOC,GAAAA,CAAAA;AACvD,IAAA,MAAM,EAAEiD,KAAKC,MAAAA,EAAM,GAAK,MAAM,IAAA,CAAK9D,KAAAA,CAChC+D,oBAAAA,CAA0BpD,KAAAA,CAAMqD,MAAAA,EAAQ;MACvCC,MAAAA,EAAQ,YAAA;MACRC,MAAAA,EAAQ,KAAA;MACRC,UAAAA,EAAY;KACd,CAAA,CACCjD,MAAAA,CAAOC,QAAAA,EAAUC,OAAAA,CAAAA;AAEpB,IAAA,OAAO;MAAEgD,MAAAA,EAAQN,MAAAA;MAAa1B,MAAAA,EAAQ,IAAA,CAAKf,cAAcwC,GAAAA;AAAuB,KAAA;AAClF;EAEAQ,cAAAA,GAAiB;AACf,IAAA,OAAO;AAAE,MAAA,GAAG,MAAMA,cAAAA,EAAAA;AAAkBtE,MAAAA,OAAAA,EAAS,IAAA,CAAKA,OAAAA;AAASC,MAAAA,KAAAA,EAAO,IAAA,CAAKA;AAAM,KAAA;AAC/E;AAEAsE,EAAAA,YAAAA,CAAaC,QAAAA,EAAwD;AACnEC,IAAAA,MAAAA,CAAOC,MAAAA,CAAO,MAAMF,QAAAA,CAAAA;AACtB;AACF","file":"chat.cjs","sourcesContent":["/**\n * Copyright 2025 © BeeAI a Series of LF Projects, LLC\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport {\n ChatModel,\n ChatModelEmitter,\n ChatModelFinishReason,\n ChatModelInput,\n ChatModelObjectInput,\n ChatModelObjectOutput,\n ChatModelOutput,\n ChatModelParameters,\n} from \"@/backend/chat.js\";\nimport { RunContext } from \"@/context.js\";\nimport { Emitter } from \"@/emitter/emitter.js\";\nimport {\n BaseChatModel,\n BaseChatModelCallOptions,\n} from \"@langchain/core/language_models/chat_models\";\nimport { AIMessageChunk, BaseMessageLike } from \"@langchain/core/messages\";\nimport { AssistantMessage, Message } from \"@/backend/message.js\";\nimport { ValueError } from \"@/errors.js\";\n\nexport class LangChainChatModel extends ChatModel {\n public readonly emitter: ChatModelEmitter;\n\n constructor(\n protected readonly lcLLM: BaseChatModel,\n public readonly parameters: ChatModelParameters = {},\n ) {\n super();\n this.emitter = Emitter.root.child({\n namespace: [\"backend\", \"langchain\", \"chat\"],\n creator: this,\n });\n }\n\n get modelId(): string {\n return this.lcLLM._modelType();\n }\n\n get providerId() {\n return \"langchain\";\n }\n\n protected async _create(input: ChatModelInput, run: RunContext<this>): Promise<ChatModelOutput> {\n const preparedInput = this.prepareInput(input, run);\n const response = this.lcLLM.bindTools\n ? await this.lcLLM\n .bindTools(input.tools ?? [])\n .invoke(preparedInput.messages, preparedInput.options)\n : await this.lcLLM.invoke(preparedInput.messages, preparedInput.options);\n\n return this.prepareOutput(response);\n }\n\n protected async *_createStream(\n input: ChatModelInput,\n run: RunContext<this>,\n ): AsyncGenerator<ChatModelOutput> {\n const preparedInput = this.prepareInput(input, run);\n\n const stream = this.lcLLM.bindTools\n ? await this.lcLLM\n .bindTools(input.tools ?? [])\n .stream(preparedInput.messages, preparedInput.options)\n : await this.lcLLM.stream(preparedInput.messages, preparedInput.options);\n\n for await (const response of stream) {\n const chunk = this.prepareOutput(response);\n yield chunk;\n }\n }\n\n protected prepareInput(input: ChatModelInput, run: RunContext<this>) {\n const messages: BaseMessageLike[] = input.messages.map((msg) => ({\n role: msg.role,\n content: msg.content,\n type: msg.role,\n // TODO\n }));\n\n const options: BaseChatModelCallOptions = {\n runId: run.runId,\n stop: input.stopSequences,\n signal: run.signal,\n tool_choice: input.toolChoice,\n };\n\n return { messages, options };\n }\n\n protected prepareOutput(output: AIMessageChunk) {\n const messages: Message[] = [];\n if (typeof output.content === \"string\") {\n messages.push(new AssistantMessage(output.content));\n } else {\n messages.push(\n new AssistantMessage(\n output.content.map((message) => {\n if (message.type === \"text\") {\n return { type: \"text\", text: message.text };\n } else if (message.type === \"image_url\") {\n return { type: \"text\", text: message.image_url.toString() };\n } else {\n throw new ValueError(`Unknown message type \"${message.type}\"`);\n }\n }),\n ),\n );\n }\n\n const stop: ChatModelFinishReason = output.response_metadata.stop_sequence || \"stop\";\n return new ChatModelOutput(\n messages,\n {\n totalTokens: output.usage_metadata?.total_tokens ?? 0,\n promptTokens: output.usage_metadata?.input_tokens ?? 0,\n completionTokens: output.usage_metadata?.output_tokens ?? 0,\n reasoningTokens: output.usage_metadata?.output_token_details?.reasoning,\n cachedPromptTokens: output.usage_metadata?.input_token_details?.cache_read,\n },\n stop,\n );\n }\n\n protected async _createStructure<T>(\n input: ChatModelObjectInput<T>,\n run: RunContext<this>,\n ): Promise<ChatModelObjectOutput<T>> {\n const { messages, options } = this.prepareInput(input, run);\n const { raw, parsed } = await this.lcLLM\n .withStructuredOutput<any>(input.schema, {\n method: \"jsonSchema\",\n strict: false,\n includeRaw: true,\n })\n .invoke(messages, options);\n\n return { object: parsed as T, output: this.prepareOutput(raw as AIMessageChunk) };\n }\n\n createSnapshot() {\n return { ...super.createSnapshot(), emitter: this.emitter, lcLLM: this.lcLLM };\n }\n\n loadSnapshot(snapshot: ReturnType<typeof this.createSnapshot>): void {\n Object.assign(this, snapshot);\n }\n}\n"]}
@@ -1,5 +1,5 @@
1
1
  import { Logger } from '../../../logger/logger.cjs';
2
- import { m as ChatModel, C as ChatModelParameters, h as ChatModelEmitter, d as ChatModelInput, n as ChatModelOutput, a as ChatModelObjectInput, b as ChatModelObjectOutput, i as ChatModelCache, l as ChatModelToolChoiceSupport } from '../../../chat-BZ55YQab.cjs';
2
+ import { m as ChatModel, C as ChatModelParameters, h as ChatModelEmitter, d as ChatModelInput, n as ChatModelOutput, a as ChatModelObjectInput, b as ChatModelObjectOutput, i as ChatModelCache, l as ChatModelToolChoiceSupport } from '../../../chat-CRb3vUVg.cjs';
3
3
  import { RunContext } from '../../../context.cjs';
4
4
  import { BaseChatModel, BaseChatModelCallOptions } from '@langchain/core/language_models/chat_models';
5
5
  import { BaseMessageLike, AIMessageChunk } from '@langchain/core/messages';
@@ -10,7 +10,7 @@ import '../../../internals/helpers/guards.cjs';
10
10
  import '../../../internals/serializable.cjs';
11
11
  import '../../../backend/message.cjs';
12
12
  import 'ai';
13
- import '../../../emitter-pJzHC_AM.cjs';
13
+ import '../../../emitter-DpqUYjXH.cjs';
14
14
  import 'promise-based-task';
15
15
  import '../../../cache/base.cjs';
16
16
  import '../../../backend/constants.cjs';
@@ -1,5 +1,5 @@
1
1
  import { Logger } from '../../../logger/logger.js';
2
- import { m as ChatModel, C as ChatModelParameters, h as ChatModelEmitter, d as ChatModelInput, n as ChatModelOutput, a as ChatModelObjectInput, b as ChatModelObjectOutput, i as ChatModelCache, l as ChatModelToolChoiceSupport } from '../../../chat-BBoOSvzm.js';
2
+ import { m as ChatModel, C as ChatModelParameters, h as ChatModelEmitter, d as ChatModelInput, n as ChatModelOutput, a as ChatModelObjectInput, b as ChatModelObjectOutput, i as ChatModelCache, l as ChatModelToolChoiceSupport } from '../../../chat-C0s-o6ll.js';
3
3
  import { RunContext } from '../../../context.js';
4
4
  import { BaseChatModel, BaseChatModelCallOptions } from '@langchain/core/language_models/chat_models';
5
5
  import { BaseMessageLike, AIMessageChunk } from '@langchain/core/messages';
@@ -10,7 +10,7 @@ import '../../../internals/helpers/guards.js';
10
10
  import '../../../internals/serializable.js';
11
11
  import '../../../backend/message.js';
12
12
  import 'ai';
13
- import '../../../emitter-BqpLJQVb.js';
13
+ import '../../../emitter-CZFbzlUi.js';
14
14
  import 'promise-based-task';
15
15
  import '../../../cache/base.js';
16
16
  import '../../../backend/constants.js';
@@ -80,13 +80,14 @@ class LangChainChatModel extends ChatModel {
80
80
  }
81
81
  })));
82
82
  }
83
- const usage = {
83
+ const stop = output.response_metadata.stop_sequence || "stop";
84
+ return new ChatModelOutput(messages, {
84
85
  totalTokens: output.usage_metadata?.total_tokens ?? 0,
85
86
  promptTokens: output.usage_metadata?.input_tokens ?? 0,
86
- completionTokens: output.usage_metadata?.output_tokens ?? 0
87
- };
88
- const stop = output.response_metadata.stop_sequence || "stop";
89
- return new ChatModelOutput(messages, usage, stop);
87
+ completionTokens: output.usage_metadata?.output_tokens ?? 0,
88
+ reasoningTokens: output.usage_metadata?.output_token_details?.reasoning,
89
+ cachedPromptTokens: output.usage_metadata?.input_token_details?.cache_read
90
+ }, stop);
90
91
  }
91
92
  async _createStructure(input, run) {
92
93
  const { messages, options } = this.prepareInput(input, run);
@@ -1 +1 @@
1
- {"version":3,"sources":["../../../../src/adapters/langchain/backend/chat.ts"],"names":["LangChainChatModel","ChatModel","emitter","lcLLM","parameters","Emitter","root","child","namespace","creator","modelId","_modelType","providerId","_create","input","run","preparedInput","prepareInput","response","bindTools","tools","invoke","messages","options","prepareOutput","_createStream","stream","chunk","map","msg","role","content","type","runId","stop","stopSequences","signal","tool_choice","toolChoice","output","push","AssistantMessage","message","text","image_url","toString","ValueError","usage","totalTokens","usage_metadata","total_tokens","promptTokens","input_tokens","completionTokens","output_tokens","response_metadata","stop_sequence","ChatModelOutput","_createStructure","raw","parsed","withStructuredOutput","schema","method","strict","includeRaw","object","createSnapshot","loadSnapshot","snapshot","Object","assign"],"mappings":";;;;;;;AA0BO,MAAMA,2BAA2BC,SAAAA,CAAAA;EA1BxC;;;;;AA2BkBC,EAAAA,OAAAA;EAEhB,WAAA,CACqBC,KAAAA,EACHC,UAAAA,GAAkC,EAAC,EACnD;AACA,IAAA,KAAA,EAAK,EAAA,IAAA,CAHcD,KAAAA,GAAAA,KAAAA,EAAAA,KACHC,UAAAA,GAAAA,UAAAA;AAGhB,IAAA,IAAA,CAAKF,OAAAA,GAAUG,OAAAA,CAAQC,IAAAA,CAAKC,KAAAA,CAAM;MAChCC,SAAAA,EAAW;AAAC,QAAA,SAAA;AAAW,QAAA,WAAA;AAAa,QAAA;;MACpCC,OAAAA,EAAS;KACX,CAAA;AACF;AAEA,EAAA,IAAIC,OAAAA,GAAkB;AACpB,IAAA,OAAO,IAAA,CAAKP,MAAMQ,UAAAA,EAAU;AAC9B;AAEA,EAAA,IAAIC,UAAAA,GAAa;AACf,IAAA,OAAO,WAAA;AACT;EAEA,MAAgBC,OAAAA,CAAQC,OAAuBC,GAAAA,EAAiD;AAC9F,IAAA,MAAMC,aAAAA,GAAgB,IAAA,CAAKC,YAAAA,CAAaH,KAAAA,EAAOC,GAAAA,CAAAA;AAC/C,IAAA,MAAMG,QAAAA,GAAW,IAAA,CAAKf,KAAAA,CAAMgB,SAAAA,GACxB,MAAM,IAAA,CAAKhB,KAAAA,CACRgB,SAAAA,CAAUL,KAAAA,CAAMM,KAAAA,IAAS,EAAE,CAAA,CAC3BC,OAAOL,aAAAA,CAAcM,QAAAA,EAAUN,aAAAA,CAAcO,OAAO,CAAA,GACvD,MAAM,IAAA,CAAKpB,KAAAA,CAAMkB,MAAAA,CAAOL,aAAAA,CAAcM,QAAAA,EAAUN,aAAAA,CAAcO,OAAO,CAAA;AAEzE,IAAA,OAAO,IAAA,CAAKC,cAAcN,QAAAA,CAAAA;AAC5B;EAEA,OAAiBO,aAAAA,CACfX,OACAC,GAAAA,EACiC;AACjC,IAAA,MAAMC,aAAAA,GAAgB,IAAA,CAAKC,YAAAA,CAAaH,KAAAA,EAAOC,GAAAA,CAAAA;AAE/C,IAAA,MAAMW,MAAAA,GAAS,IAAA,CAAKvB,KAAAA,CAAMgB,SAAAA,GACtB,MAAM,IAAA,CAAKhB,KAAAA,CACRgB,SAAAA,CAAUL,KAAAA,CAAMM,KAAAA,IAAS,EAAE,CAAA,CAC3BM,OAAOV,aAAAA,CAAcM,QAAAA,EAAUN,aAAAA,CAAcO,OAAO,CAAA,GACvD,MAAM,IAAA,CAAKpB,KAAAA,CAAMuB,MAAAA,CAAOV,aAAAA,CAAcM,QAAAA,EAAUN,aAAAA,CAAcO,OAAO,CAAA;AAEzE,IAAA,WAAA,MAAiBL,YAAYQ,MAAAA,EAAQ;AACnC,MAAA,MAAMC,KAAAA,GAAQ,IAAA,CAAKH,aAAAA,CAAcN,QAAAA,CAAAA;AACjC,MAAA,MAAMS,KAAAA;AACR;AACF;AAEUV,EAAAA,YAAAA,CAAaH,OAAuBC,GAAAA,EAAuB;AACnE,IAAA,MAAMO,QAAAA,GAA8BR,KAAAA,CAAMQ,QAAAA,CAASM,GAAAA,CAAI,CAACC,GAAAA,MAAS;AAC/DC,MAAAA,IAAAA,EAAMD,GAAAA,CAAIC,IAAAA;AACVC,MAAAA,OAAAA,EAASF,GAAAA,CAAIE,OAAAA;AACbC,MAAAA,IAAAA,EAAMH,GAAAA,CAAIC;KAEZ,CAAA,CAAA;AAEA,IAAA,MAAMP,OAAAA,GAAoC;AACxCU,MAAAA,KAAAA,EAAOlB,GAAAA,CAAIkB,KAAAA;AACXC,MAAAA,IAAAA,EAAMpB,KAAAA,CAAMqB,aAAAA;AACZC,MAAAA,MAAAA,EAAQrB,GAAAA,CAAIqB,MAAAA;AACZC,MAAAA,WAAAA,EAAavB,KAAAA,CAAMwB;AACrB,KAAA;AAEA,IAAA,OAAO;AAAEhB,MAAAA,QAAAA;AAAUC,MAAAA;AAAQ,KAAA;AAC7B;AAEUC,EAAAA,aAAAA,CAAce,MAAAA,EAAwB;AAC9C,IAAA,MAAMjB,WAAsB,EAAA;AAC5B,IAAA,IAAI,OAAOiB,MAAAA,CAAOR,OAAAA,KAAY,QAAA,EAAU;AACtCT,MAAAA,QAAAA,CAASkB,IAAAA,CAAK,IAAIC,gBAAAA,CAAiBF,MAAAA,CAAOR,OAAO,CAAA,CAAA;KACnD,MAAO;AACLT,MAAAA,QAAAA,CAASkB,KACP,IAAIC,gBAAAA,CACFF,OAAOR,OAAAA,CAAQH,GAAAA,CAAI,CAACc,OAAAA,KAAAA;AAClB,QAAA,IAAIA,OAAAA,CAAQV,SAAS,MAAA,EAAQ;AAC3B,UAAA,OAAO;YAAEA,IAAAA,EAAM,MAAA;AAAQW,YAAAA,IAAAA,EAAMD,OAAAA,CAAQC;AAAK,WAAA;SAC5C,MAAA,IAAWD,OAAAA,CAAQV,SAAS,WAAA,EAAa;AACvC,UAAA,OAAO;YAAEA,IAAAA,EAAM,MAAA;YAAQW,IAAAA,EAAMD,OAAAA,CAAQE,UAAUC,QAAAA;AAAW,WAAA;SAC5D,MAAO;AACL,UAAA,MAAM,IAAIC,UAAAA,CAAW,CAAA,sBAAA,EAAyBJ,OAAAA,CAAQV,IAAI,CAAA,CAAA,CAAG,CAAA;AAC/D;AACF,OAAA,CAAA,CAAA,CAAA;AAGN;AAEA,IAAA,MAAMe,KAAAA,GAAwB;MAC5BC,WAAAA,EAAaT,MAAAA,CAAOU,gBAAgBC,YAAAA,IAAgB,CAAA;MACpDC,YAAAA,EAAcZ,MAAAA,CAAOU,gBAAgBG,YAAAA,IAAgB,CAAA;MACrDC,gBAAAA,EAAkBd,MAAAA,CAAOU,gBAAgBK,aAAAA,IAAiB;AAC5D,KAAA;AAEA,IAAA,MAAMpB,IAAAA,GAA8BK,MAAAA,CAAOgB,iBAAAA,CAAkBC,aAAAA,IAAiB,MAAA;AAE9E,IAAA,OAAO,IAAIC,eAAAA,CAAgBnC,QAAAA,EAAUyB,KAAAA,EAAOb,IAAAA,CAAAA;AAC9C;EAEA,MAAgBwB,gBAAAA,CACd5C,OACAC,GAAAA,EACmC;AACnC,IAAA,MAAM,EAAEO,QAAAA,EAAUC,OAAAA,KAAY,IAAA,CAAKN,YAAAA,CAAaH,OAAOC,GAAAA,CAAAA;AACvD,IAAA,MAAM,EAAE4C,KAAKC,MAAAA,EAAM,GAAK,MAAM,IAAA,CAAKzD,KAAAA,CAChC0D,oBAAAA,CAA0B/C,KAAAA,CAAMgD,MAAAA,EAAQ;MACvCC,MAAAA,EAAQ,YAAA;MACRC,MAAAA,EAAQ,KAAA;MACRC,UAAAA,EAAY;KACd,CAAA,CACC5C,MAAAA,CAAOC,QAAAA,EAAUC,OAAAA,CAAAA;AAEpB,IAAA,OAAO;MAAE2C,MAAAA,EAAQN,MAAAA;MAAarB,MAAAA,EAAQ,IAAA,CAAKf,cAAcmC,GAAAA;AAAuB,KAAA;AAClF;EAEAQ,cAAAA,GAAiB;AACf,IAAA,OAAO;AAAE,MAAA,GAAG,MAAMA,cAAAA,EAAAA;AAAkBjE,MAAAA,OAAAA,EAAS,IAAA,CAAKA,OAAAA;AAASC,MAAAA,KAAAA,EAAO,IAAA,CAAKA;AAAM,KAAA;AAC/E;AAEAiE,EAAAA,YAAAA,CAAaC,QAAAA,EAAwD;AACnEC,IAAAA,MAAAA,CAAOC,MAAAA,CAAO,MAAMF,QAAAA,CAAAA;AACtB;AACF","file":"chat.js","sourcesContent":["/**\n * Copyright 2025 © BeeAI a Series of LF Projects, LLC\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport {\n ChatModel,\n ChatModelEmitter,\n ChatModelFinishReason,\n ChatModelInput,\n ChatModelObjectInput,\n ChatModelObjectOutput,\n ChatModelOutput,\n ChatModelParameters,\n ChatModelUsage,\n} from \"@/backend/chat.js\";\nimport { RunContext } from \"@/context.js\";\nimport { Emitter } from \"@/emitter/emitter.js\";\nimport {\n BaseChatModel,\n BaseChatModelCallOptions,\n} from \"@langchain/core/language_models/chat_models\";\nimport { AIMessageChunk, BaseMessageLike } from \"@langchain/core/messages\";\nimport { AssistantMessage, Message } from \"@/backend/message.js\";\nimport { ValueError } from \"@/errors.js\";\n\nexport class LangChainChatModel extends ChatModel {\n public readonly emitter: ChatModelEmitter;\n\n constructor(\n protected readonly lcLLM: BaseChatModel,\n public readonly parameters: ChatModelParameters = {},\n ) {\n super();\n this.emitter = Emitter.root.child({\n namespace: [\"backend\", \"langchain\", \"chat\"],\n creator: this,\n });\n }\n\n get modelId(): string {\n return this.lcLLM._modelType();\n }\n\n get providerId() {\n return \"langchain\";\n }\n\n protected async _create(input: ChatModelInput, run: RunContext<this>): Promise<ChatModelOutput> {\n const preparedInput = this.prepareInput(input, run);\n const response = this.lcLLM.bindTools\n ? await this.lcLLM\n .bindTools(input.tools ?? [])\n .invoke(preparedInput.messages, preparedInput.options)\n : await this.lcLLM.invoke(preparedInput.messages, preparedInput.options);\n\n return this.prepareOutput(response);\n }\n\n protected async *_createStream(\n input: ChatModelInput,\n run: RunContext<this>,\n ): AsyncGenerator<ChatModelOutput> {\n const preparedInput = this.prepareInput(input, run);\n\n const stream = this.lcLLM.bindTools\n ? await this.lcLLM\n .bindTools(input.tools ?? [])\n .stream(preparedInput.messages, preparedInput.options)\n : await this.lcLLM.stream(preparedInput.messages, preparedInput.options);\n\n for await (const response of stream) {\n const chunk = this.prepareOutput(response);\n yield chunk;\n }\n }\n\n protected prepareInput(input: ChatModelInput, run: RunContext<this>) {\n const messages: BaseMessageLike[] = input.messages.map((msg) => ({\n role: msg.role,\n content: msg.content,\n type: msg.role,\n // TODO\n }));\n\n const options: BaseChatModelCallOptions = {\n runId: run.runId,\n stop: input.stopSequences,\n signal: run.signal,\n tool_choice: input.toolChoice,\n };\n\n return { messages, options };\n }\n\n protected prepareOutput(output: AIMessageChunk) {\n const messages: Message[] = [];\n if (typeof output.content === \"string\") {\n messages.push(new AssistantMessage(output.content));\n } else {\n messages.push(\n new AssistantMessage(\n output.content.map((message) => {\n if (message.type === \"text\") {\n return { type: \"text\", text: message.text };\n } else if (message.type === \"image_url\") {\n return { type: \"text\", text: message.image_url.toString() };\n } else {\n throw new ValueError(`Unknown message type \"${message.type}\"`);\n }\n }),\n ),\n );\n }\n\n const usage: ChatModelUsage = {\n totalTokens: output.usage_metadata?.total_tokens ?? 0,\n promptTokens: output.usage_metadata?.input_tokens ?? 0,\n completionTokens: output.usage_metadata?.output_tokens ?? 0,\n };\n\n const stop: ChatModelFinishReason = output.response_metadata.stop_sequence || \"stop\";\n\n return new ChatModelOutput(messages, usage, stop);\n }\n\n protected async _createStructure<T>(\n input: ChatModelObjectInput<T>,\n run: RunContext<this>,\n ): Promise<ChatModelObjectOutput<T>> {\n const { messages, options } = this.prepareInput(input, run);\n const { raw, parsed } = await this.lcLLM\n .withStructuredOutput<any>(input.schema, {\n method: \"jsonSchema\",\n strict: false,\n includeRaw: true,\n })\n .invoke(messages, options);\n\n return { object: parsed as T, output: this.prepareOutput(raw as AIMessageChunk) };\n }\n\n createSnapshot() {\n return { ...super.createSnapshot(), emitter: this.emitter, lcLLM: this.lcLLM };\n }\n\n loadSnapshot(snapshot: ReturnType<typeof this.createSnapshot>): void {\n Object.assign(this, snapshot);\n }\n}\n"]}
1
+ {"version":3,"sources":["../../../../src/adapters/langchain/backend/chat.ts"],"names":["LangChainChatModel","ChatModel","emitter","lcLLM","parameters","Emitter","root","child","namespace","creator","modelId","_modelType","providerId","_create","input","run","preparedInput","prepareInput","response","bindTools","tools","invoke","messages","options","prepareOutput","_createStream","stream","chunk","map","msg","role","content","type","runId","stop","stopSequences","signal","tool_choice","toolChoice","output","push","AssistantMessage","message","text","image_url","toString","ValueError","response_metadata","stop_sequence","ChatModelOutput","totalTokens","usage_metadata","total_tokens","promptTokens","input_tokens","completionTokens","output_tokens","reasoningTokens","output_token_details","reasoning","cachedPromptTokens","input_token_details","cache_read","_createStructure","raw","parsed","withStructuredOutput","schema","method","strict","includeRaw","object","createSnapshot","loadSnapshot","snapshot","Object","assign"],"mappings":";;;;;;;AAyBO,MAAMA,2BAA2BC,SAAAA,CAAAA;EAzBxC;;;;;AA0BkBC,EAAAA,OAAAA;EAEhB,WAAA,CACqBC,KAAAA,EACHC,UAAAA,GAAkC,EAAC,EACnD;AACA,IAAA,KAAA,EAAK,EAAA,IAAA,CAHcD,KAAAA,GAAAA,KAAAA,EAAAA,KACHC,UAAAA,GAAAA,UAAAA;AAGhB,IAAA,IAAA,CAAKF,OAAAA,GAAUG,OAAAA,CAAQC,IAAAA,CAAKC,KAAAA,CAAM;MAChCC,SAAAA,EAAW;AAAC,QAAA,SAAA;AAAW,QAAA,WAAA;AAAa,QAAA;;MACpCC,OAAAA,EAAS;KACX,CAAA;AACF;AAEA,EAAA,IAAIC,OAAAA,GAAkB;AACpB,IAAA,OAAO,IAAA,CAAKP,MAAMQ,UAAAA,EAAU;AAC9B;AAEA,EAAA,IAAIC,UAAAA,GAAa;AACf,IAAA,OAAO,WAAA;AACT;EAEA,MAAgBC,OAAAA,CAAQC,OAAuBC,GAAAA,EAAiD;AAC9F,IAAA,MAAMC,aAAAA,GAAgB,IAAA,CAAKC,YAAAA,CAAaH,KAAAA,EAAOC,GAAAA,CAAAA;AAC/C,IAAA,MAAMG,QAAAA,GAAW,IAAA,CAAKf,KAAAA,CAAMgB,SAAAA,GACxB,MAAM,IAAA,CAAKhB,KAAAA,CACRgB,SAAAA,CAAUL,KAAAA,CAAMM,KAAAA,IAAS,EAAE,CAAA,CAC3BC,OAAOL,aAAAA,CAAcM,QAAAA,EAAUN,aAAAA,CAAcO,OAAO,CAAA,GACvD,MAAM,IAAA,CAAKpB,KAAAA,CAAMkB,MAAAA,CAAOL,aAAAA,CAAcM,QAAAA,EAAUN,aAAAA,CAAcO,OAAO,CAAA;AAEzE,IAAA,OAAO,IAAA,CAAKC,cAAcN,QAAAA,CAAAA;AAC5B;EAEA,OAAiBO,aAAAA,CACfX,OACAC,GAAAA,EACiC;AACjC,IAAA,MAAMC,aAAAA,GAAgB,IAAA,CAAKC,YAAAA,CAAaH,KAAAA,EAAOC,GAAAA,CAAAA;AAE/C,IAAA,MAAMW,MAAAA,GAAS,IAAA,CAAKvB,KAAAA,CAAMgB,SAAAA,GACtB,MAAM,IAAA,CAAKhB,KAAAA,CACRgB,SAAAA,CAAUL,KAAAA,CAAMM,KAAAA,IAAS,EAAE,CAAA,CAC3BM,OAAOV,aAAAA,CAAcM,QAAAA,EAAUN,aAAAA,CAAcO,OAAO,CAAA,GACvD,MAAM,IAAA,CAAKpB,KAAAA,CAAMuB,MAAAA,CAAOV,aAAAA,CAAcM,QAAAA,EAAUN,aAAAA,CAAcO,OAAO,CAAA;AAEzE,IAAA,WAAA,MAAiBL,YAAYQ,MAAAA,EAAQ;AACnC,MAAA,MAAMC,KAAAA,GAAQ,IAAA,CAAKH,aAAAA,CAAcN,QAAAA,CAAAA;AACjC,MAAA,MAAMS,KAAAA;AACR;AACF;AAEUV,EAAAA,YAAAA,CAAaH,OAAuBC,GAAAA,EAAuB;AACnE,IAAA,MAAMO,QAAAA,GAA8BR,KAAAA,CAAMQ,QAAAA,CAASM,GAAAA,CAAI,CAACC,GAAAA,MAAS;AAC/DC,MAAAA,IAAAA,EAAMD,GAAAA,CAAIC,IAAAA;AACVC,MAAAA,OAAAA,EAASF,GAAAA,CAAIE,OAAAA;AACbC,MAAAA,IAAAA,EAAMH,GAAAA,CAAIC;KAEZ,CAAA,CAAA;AAEA,IAAA,MAAMP,OAAAA,GAAoC;AACxCU,MAAAA,KAAAA,EAAOlB,GAAAA,CAAIkB,KAAAA;AACXC,MAAAA,IAAAA,EAAMpB,KAAAA,CAAMqB,aAAAA;AACZC,MAAAA,MAAAA,EAAQrB,GAAAA,CAAIqB,MAAAA;AACZC,MAAAA,WAAAA,EAAavB,KAAAA,CAAMwB;AACrB,KAAA;AAEA,IAAA,OAAO;AAAEhB,MAAAA,QAAAA;AAAUC,MAAAA;AAAQ,KAAA;AAC7B;AAEUC,EAAAA,aAAAA,CAAce,MAAAA,EAAwB;AAC9C,IAAA,MAAMjB,WAAsB,EAAA;AAC5B,IAAA,IAAI,OAAOiB,MAAAA,CAAOR,OAAAA,KAAY,QAAA,EAAU;AACtCT,MAAAA,QAAAA,CAASkB,IAAAA,CAAK,IAAIC,gBAAAA,CAAiBF,MAAAA,CAAOR,OAAO,CAAA,CAAA;KACnD,MAAO;AACLT,MAAAA,QAAAA,CAASkB,KACP,IAAIC,gBAAAA,CACFF,OAAOR,OAAAA,CAAQH,GAAAA,CAAI,CAACc,OAAAA,KAAAA;AAClB,QAAA,IAAIA,OAAAA,CAAQV,SAAS,MAAA,EAAQ;AAC3B,UAAA,OAAO;YAAEA,IAAAA,EAAM,MAAA;AAAQW,YAAAA,IAAAA,EAAMD,OAAAA,CAAQC;AAAK,WAAA;SAC5C,MAAA,IAAWD,OAAAA,CAAQV,SAAS,WAAA,EAAa;AACvC,UAAA,OAAO;YAAEA,IAAAA,EAAM,MAAA;YAAQW,IAAAA,EAAMD,OAAAA,CAAQE,UAAUC,QAAAA;AAAW,WAAA;SAC5D,MAAO;AACL,UAAA,MAAM,IAAIC,UAAAA,CAAW,CAAA,sBAAA,EAAyBJ,OAAAA,CAAQV,IAAI,CAAA,CAAA,CAAG,CAAA;AAC/D;AACF,OAAA,CAAA,CAAA,CAAA;AAGN;AAEA,IAAA,MAAME,IAAAA,GAA8BK,MAAAA,CAAOQ,iBAAAA,CAAkBC,aAAAA,IAAiB,MAAA;AAC9E,IAAA,OAAO,IAAIC,gBACT3B,QAAAA,EACA;MACE4B,WAAAA,EAAaX,MAAAA,CAAOY,gBAAgBC,YAAAA,IAAgB,CAAA;MACpDC,YAAAA,EAAcd,MAAAA,CAAOY,gBAAgBG,YAAAA,IAAgB,CAAA;MACrDC,gBAAAA,EAAkBhB,MAAAA,CAAOY,gBAAgBK,aAAAA,IAAiB,CAAA;MAC1DC,eAAAA,EAAiBlB,MAAAA,CAAOY,gBAAgBO,oBAAAA,EAAsBC,SAAAA;MAC9DC,kBAAAA,EAAoBrB,MAAAA,CAAOY,gBAAgBU,mBAAAA,EAAqBC;AAClE,KAAA,EACA5B,IAAAA,CAAAA;AAEJ;EAEA,MAAgB6B,gBAAAA,CACdjD,OACAC,GAAAA,EACmC;AACnC,IAAA,MAAM,EAAEO,QAAAA,EAAUC,OAAAA,KAAY,IAAA,CAAKN,YAAAA,CAAaH,OAAOC,GAAAA,CAAAA;AACvD,IAAA,MAAM,EAAEiD,KAAKC,MAAAA,EAAM,GAAK,MAAM,IAAA,CAAK9D,KAAAA,CAChC+D,oBAAAA,CAA0BpD,KAAAA,CAAMqD,MAAAA,EAAQ;MACvCC,MAAAA,EAAQ,YAAA;MACRC,MAAAA,EAAQ,KAAA;MACRC,UAAAA,EAAY;KACd,CAAA,CACCjD,MAAAA,CAAOC,QAAAA,EAAUC,OAAAA,CAAAA;AAEpB,IAAA,OAAO;MAAEgD,MAAAA,EAAQN,MAAAA;MAAa1B,MAAAA,EAAQ,IAAA,CAAKf,cAAcwC,GAAAA;AAAuB,KAAA;AAClF;EAEAQ,cAAAA,GAAiB;AACf,IAAA,OAAO;AAAE,MAAA,GAAG,MAAMA,cAAAA,EAAAA;AAAkBtE,MAAAA,OAAAA,EAAS,IAAA,CAAKA,OAAAA;AAASC,MAAAA,KAAAA,EAAO,IAAA,CAAKA;AAAM,KAAA;AAC/E;AAEAsE,EAAAA,YAAAA,CAAaC,QAAAA,EAAwD;AACnEC,IAAAA,MAAAA,CAAOC,MAAAA,CAAO,MAAMF,QAAAA,CAAAA;AACtB;AACF","file":"chat.js","sourcesContent":["/**\n * Copyright 2025 © BeeAI a Series of LF Projects, LLC\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport {\n ChatModel,\n ChatModelEmitter,\n ChatModelFinishReason,\n ChatModelInput,\n ChatModelObjectInput,\n ChatModelObjectOutput,\n ChatModelOutput,\n ChatModelParameters,\n} from \"@/backend/chat.js\";\nimport { RunContext } from \"@/context.js\";\nimport { Emitter } from \"@/emitter/emitter.js\";\nimport {\n BaseChatModel,\n BaseChatModelCallOptions,\n} from \"@langchain/core/language_models/chat_models\";\nimport { AIMessageChunk, BaseMessageLike } from \"@langchain/core/messages\";\nimport { AssistantMessage, Message } from \"@/backend/message.js\";\nimport { ValueError } from \"@/errors.js\";\n\nexport class LangChainChatModel extends ChatModel {\n public readonly emitter: ChatModelEmitter;\n\n constructor(\n protected readonly lcLLM: BaseChatModel,\n public readonly parameters: ChatModelParameters = {},\n ) {\n super();\n this.emitter = Emitter.root.child({\n namespace: [\"backend\", \"langchain\", \"chat\"],\n creator: this,\n });\n }\n\n get modelId(): string {\n return this.lcLLM._modelType();\n }\n\n get providerId() {\n return \"langchain\";\n }\n\n protected async _create(input: ChatModelInput, run: RunContext<this>): Promise<ChatModelOutput> {\n const preparedInput = this.prepareInput(input, run);\n const response = this.lcLLM.bindTools\n ? await this.lcLLM\n .bindTools(input.tools ?? [])\n .invoke(preparedInput.messages, preparedInput.options)\n : await this.lcLLM.invoke(preparedInput.messages, preparedInput.options);\n\n return this.prepareOutput(response);\n }\n\n protected async *_createStream(\n input: ChatModelInput,\n run: RunContext<this>,\n ): AsyncGenerator<ChatModelOutput> {\n const preparedInput = this.prepareInput(input, run);\n\n const stream = this.lcLLM.bindTools\n ? await this.lcLLM\n .bindTools(input.tools ?? [])\n .stream(preparedInput.messages, preparedInput.options)\n : await this.lcLLM.stream(preparedInput.messages, preparedInput.options);\n\n for await (const response of stream) {\n const chunk = this.prepareOutput(response);\n yield chunk;\n }\n }\n\n protected prepareInput(input: ChatModelInput, run: RunContext<this>) {\n const messages: BaseMessageLike[] = input.messages.map((msg) => ({\n role: msg.role,\n content: msg.content,\n type: msg.role,\n // TODO\n }));\n\n const options: BaseChatModelCallOptions = {\n runId: run.runId,\n stop: input.stopSequences,\n signal: run.signal,\n tool_choice: input.toolChoice,\n };\n\n return { messages, options };\n }\n\n protected prepareOutput(output: AIMessageChunk) {\n const messages: Message[] = [];\n if (typeof output.content === \"string\") {\n messages.push(new AssistantMessage(output.content));\n } else {\n messages.push(\n new AssistantMessage(\n output.content.map((message) => {\n if (message.type === \"text\") {\n return { type: \"text\", text: message.text };\n } else if (message.type === \"image_url\") {\n return { type: \"text\", text: message.image_url.toString() };\n } else {\n throw new ValueError(`Unknown message type \"${message.type}\"`);\n }\n }),\n ),\n );\n }\n\n const stop: ChatModelFinishReason = output.response_metadata.stop_sequence || \"stop\";\n return new ChatModelOutput(\n messages,\n {\n totalTokens: output.usage_metadata?.total_tokens ?? 0,\n promptTokens: output.usage_metadata?.input_tokens ?? 0,\n completionTokens: output.usage_metadata?.output_tokens ?? 0,\n reasoningTokens: output.usage_metadata?.output_token_details?.reasoning,\n cachedPromptTokens: output.usage_metadata?.input_token_details?.cache_read,\n },\n stop,\n );\n }\n\n protected async _createStructure<T>(\n input: ChatModelObjectInput<T>,\n run: RunContext<this>,\n ): Promise<ChatModelObjectOutput<T>> {\n const { messages, options } = this.prepareInput(input, run);\n const { raw, parsed } = await this.lcLLM\n .withStructuredOutput<any>(input.schema, {\n method: \"jsonSchema\",\n strict: false,\n includeRaw: true,\n })\n .invoke(messages, options);\n\n return { object: parsed as T, output: this.prepareOutput(raw as AIMessageChunk) };\n }\n\n createSnapshot() {\n return { ...super.createSnapshot(), emitter: this.emitter, lcLLM: this.lcLLM };\n }\n\n loadSnapshot(snapshot: ReturnType<typeof this.createSnapshot>): void {\n Object.assign(this, snapshot);\n }\n}\n"]}
@@ -1,12 +1,12 @@
1
1
  import { EmbeddingModel, EmbeddingModelEvents, EmbeddingModelInput, EmbeddingModelOutput } from '../../../backend/embedding.cjs';
2
2
  import { RunContext } from '../../../context.cjs';
3
- import { E as Emitter } from '../../../emitter-pJzHC_AM.cjs';
3
+ import { E as Emitter } from '../../../emitter-DpqUYjXH.cjs';
4
4
  import { Embeddings } from '@langchain/core/embeddings';
5
5
  import '../../../internals/serializable.cjs';
6
6
  import '../../../internals/types.cjs';
7
7
  import '../../../internals/helpers/guards.cjs';
8
8
  import '../../../errors.cjs';
9
- import '../../../chat-BZ55YQab.cjs';
9
+ import '../../../chat-CRb3vUVg.cjs';
10
10
  import '../../../backend/message.cjs';
11
11
  import 'ai';
12
12
  import 'promise-based-task';
@@ -35,7 +35,7 @@ declare class LangChainEmbeddingModel extends EmbeddingModel {
35
35
  get providerId(): string;
36
36
  protected _create(input: EmbeddingModelInput, run: RunContext<this>): Promise<EmbeddingModelOutput>;
37
37
  createSnapshot(): {
38
- lcEmbedding: Embeddings;
38
+ lcEmbedding: Embeddings<number[]>;
39
39
  emitter: Emitter<EmbeddingModelEvents>;
40
40
  };
41
41
  loadSnapshot(snapshot: ReturnType<typeof this.createSnapshot>): void;