beeai-framework 0.1.22 → 0.1.23

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (267) hide show
  1. package/dist/adapters/a2a/agents/agent.d.cts +2 -2
  2. package/dist/adapters/a2a/agents/agent.d.ts +2 -2
  3. package/dist/adapters/a2a/agents/events.d.cts +1 -1
  4. package/dist/adapters/a2a/agents/events.d.ts +1 -1
  5. package/dist/adapters/a2a/agents/types.d.cts +2 -2
  6. package/dist/adapters/a2a/agents/types.d.ts +2 -2
  7. package/dist/adapters/a2a/serve/agent_executor.d.cts +3 -3
  8. package/dist/adapters/a2a/serve/agent_executor.d.ts +3 -3
  9. package/dist/adapters/a2a/serve/server.d.cts +2 -2
  10. package/dist/adapters/a2a/serve/server.d.ts +2 -2
  11. package/dist/adapters/acp/agents/agent.d.cts +2 -2
  12. package/dist/adapters/acp/agents/agent.d.ts +2 -2
  13. package/dist/adapters/acp/agents/events.d.cts +1 -1
  14. package/dist/adapters/acp/agents/events.d.ts +1 -1
  15. package/dist/adapters/agentstack/agents/agent.d.cts +2 -2
  16. package/dist/adapters/agentstack/agents/agent.d.ts +2 -2
  17. package/dist/adapters/agentstack/agents/events.d.cts +1 -1
  18. package/dist/adapters/agentstack/agents/events.d.ts +1 -1
  19. package/dist/adapters/amazon-bedrock/backend/chat.d.cts +2 -2
  20. package/dist/adapters/amazon-bedrock/backend/chat.d.ts +2 -2
  21. package/dist/adapters/amazon-bedrock/backend/embedding.d.cts +2 -2
  22. package/dist/adapters/amazon-bedrock/backend/embedding.d.ts +2 -2
  23. package/dist/adapters/anthropic/backend/chat.d.cts +2 -2
  24. package/dist/adapters/anthropic/backend/chat.d.ts +2 -2
  25. package/dist/adapters/anthropic/backend/embedding.d.cts +2 -2
  26. package/dist/adapters/anthropic/backend/embedding.d.ts +2 -2
  27. package/dist/adapters/azure-openai/backend/chat.d.cts +2 -2
  28. package/dist/adapters/azure-openai/backend/chat.d.ts +2 -2
  29. package/dist/adapters/azure-openai/backend/embedding.d.cts +2 -2
  30. package/dist/adapters/azure-openai/backend/embedding.d.ts +2 -2
  31. package/dist/adapters/dummy/backend/chat.d.cts +2 -2
  32. package/dist/adapters/dummy/backend/chat.d.ts +2 -2
  33. package/dist/adapters/dummy/backend/embedding.d.cts +2 -2
  34. package/dist/adapters/dummy/backend/embedding.d.ts +2 -2
  35. package/dist/adapters/google-vertex/backend/chat.d.cts +2 -2
  36. package/dist/adapters/google-vertex/backend/chat.d.ts +2 -2
  37. package/dist/adapters/google-vertex/backend/embedding.d.cts +2 -2
  38. package/dist/adapters/google-vertex/backend/embedding.d.ts +2 -2
  39. package/dist/adapters/groq/backend/chat.d.cts +2 -2
  40. package/dist/adapters/groq/backend/chat.d.ts +2 -2
  41. package/dist/adapters/groq/backend/embedding.d.cts +2 -2
  42. package/dist/adapters/groq/backend/embedding.d.ts +2 -2
  43. package/dist/adapters/langchain/backend/chat.d.cts +2 -2
  44. package/dist/adapters/langchain/backend/chat.d.ts +2 -2
  45. package/dist/adapters/langchain/backend/embedding.d.cts +2 -2
  46. package/dist/adapters/langchain/backend/embedding.d.ts +2 -2
  47. package/dist/adapters/langchain/tools.d.cts +1 -1
  48. package/dist/adapters/langchain/tools.d.ts +1 -1
  49. package/dist/adapters/ollama/backend/chat.cjs +1 -2
  50. package/dist/adapters/ollama/backend/chat.cjs.map +1 -1
  51. package/dist/adapters/ollama/backend/chat.d.cts +2 -3
  52. package/dist/adapters/ollama/backend/chat.d.ts +2 -3
  53. package/dist/adapters/ollama/backend/chat.js +1 -2
  54. package/dist/adapters/ollama/backend/chat.js.map +1 -1
  55. package/dist/adapters/ollama/backend/embedding.d.cts +2 -2
  56. package/dist/adapters/ollama/backend/embedding.d.ts +2 -2
  57. package/dist/adapters/openai/backend/chat.d.cts +2 -2
  58. package/dist/adapters/openai/backend/chat.d.ts +2 -2
  59. package/dist/adapters/openai/backend/embedding.d.cts +2 -2
  60. package/dist/adapters/openai/backend/embedding.d.ts +2 -2
  61. package/dist/adapters/vercel/backend/chat.cjs +77 -17
  62. package/dist/adapters/vercel/backend/chat.cjs.map +1 -1
  63. package/dist/adapters/vercel/backend/chat.d.cts +2 -2
  64. package/dist/adapters/vercel/backend/chat.d.ts +2 -2
  65. package/dist/adapters/vercel/backend/chat.js +77 -17
  66. package/dist/adapters/vercel/backend/chat.js.map +1 -1
  67. package/dist/adapters/vercel/backend/embedding.d.cts +2 -2
  68. package/dist/adapters/vercel/backend/embedding.d.ts +2 -2
  69. package/dist/adapters/vercel/backend/utils.cjs +17 -0
  70. package/dist/adapters/vercel/backend/utils.cjs.map +1 -1
  71. package/dist/adapters/vercel/backend/utils.d.cts +5 -4
  72. package/dist/adapters/vercel/backend/utils.d.ts +5 -4
  73. package/dist/adapters/vercel/backend/utils.js +17 -1
  74. package/dist/adapters/vercel/backend/utils.js.map +1 -1
  75. package/dist/adapters/watsonx/backend/chat.cjs +7 -8
  76. package/dist/adapters/watsonx/backend/chat.cjs.map +1 -1
  77. package/dist/adapters/watsonx/backend/chat.d.cts +4 -4
  78. package/dist/adapters/watsonx/backend/chat.d.ts +4 -4
  79. package/dist/adapters/watsonx/backend/chat.js +7 -8
  80. package/dist/adapters/watsonx/backend/chat.js.map +1 -1
  81. package/dist/adapters/watsonx/backend/embedding.d.cts +2 -2
  82. package/dist/adapters/watsonx/backend/embedding.d.ts +2 -2
  83. package/dist/adapters/xai/backend/chat.d.cts +2 -2
  84. package/dist/adapters/xai/backend/chat.d.ts +2 -2
  85. package/dist/{agent-BMfyig7X.d.cts → agent-BVuc9Nfn.d.cts} +3 -3
  86. package/dist/{agent-Dxw8AGWb.d.ts → agent-Cuy8Q4gw.d.ts} +3 -3
  87. package/dist/agents/base.d.cts +2 -2
  88. package/dist/agents/base.d.ts +2 -2
  89. package/dist/agents/experimental/replan/agent.d.cts +3 -3
  90. package/dist/agents/experimental/replan/agent.d.ts +3 -3
  91. package/dist/agents/experimental/replan/prompts.d.cts +1 -1
  92. package/dist/agents/experimental/replan/prompts.d.ts +1 -1
  93. package/dist/agents/experimental/streamlit/agent.d.cts +3 -3
  94. package/dist/agents/experimental/streamlit/agent.d.ts +3 -3
  95. package/dist/agents/react/agent.d.cts +4 -4
  96. package/dist/agents/react/agent.d.ts +4 -4
  97. package/dist/agents/react/prompts.d.cts +1 -1
  98. package/dist/agents/react/prompts.d.ts +1 -1
  99. package/dist/agents/react/runners/base.d.cts +4 -4
  100. package/dist/agents/react/runners/base.d.ts +4 -4
  101. package/dist/agents/react/runners/deep-think/prompts.d.cts +1 -1
  102. package/dist/agents/react/runners/deep-think/prompts.d.ts +1 -1
  103. package/dist/agents/react/runners/deep-think/runner.d.cts +5 -5
  104. package/dist/agents/react/runners/deep-think/runner.d.ts +5 -5
  105. package/dist/agents/react/runners/default/runner.d.cts +6 -6
  106. package/dist/agents/react/runners/default/runner.d.ts +6 -6
  107. package/dist/agents/react/runners/granite/prompts.d.cts +1 -1
  108. package/dist/agents/react/runners/granite/prompts.d.ts +1 -1
  109. package/dist/agents/react/runners/granite/runner.d.cts +5 -5
  110. package/dist/agents/react/runners/granite/runner.d.ts +5 -5
  111. package/dist/agents/react/types.d.cts +2 -2
  112. package/dist/agents/react/types.d.ts +2 -2
  113. package/dist/agents/requirement/agent.cjs +1 -1
  114. package/dist/agents/requirement/agent.cjs.map +1 -1
  115. package/dist/agents/requirement/agent.d.cts +4 -4
  116. package/dist/agents/requirement/agent.d.ts +4 -4
  117. package/dist/agents/requirement/agent.js +1 -1
  118. package/dist/agents/requirement/agent.js.map +1 -1
  119. package/dist/agents/requirement/prompts.d.cts +2 -2
  120. package/dist/agents/requirement/prompts.d.ts +2 -2
  121. package/dist/agents/requirement/requirements/conditional.d.cts +4 -4
  122. package/dist/agents/requirement/requirements/conditional.d.ts +4 -4
  123. package/dist/agents/requirement/requirements/requirement.d.cts +4 -4
  124. package/dist/agents/requirement/requirements/requirement.d.ts +4 -4
  125. package/dist/agents/requirement/requirements/utils.d.cts +1 -1
  126. package/dist/agents/requirement/requirements/utils.d.ts +1 -1
  127. package/dist/agents/requirement/runner.cjs +31 -15
  128. package/dist/agents/requirement/runner.cjs.map +1 -1
  129. package/dist/agents/requirement/runner.d.cts +13 -15
  130. package/dist/agents/requirement/runner.d.ts +13 -15
  131. package/dist/agents/requirement/runner.js +31 -15
  132. package/dist/agents/requirement/runner.js.map +1 -1
  133. package/dist/agents/requirement/types.d.cts +4 -4
  134. package/dist/agents/requirement/types.d.ts +4 -4
  135. package/dist/agents/requirement/utils/llm.d.cts +4 -4
  136. package/dist/agents/requirement/utils/llm.d.ts +4 -4
  137. package/dist/agents/requirement/utils/tool.cjs +2 -1
  138. package/dist/agents/requirement/utils/tool.cjs.map +1 -1
  139. package/dist/agents/requirement/utils/tool.d.cts +4 -4
  140. package/dist/agents/requirement/utils/tool.d.ts +4 -4
  141. package/dist/agents/requirement/utils/tool.js +2 -1
  142. package/dist/agents/requirement/utils/tool.js.map +1 -1
  143. package/dist/agents/toolCalling/agent.cjs +4 -1
  144. package/dist/agents/toolCalling/agent.cjs.map +1 -1
  145. package/dist/agents/toolCalling/agent.d.cts +6 -3
  146. package/dist/agents/toolCalling/agent.d.ts +6 -3
  147. package/dist/agents/toolCalling/agent.js +5 -2
  148. package/dist/agents/toolCalling/agent.js.map +1 -1
  149. package/dist/agents/toolCalling/types.d.cts +1 -1
  150. package/dist/agents/toolCalling/types.d.ts +1 -1
  151. package/dist/agents/types.d.cts +2 -2
  152. package/dist/agents/types.d.ts +2 -2
  153. package/dist/backend/backend.d.cts +2 -2
  154. package/dist/backend/backend.d.ts +2 -2
  155. package/dist/backend/chat.cjs +122 -4
  156. package/dist/backend/chat.cjs.map +1 -1
  157. package/dist/backend/chat.d.cts +3 -3
  158. package/dist/backend/chat.d.ts +3 -3
  159. package/dist/backend/chat.js +123 -5
  160. package/dist/backend/chat.js.map +1 -1
  161. package/dist/backend/core.d.cts +3 -3
  162. package/dist/backend/core.d.ts +3 -3
  163. package/dist/backend/embedding.d.cts +2 -2
  164. package/dist/backend/embedding.d.ts +2 -2
  165. package/dist/backend/message.cjs +20 -13
  166. package/dist/backend/message.cjs.map +1 -1
  167. package/dist/backend/message.d.cts +7 -3
  168. package/dist/backend/message.d.ts +7 -3
  169. package/dist/backend/message.js +20 -13
  170. package/dist/backend/message.js.map +1 -1
  171. package/dist/backend/utils.d.cts +2 -2
  172. package/dist/backend/utils.d.ts +2 -2
  173. package/dist/{base-BurRTBA7.d.ts → base-CyFx7BRL.d.ts} +2 -2
  174. package/dist/{base-v-PYIq5G.d.cts → base-XfWeXQ-S.d.cts} +2 -2
  175. package/dist/{chat-DB_cf15S.d.cts → chat-BOb9n7AJ.d.cts} +7 -8
  176. package/dist/{chat-BbCYMAdE.d.ts → chat-DrqwUgD8.d.ts} +7 -8
  177. package/dist/context.d.cts +1 -1
  178. package/dist/context.d.ts +1 -1
  179. package/dist/emitter/emitter.cjs +2 -1
  180. package/dist/emitter/emitter.cjs.map +1 -1
  181. package/dist/emitter/emitter.d.cts +1 -1
  182. package/dist/emitter/emitter.d.ts +1 -1
  183. package/dist/emitter/emitter.js +2 -1
  184. package/dist/emitter/emitter.js.map +1 -1
  185. package/dist/emitter/types.d.cts +1 -1
  186. package/dist/emitter/types.d.ts +1 -1
  187. package/dist/{emitter-CN7I4uSV.d.ts → emitter-36-9MnvA.d.ts} +2 -2
  188. package/dist/{emitter-CGKd_-kK.d.cts → emitter-jN55XZZq.d.cts} +2 -2
  189. package/dist/index.d.cts +3 -3
  190. package/dist/index.d.ts +3 -3
  191. package/dist/internals/fetcher.d.cts +1 -1
  192. package/dist/internals/fetcher.d.ts +1 -1
  193. package/dist/memory/summarizeMemory.d.cts +2 -2
  194. package/dist/memory/summarizeMemory.d.ts +2 -2
  195. package/dist/middleware/streamToolCall.cjs +132 -0
  196. package/dist/middleware/streamToolCall.cjs.map +1 -0
  197. package/dist/middleware/streamToolCall.d.cts +94 -0
  198. package/dist/middleware/streamToolCall.d.ts +94 -0
  199. package/dist/middleware/streamToolCall.js +130 -0
  200. package/dist/middleware/streamToolCall.js.map +1 -0
  201. package/dist/middleware/trajectory.cjs +14 -5
  202. package/dist/middleware/trajectory.cjs.map +1 -1
  203. package/dist/middleware/trajectory.d.cts +1 -1
  204. package/dist/middleware/trajectory.d.ts +1 -1
  205. package/dist/middleware/trajectory.js +13 -4
  206. package/dist/middleware/trajectory.js.map +1 -1
  207. package/dist/parsers/linePrefix.d.cts +1 -1
  208. package/dist/parsers/linePrefix.d.ts +1 -1
  209. package/dist/tools/arxiv.d.cts +1 -1
  210. package/dist/tools/arxiv.d.ts +1 -1
  211. package/dist/tools/base.d.cts +1 -1
  212. package/dist/tools/base.d.ts +1 -1
  213. package/dist/tools/calculator.d.cts +1 -1
  214. package/dist/tools/calculator.d.ts +1 -1
  215. package/dist/tools/custom.d.cts +2 -2
  216. package/dist/tools/custom.d.ts +2 -2
  217. package/dist/tools/database/elasticsearch.d.cts +1 -1
  218. package/dist/tools/database/elasticsearch.d.ts +1 -1
  219. package/dist/tools/database/milvus.d.cts +3 -3
  220. package/dist/tools/database/milvus.d.ts +3 -3
  221. package/dist/tools/database/qdrant.d.cts +1 -1
  222. package/dist/tools/database/qdrant.d.ts +1 -1
  223. package/dist/tools/database/sql.d.cts +1 -1
  224. package/dist/tools/database/sql.d.ts +1 -1
  225. package/dist/tools/handoff.cjs +68 -0
  226. package/dist/tools/handoff.cjs.map +1 -0
  227. package/dist/tools/handoff.d.cts +47 -0
  228. package/dist/tools/handoff.d.ts +47 -0
  229. package/dist/tools/handoff.js +66 -0
  230. package/dist/tools/handoff.js.map +1 -0
  231. package/dist/tools/llm.d.cts +2 -2
  232. package/dist/tools/llm.d.ts +2 -2
  233. package/dist/tools/mcp.d.cts +1 -1
  234. package/dist/tools/mcp.d.ts +1 -1
  235. package/dist/tools/openapi.d.cts +1 -1
  236. package/dist/tools/openapi.d.ts +1 -1
  237. package/dist/tools/python/output.d.cts +1 -1
  238. package/dist/tools/python/output.d.ts +1 -1
  239. package/dist/tools/python/python.d.cts +2 -2
  240. package/dist/tools/python/python.d.ts +2 -2
  241. package/dist/tools/search/base.d.cts +1 -1
  242. package/dist/tools/search/base.d.ts +1 -1
  243. package/dist/tools/search/duckDuckGoSearch.d.cts +1 -1
  244. package/dist/tools/search/duckDuckGoSearch.d.ts +1 -1
  245. package/dist/tools/search/googleSearch.d.cts +1 -1
  246. package/dist/tools/search/googleSearch.d.ts +1 -1
  247. package/dist/tools/search/searXNGSearch.d.cts +1 -1
  248. package/dist/tools/search/searXNGSearch.d.ts +1 -1
  249. package/dist/tools/search/wikipedia.d.cts +1 -1
  250. package/dist/tools/search/wikipedia.d.ts +1 -1
  251. package/dist/tools/similarity.d.cts +1 -1
  252. package/dist/tools/similarity.d.ts +1 -1
  253. package/dist/tools/think.d.cts +1 -1
  254. package/dist/tools/think.d.ts +1 -1
  255. package/dist/tools/weather/openMeteo.d.cts +1 -1
  256. package/dist/tools/weather/openMeteo.d.ts +1 -1
  257. package/dist/tools/web/webCrawler.d.cts +1 -1
  258. package/dist/tools/web/webCrawler.d.ts +1 -1
  259. package/dist/{types-BXkkbABd.d.ts → types-CJwiMH3u.d.ts} +11 -11
  260. package/dist/{types-Cx6f43du.d.cts → types-W0QttuP_.d.cts} +11 -11
  261. package/dist/version.cjs +1 -1
  262. package/dist/version.js +1 -1
  263. package/dist/workflows/agent.d.cts +3 -3
  264. package/dist/workflows/agent.d.ts +3 -3
  265. package/dist/workflows/workflow.d.cts +1 -1
  266. package/dist/workflows/workflow.d.ts +1 -1
  267. package/package.json +8 -8
@@ -3,13 +3,13 @@ import { GroqClientSettings, GroqClient } from './client.js';
3
3
  import { GroqProvider } from '@ai-sdk/groq';
4
4
  import '../../../backend/embedding.js';
5
5
  import '../../../context.js';
6
- import '../../../emitter-CN7I4uSV.js';
6
+ import '../../../emitter-36-9MnvA.js';
7
7
  import '../../../internals/types.js';
8
8
  import '../../../internals/helpers/guards.js';
9
9
  import '../../../internals/serializable.js';
10
10
  import '../../../internals/helpers/promise.js';
11
11
  import '../../../errors.js';
12
- import '../../../chat-BbCYMAdE.js';
12
+ import '../../../chat-DrqwUgD8.js';
13
13
  import '../../../backend/message.js';
14
14
  import 'ai';
15
15
  import 'promise-based-task';
@@ -1,5 +1,5 @@
1
1
  import { Logger } from '../../../logger/logger.cjs';
2
- import { m as ChatModel, C as ChatModelParameters, h as ChatModelEmitter, d as ChatModelInput, n as ChatModelOutput, a as ChatModelObjectInput, b as ChatModelObjectOutput, i as ChatModelCache, l as ChatModelToolChoiceSupport } from '../../../chat-DB_cf15S.cjs';
2
+ import { m as ChatModel, C as ChatModelParameters, h as ChatModelEmitter, d as ChatModelInput, n as ChatModelOutput, a as ChatModelObjectInput, b as ChatModelObjectOutput, i as ChatModelCache, l as ChatModelToolChoiceSupport } from '../../../chat-BOb9n7AJ.cjs';
3
3
  import { RunContext } from '../../../context.cjs';
4
4
  import { BaseChatModel, BaseChatModelCallOptions } from '@langchain/core/language_models/chat_models';
5
5
  import { BaseMessageLike, AIMessageChunk } from '@langchain/core/messages';
@@ -10,7 +10,7 @@ import '../../../internals/helpers/guards.cjs';
10
10
  import '../../../internals/serializable.cjs';
11
11
  import '../../../backend/message.cjs';
12
12
  import 'ai';
13
- import '../../../emitter-CGKd_-kK.cjs';
13
+ import '../../../emitter-jN55XZZq.cjs';
14
14
  import 'promise-based-task';
15
15
  import '../../../cache/base.cjs';
16
16
  import '../../../backend/constants.cjs';
@@ -1,5 +1,5 @@
1
1
  import { Logger } from '../../../logger/logger.js';
2
- import { m as ChatModel, C as ChatModelParameters, h as ChatModelEmitter, d as ChatModelInput, n as ChatModelOutput, a as ChatModelObjectInput, b as ChatModelObjectOutput, i as ChatModelCache, l as ChatModelToolChoiceSupport } from '../../../chat-BbCYMAdE.js';
2
+ import { m as ChatModel, C as ChatModelParameters, h as ChatModelEmitter, d as ChatModelInput, n as ChatModelOutput, a as ChatModelObjectInput, b as ChatModelObjectOutput, i as ChatModelCache, l as ChatModelToolChoiceSupport } from '../../../chat-DrqwUgD8.js';
3
3
  import { RunContext } from '../../../context.js';
4
4
  import { BaseChatModel, BaseChatModelCallOptions } from '@langchain/core/language_models/chat_models';
5
5
  import { BaseMessageLike, AIMessageChunk } from '@langchain/core/messages';
@@ -10,7 +10,7 @@ import '../../../internals/helpers/guards.js';
10
10
  import '../../../internals/serializable.js';
11
11
  import '../../../backend/message.js';
12
12
  import 'ai';
13
- import '../../../emitter-CN7I4uSV.js';
13
+ import '../../../emitter-36-9MnvA.js';
14
14
  import 'promise-based-task';
15
15
  import '../../../cache/base.js';
16
16
  import '../../../backend/constants.js';
@@ -1,12 +1,12 @@
1
1
  import { EmbeddingModel, EmbeddingModelEvents, EmbeddingModelInput, EmbeddingModelOutput } from '../../../backend/embedding.cjs';
2
2
  import { RunContext } from '../../../context.cjs';
3
- import { E as Emitter } from '../../../emitter-CGKd_-kK.cjs';
3
+ import { E as Emitter } from '../../../emitter-jN55XZZq.cjs';
4
4
  import { Embeddings } from '@langchain/core/embeddings';
5
5
  import '../../../internals/serializable.cjs';
6
6
  import '../../../internals/types.cjs';
7
7
  import '../../../internals/helpers/guards.cjs';
8
8
  import '../../../errors.cjs';
9
- import '../../../chat-DB_cf15S.cjs';
9
+ import '../../../chat-BOb9n7AJ.cjs';
10
10
  import '../../../backend/message.cjs';
11
11
  import 'ai';
12
12
  import 'promise-based-task';
@@ -1,12 +1,12 @@
1
1
  import { EmbeddingModel, EmbeddingModelEvents, EmbeddingModelInput, EmbeddingModelOutput } from '../../../backend/embedding.js';
2
2
  import { RunContext } from '../../../context.js';
3
- import { E as Emitter } from '../../../emitter-CN7I4uSV.js';
3
+ import { E as Emitter } from '../../../emitter-36-9MnvA.js';
4
4
  import { Embeddings } from '@langchain/core/embeddings';
5
5
  import '../../../internals/serializable.js';
6
6
  import '../../../internals/types.js';
7
7
  import '../../../internals/helpers/guards.js';
8
8
  import '../../../errors.js';
9
- import '../../../chat-BbCYMAdE.js';
9
+ import '../../../chat-DrqwUgD8.js';
10
10
  import '../../../backend/message.js';
11
11
  import 'ai';
12
12
  import 'promise-based-task';
@@ -4,7 +4,7 @@ import { BaseToolRunOptions, BaseToolOptions, JSONToolOutput, Tool, ToolEmitter,
4
4
  import { GetRunContext } from '../../context.cjs';
5
5
  import { RunnableConfig } from '@langchain/core/runnables';
6
6
  import * as LCTools from '@langchain/core/tools';
7
- import { E as Emitter } from '../../emitter-CGKd_-kK.cjs';
7
+ import { E as Emitter } from '../../emitter-jN55XZZq.cjs';
8
8
  import '../../internals/serializable.cjs';
9
9
  import '../../internals/types.cjs';
10
10
  import '../../internals/helpers/guards.cjs';
@@ -4,7 +4,7 @@ import { BaseToolRunOptions, BaseToolOptions, JSONToolOutput, Tool, ToolEmitter,
4
4
  import { GetRunContext } from '../../context.js';
5
5
  import { RunnableConfig } from '@langchain/core/runnables';
6
6
  import * as LCTools from '@langchain/core/tools';
7
- import { E as Emitter } from '../../emitter-CN7I4uSV.js';
7
+ import { E as Emitter } from '../../emitter-36-9MnvA.js';
8
8
  import '../../internals/serializable.js';
9
9
  import '../../internals/types.js';
10
10
  import '../../internals/helpers/guards.js';
@@ -10,12 +10,11 @@ class OllamaChatModel extends chat_cjs.VercelChatModel {
10
10
  static {
11
11
  __name(this, "OllamaChatModel");
12
12
  }
13
- supportsToolStreaming = false;
14
13
  toolChoiceSupport = [
15
14
  "none",
16
15
  "auto"
17
16
  ];
18
- constructor(modelId = env_cjs.getEnv("OLLAMA_CHAT_MODEL", "llama3.1:8b"), settings = {}, client) {
17
+ constructor(modelId = env_cjs.getEnv("OLLAMA_CHAT_MODEL", "granite4:micro"), settings = {}, client) {
19
18
  const model = client_cjs.OllamaClient.ensure(client).instance.chat(modelId, settings);
20
19
  super(model);
21
20
  }
@@ -1 +1 @@
1
- {"version":3,"sources":["../../../../src/adapters/ollama/backend/chat.ts"],"names":["OllamaChatModel","VercelChatModel","supportsToolStreaming","toolChoiceSupport","modelId","getEnv","settings","client","model","OllamaClient","ensure","instance","chat","register"],"mappings":";;;;;;;;AAeO,MAAMA,wBAAwBC,wBAAAA,CAAAA;EAfrC;;;EAgBWC,qBAAAA,GAAwB,KAAA;EACjBC,iBAAAA,GAAkD;AAAC,IAAA,MAAA;AAAQ,IAAA;;EAE3E,WAAA,CACEC,OAAAA,GAA6BC,eAAO,mBAAA,EAAqB,aAAA,GACzDC,QAAAA,GAAoC,IACpCC,MAAAA,EACA;AACA,IAAA,MAAMC,KAAAA,GAAQC,wBAAaC,MAAAA,CAAOH,MAAAA,EAAQI,QAAAA,CAASC,IAAAA,CAAKR,SAASE,QAAAA,CAAAA;AACjE,IAAA,KAAA,CAAME,KAAAA,CAAAA;AACR;EAEA;AACE,IAAA,IAAA,CAAKK,QAAAA,EAAQ;AACf;AACF","file":"chat.cjs","sourcesContent":["/**\n * Copyright 2025 © BeeAI a Series of LF Projects, LLC\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport { VercelChatModel } from \"@/adapters/vercel/backend/chat.js\";\nimport { OllamaProvider } from \"ollama-ai-provider-v2\";\nimport { OllamaClient, OllamaClientSettings } from \"@/adapters/ollama/backend/client.js\";\nimport { getEnv } from \"@/internals/env.js\";\nimport { ChatModelToolChoiceSupport } from \"@/backend/chat.js\";\n\ntype OllamaParameters = Parameters<OllamaProvider[\"chat\"]>;\nexport type OllamaChatModelId = NonNullable<OllamaParameters[0]>;\nexport type OllamaChatModelSettings = NonNullable<OllamaParameters[1]>;\n\nexport class OllamaChatModel extends VercelChatModel {\n readonly supportsToolStreaming = false;\n public readonly toolChoiceSupport: ChatModelToolChoiceSupport[] = [\"none\", \"auto\"];\n\n constructor(\n modelId: OllamaChatModelId = getEnv(\"OLLAMA_CHAT_MODEL\", \"llama3.1:8b\"),\n settings: OllamaChatModelSettings = {},\n client?: OllamaClient | OllamaClientSettings,\n ) {\n const model = OllamaClient.ensure(client).instance.chat(modelId, settings);\n super(model);\n }\n\n static {\n this.register();\n }\n}\n"]}
1
+ {"version":3,"sources":["../../../../src/adapters/ollama/backend/chat.ts"],"names":["OllamaChatModel","VercelChatModel","toolChoiceSupport","modelId","getEnv","settings","client","model","OllamaClient","ensure","instance","chat","register"],"mappings":";;;;;;;;AAeO,MAAMA,wBAAwBC,wBAAAA,CAAAA;EAfrC;;;EAgBkBC,iBAAAA,GAAkD;AAAC,IAAA,MAAA;AAAQ,IAAA;;EAE3E,WAAA,CACEC,OAAAA,GAA6BC,eAAO,mBAAA,EAAqB,gBAAA,GACzDC,QAAAA,GAAoC,IACpCC,MAAAA,EACA;AACA,IAAA,MAAMC,KAAAA,GAAQC,wBAAaC,MAAAA,CAAOH,MAAAA,EAAQI,QAAAA,CAASC,IAAAA,CAAKR,SAASE,QAAAA,CAAAA;AACjE,IAAA,KAAA,CAAME,KAAAA,CAAAA;AACR;EAEA;AACE,IAAA,IAAA,CAAKK,QAAAA,EAAQ;AACf;AACF","file":"chat.cjs","sourcesContent":["/**\n * Copyright 2025 © BeeAI a Series of LF Projects, LLC\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport { VercelChatModel } from \"@/adapters/vercel/backend/chat.js\";\nimport { OllamaProvider } from \"ollama-ai-provider-v2\";\nimport { OllamaClient, OllamaClientSettings } from \"@/adapters/ollama/backend/client.js\";\nimport { getEnv } from \"@/internals/env.js\";\nimport { ChatModelToolChoiceSupport } from \"@/backend/chat.js\";\n\ntype OllamaParameters = Parameters<OllamaProvider[\"chat\"]>;\nexport type OllamaChatModelId = NonNullable<OllamaParameters[0]>;\nexport type OllamaChatModelSettings = NonNullable<OllamaParameters[1]>;\n\nexport class OllamaChatModel extends VercelChatModel {\n public readonly toolChoiceSupport: ChatModelToolChoiceSupport[] = [\"none\", \"auto\"];\n\n constructor(\n modelId: OllamaChatModelId = getEnv(\"OLLAMA_CHAT_MODEL\", \"granite4:micro\"),\n settings: OllamaChatModelSettings = {},\n client?: OllamaClient | OllamaClientSettings,\n ) {\n const model = OllamaClient.ensure(client).instance.chat(modelId, settings);\n super(model);\n }\n\n static {\n this.register();\n }\n}\n"]}
@@ -1,7 +1,7 @@
1
1
  import { VercelChatModel } from '../../vercel/backend/chat.cjs';
2
2
  import { OllamaProvider } from 'ollama-ai-provider-v2';
3
3
  import { OllamaClient, OllamaClientSettings } from './client.cjs';
4
- import { l as ChatModelToolChoiceSupport } from '../../../chat-DB_cf15S.cjs';
4
+ import { l as ChatModelToolChoiceSupport } from '../../../chat-BOb9n7AJ.cjs';
5
5
  import '../../../logger/logger.cjs';
6
6
  import 'pino';
7
7
  import '../../../errors.cjs';
@@ -9,7 +9,7 @@ import '../../../internals/types.cjs';
9
9
  import '../../../internals/helpers/guards.cjs';
10
10
  import '../../../internals/serializable.cjs';
11
11
  import 'ai';
12
- import '../../../emitter-CGKd_-kK.cjs';
12
+ import '../../../emitter-jN55XZZq.cjs';
13
13
  import '../../../backend/message.cjs';
14
14
  import '../../../context.cjs';
15
15
  import '../../../internals/helpers/promise.cjs';
@@ -33,7 +33,6 @@ type OllamaParameters = Parameters<OllamaProvider["chat"]>;
33
33
  type OllamaChatModelId = NonNullable<OllamaParameters[0]>;
34
34
  type OllamaChatModelSettings = NonNullable<OllamaParameters[1]>;
35
35
  declare class OllamaChatModel extends VercelChatModel {
36
- readonly supportsToolStreaming = false;
37
36
  readonly toolChoiceSupport: ChatModelToolChoiceSupport[];
38
37
  constructor(modelId?: OllamaChatModelId, settings?: OllamaChatModelSettings, client?: OllamaClient | OllamaClientSettings);
39
38
  }
@@ -1,7 +1,7 @@
1
1
  import { VercelChatModel } from '../../vercel/backend/chat.js';
2
2
  import { OllamaProvider } from 'ollama-ai-provider-v2';
3
3
  import { OllamaClient, OllamaClientSettings } from './client.js';
4
- import { l as ChatModelToolChoiceSupport } from '../../../chat-BbCYMAdE.js';
4
+ import { l as ChatModelToolChoiceSupport } from '../../../chat-DrqwUgD8.js';
5
5
  import '../../../logger/logger.js';
6
6
  import 'pino';
7
7
  import '../../../errors.js';
@@ -9,7 +9,7 @@ import '../../../internals/types.js';
9
9
  import '../../../internals/helpers/guards.js';
10
10
  import '../../../internals/serializable.js';
11
11
  import 'ai';
12
- import '../../../emitter-CN7I4uSV.js';
12
+ import '../../../emitter-36-9MnvA.js';
13
13
  import '../../../backend/message.js';
14
14
  import '../../../context.js';
15
15
  import '../../../internals/helpers/promise.js';
@@ -33,7 +33,6 @@ type OllamaParameters = Parameters<OllamaProvider["chat"]>;
33
33
  type OllamaChatModelId = NonNullable<OllamaParameters[0]>;
34
34
  type OllamaChatModelSettings = NonNullable<OllamaParameters[1]>;
35
35
  declare class OllamaChatModel extends VercelChatModel {
36
- readonly supportsToolStreaming = false;
37
36
  readonly toolChoiceSupport: ChatModelToolChoiceSupport[];
38
37
  constructor(modelId?: OllamaChatModelId, settings?: OllamaChatModelSettings, client?: OllamaClient | OllamaClientSettings);
39
38
  }
@@ -8,12 +8,11 @@ class OllamaChatModel extends VercelChatModel {
8
8
  static {
9
9
  __name(this, "OllamaChatModel");
10
10
  }
11
- supportsToolStreaming = false;
12
11
  toolChoiceSupport = [
13
12
  "none",
14
13
  "auto"
15
14
  ];
16
- constructor(modelId = getEnv("OLLAMA_CHAT_MODEL", "llama3.1:8b"), settings = {}, client) {
15
+ constructor(modelId = getEnv("OLLAMA_CHAT_MODEL", "granite4:micro"), settings = {}, client) {
17
16
  const model = OllamaClient.ensure(client).instance.chat(modelId, settings);
18
17
  super(model);
19
18
  }
@@ -1 +1 @@
1
- {"version":3,"sources":["../../../../src/adapters/ollama/backend/chat.ts"],"names":["OllamaChatModel","VercelChatModel","supportsToolStreaming","toolChoiceSupport","modelId","getEnv","settings","client","model","OllamaClient","ensure","instance","chat","register"],"mappings":";;;;;;AAeO,MAAMA,wBAAwBC,eAAAA,CAAAA;EAfrC;;;EAgBWC,qBAAAA,GAAwB,KAAA;EACjBC,iBAAAA,GAAkD;AAAC,IAAA,MAAA;AAAQ,IAAA;;EAE3E,WAAA,CACEC,OAAAA,GAA6BC,OAAO,mBAAA,EAAqB,aAAA,GACzDC,QAAAA,GAAoC,IACpCC,MAAAA,EACA;AACA,IAAA,MAAMC,KAAAA,GAAQC,aAAaC,MAAAA,CAAOH,MAAAA,EAAQI,QAAAA,CAASC,IAAAA,CAAKR,SAASE,QAAAA,CAAAA;AACjE,IAAA,KAAA,CAAME,KAAAA,CAAAA;AACR;EAEA;AACE,IAAA,IAAA,CAAKK,QAAAA,EAAQ;AACf;AACF","file":"chat.js","sourcesContent":["/**\n * Copyright 2025 © BeeAI a Series of LF Projects, LLC\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport { VercelChatModel } from \"@/adapters/vercel/backend/chat.js\";\nimport { OllamaProvider } from \"ollama-ai-provider-v2\";\nimport { OllamaClient, OllamaClientSettings } from \"@/adapters/ollama/backend/client.js\";\nimport { getEnv } from \"@/internals/env.js\";\nimport { ChatModelToolChoiceSupport } from \"@/backend/chat.js\";\n\ntype OllamaParameters = Parameters<OllamaProvider[\"chat\"]>;\nexport type OllamaChatModelId = NonNullable<OllamaParameters[0]>;\nexport type OllamaChatModelSettings = NonNullable<OllamaParameters[1]>;\n\nexport class OllamaChatModel extends VercelChatModel {\n readonly supportsToolStreaming = false;\n public readonly toolChoiceSupport: ChatModelToolChoiceSupport[] = [\"none\", \"auto\"];\n\n constructor(\n modelId: OllamaChatModelId = getEnv(\"OLLAMA_CHAT_MODEL\", \"llama3.1:8b\"),\n settings: OllamaChatModelSettings = {},\n client?: OllamaClient | OllamaClientSettings,\n ) {\n const model = OllamaClient.ensure(client).instance.chat(modelId, settings);\n super(model);\n }\n\n static {\n this.register();\n }\n}\n"]}
1
+ {"version":3,"sources":["../../../../src/adapters/ollama/backend/chat.ts"],"names":["OllamaChatModel","VercelChatModel","toolChoiceSupport","modelId","getEnv","settings","client","model","OllamaClient","ensure","instance","chat","register"],"mappings":";;;;;;AAeO,MAAMA,wBAAwBC,eAAAA,CAAAA;EAfrC;;;EAgBkBC,iBAAAA,GAAkD;AAAC,IAAA,MAAA;AAAQ,IAAA;;EAE3E,WAAA,CACEC,OAAAA,GAA6BC,OAAO,mBAAA,EAAqB,gBAAA,GACzDC,QAAAA,GAAoC,IACpCC,MAAAA,EACA;AACA,IAAA,MAAMC,KAAAA,GAAQC,aAAaC,MAAAA,CAAOH,MAAAA,EAAQI,QAAAA,CAASC,IAAAA,CAAKR,SAASE,QAAAA,CAAAA;AACjE,IAAA,KAAA,CAAME,KAAAA,CAAAA;AACR;EAEA;AACE,IAAA,IAAA,CAAKK,QAAAA,EAAQ;AACf;AACF","file":"chat.js","sourcesContent":["/**\n * Copyright 2025 © BeeAI a Series of LF Projects, LLC\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport { VercelChatModel } from \"@/adapters/vercel/backend/chat.js\";\nimport { OllamaProvider } from \"ollama-ai-provider-v2\";\nimport { OllamaClient, OllamaClientSettings } from \"@/adapters/ollama/backend/client.js\";\nimport { getEnv } from \"@/internals/env.js\";\nimport { ChatModelToolChoiceSupport } from \"@/backend/chat.js\";\n\ntype OllamaParameters = Parameters<OllamaProvider[\"chat\"]>;\nexport type OllamaChatModelId = NonNullable<OllamaParameters[0]>;\nexport type OllamaChatModelSettings = NonNullable<OllamaParameters[1]>;\n\nexport class OllamaChatModel extends VercelChatModel {\n public readonly toolChoiceSupport: ChatModelToolChoiceSupport[] = [\"none\", \"auto\"];\n\n constructor(\n modelId: OllamaChatModelId = getEnv(\"OLLAMA_CHAT_MODEL\", \"granite4:micro\"),\n settings: OllamaChatModelSettings = {},\n client?: OllamaClient | OllamaClientSettings,\n ) {\n const model = OllamaClient.ensure(client).instance.chat(modelId, settings);\n super(model);\n }\n\n static {\n this.register();\n }\n}\n"]}
@@ -7,10 +7,10 @@ import '../../../internals/types.cjs';
7
7
  import '../../../internals/helpers/guards.cjs';
8
8
  import '../../../backend/embedding.cjs';
9
9
  import '../../../context.cjs';
10
- import '../../../emitter-CGKd_-kK.cjs';
10
+ import '../../../emitter-jN55XZZq.cjs';
11
11
  import '../../../internals/helpers/promise.cjs';
12
12
  import '../../../errors.cjs';
13
- import '../../../chat-DB_cf15S.cjs';
13
+ import '../../../chat-BOb9n7AJ.cjs';
14
14
  import '../../../backend/message.cjs';
15
15
  import 'ai';
16
16
  import 'promise-based-task';
@@ -7,10 +7,10 @@ import '../../../internals/types.js';
7
7
  import '../../../internals/helpers/guards.js';
8
8
  import '../../../backend/embedding.js';
9
9
  import '../../../context.js';
10
- import '../../../emitter-CN7I4uSV.js';
10
+ import '../../../emitter-36-9MnvA.js';
11
11
  import '../../../internals/helpers/promise.js';
12
12
  import '../../../errors.js';
13
- import '../../../chat-BbCYMAdE.js';
13
+ import '../../../chat-DrqwUgD8.js';
14
14
  import '../../../backend/message.js';
15
15
  import 'ai';
16
16
  import 'promise-based-task';
@@ -1,7 +1,7 @@
1
1
  import { OpenAIProvider } from '@ai-sdk/openai';
2
2
  import { OpenAIClient, OpenAIClientSettings } from './client.cjs';
3
3
  import { VercelChatModel } from '../../vercel/backend/chat.cjs';
4
- import { C as ChatModelParameters } from '../../../chat-DB_cf15S.cjs';
4
+ import { C as ChatModelParameters } from '../../../chat-BOb9n7AJ.cjs';
5
5
  import '../../../backend/client.cjs';
6
6
  import '../../../internals/serializable.cjs';
7
7
  import '../../../internals/types.cjs';
@@ -10,7 +10,7 @@ import '../../../logger/logger.cjs';
10
10
  import 'pino';
11
11
  import '../../../errors.cjs';
12
12
  import 'ai';
13
- import '../../../emitter-CGKd_-kK.cjs';
13
+ import '../../../emitter-jN55XZZq.cjs';
14
14
  import '../../../backend/message.cjs';
15
15
  import '../../../context.cjs';
16
16
  import '../../../internals/helpers/promise.cjs';
@@ -1,7 +1,7 @@
1
1
  import { OpenAIProvider } from '@ai-sdk/openai';
2
2
  import { OpenAIClient, OpenAIClientSettings } from './client.js';
3
3
  import { VercelChatModel } from '../../vercel/backend/chat.js';
4
- import { C as ChatModelParameters } from '../../../chat-BbCYMAdE.js';
4
+ import { C as ChatModelParameters } from '../../../chat-DrqwUgD8.js';
5
5
  import '../../../backend/client.js';
6
6
  import '../../../internals/serializable.js';
7
7
  import '../../../internals/types.js';
@@ -10,7 +10,7 @@ import '../../../logger/logger.js';
10
10
  import 'pino';
11
11
  import '../../../errors.js';
12
12
  import 'ai';
13
- import '../../../emitter-CN7I4uSV.js';
13
+ import '../../../emitter-36-9MnvA.js';
14
14
  import '../../../backend/message.js';
15
15
  import '../../../context.js';
16
16
  import '../../../internals/helpers/promise.js';
@@ -7,10 +7,10 @@ import '../../../internals/types.cjs';
7
7
  import '../../../internals/helpers/guards.cjs';
8
8
  import '../../../backend/embedding.cjs';
9
9
  import '../../../context.cjs';
10
- import '../../../emitter-CGKd_-kK.cjs';
10
+ import '../../../emitter-jN55XZZq.cjs';
11
11
  import '../../../internals/helpers/promise.cjs';
12
12
  import '../../../errors.cjs';
13
- import '../../../chat-DB_cf15S.cjs';
13
+ import '../../../chat-BOb9n7AJ.cjs';
14
14
  import '../../../backend/message.cjs';
15
15
  import 'ai';
16
16
  import 'promise-based-task';
@@ -7,10 +7,10 @@ import '../../../internals/types.js';
7
7
  import '../../../internals/helpers/guards.js';
8
8
  import '../../../backend/embedding.js';
9
9
  import '../../../context.js';
10
- import '../../../emitter-CN7I4uSV.js';
10
+ import '../../../emitter-36-9MnvA.js';
11
11
  import '../../../internals/helpers/promise.js';
12
12
  import '../../../errors.js';
13
- import '../../../chat-BbCYMAdE.js';
13
+ import '../../../chat-DrqwUgD8.js';
14
14
  import '../../../backend/message.js';
15
15
  import 'ai';
16
16
  import 'promise-based-task';
@@ -50,7 +50,11 @@ class VercelChatModel extends chat_cjs.ChatModel {
50
50
  }, run);
51
51
  return output;
52
52
  }
53
- const { finishReason, usage, response: { messages } } = await ai.generateText(await this.transformInput(input));
53
+ const { finishReason, usage, response: { messages } } = await ai.generateText({
54
+ temperature: 0,
55
+ ...await this.transformInput(input),
56
+ abortSignal: run.signal
57
+ });
54
58
  return new chat_cjs.ChatModelOutput(this.transformMessages(messages), utils_cjs.extractTokenUsage(usage), finishReason);
55
59
  }
56
60
  async _createStructure({ schema, ...input }, run) {
@@ -58,7 +62,6 @@ class VercelChatModel extends chat_cjs.ChatModel {
58
62
  temperature: 0,
59
63
  ...await this.transformInput(input),
60
64
  abortSignal: run.signal,
61
- model: this.model,
62
65
  ...schema instanceof zod.ZodSchema ? {
63
66
  schema,
64
67
  output: (schema._input || schema) instanceof zod.ZodArray ? "array" : (schema._input || schema) instanceof zod.ZodEnum ? "enum" : "object"
@@ -76,6 +79,15 @@ class VercelChatModel extends chat_cjs.ChatModel {
76
79
  };
77
80
  }
78
81
  async *_createStream(input, run) {
82
+ const responseFormat = input.responseFormat;
83
+ if (responseFormat && (responseFormat instanceof zod.ZodSchema || responseFormat.schema)) {
84
+ const { output } = await this._createStructure({
85
+ ...input,
86
+ schema: responseFormat
87
+ }, run);
88
+ yield output;
89
+ return;
90
+ }
79
91
  if (!this.supportsToolStreaming && !remeda.isEmpty(input.tools ?? [])) {
80
92
  const response = await this._create(input, run);
81
93
  yield response;
@@ -85,45 +97,91 @@ class VercelChatModel extends chat_cjs.ChatModel {
85
97
  ...await this.transformInput(input),
86
98
  abortSignal: run.signal
87
99
  });
88
- let lastChunk = null;
100
+ let streamEmpty = true;
101
+ const streamedToolCalls = /* @__PURE__ */ new Map();
89
102
  for await (const event of fullStream) {
90
103
  let message;
91
104
  switch (event.type) {
92
105
  case "text-delta":
93
- message = new message_cjs.AssistantMessage(event.text);
106
+ streamEmpty = false;
107
+ message = new message_cjs.AssistantMessage(event.text, {}, event.id);
108
+ yield new chat_cjs.ChatModelOutput([
109
+ message
110
+ ]);
111
+ break;
112
+ case "text-end":
113
+ streamEmpty = false;
114
+ break;
115
+ case "tool-input-start": {
116
+ if (!input.streamPartialToolCalls) {
117
+ break;
118
+ }
119
+ const chunk = {
120
+ type: "tool-call",
121
+ toolName: event.toolName,
122
+ toolCallId: event.id,
123
+ input: ""
124
+ };
125
+ streamedToolCalls.set(event.id, chunk);
126
+ const message2 = new message_cjs.AssistantMessage(chunk, {}, event.id);
127
+ yield new chat_cjs.ChatModelOutput([
128
+ message2
129
+ ]);
130
+ break;
131
+ }
132
+ case "tool-input-delta": {
133
+ if (!input.streamPartialToolCalls) {
134
+ break;
135
+ }
136
+ if (event.delta) {
137
+ const chunk = streamedToolCalls.get(event.id);
138
+ const message2 = new message_cjs.AssistantMessage({
139
+ ...chunk,
140
+ input: event.delta
141
+ }, {}, event.id);
142
+ yield new chat_cjs.ChatModelOutput([
143
+ message2
144
+ ]);
145
+ }
94
146
  break;
95
- case "tool-call":
147
+ }
148
+ case "tool-call": {
149
+ streamEmpty = false;
150
+ const existingToolCall = streamedToolCalls.get(event.toolCallId);
151
+ if (existingToolCall) {
152
+ streamedToolCalls.delete(event.toolCallId);
153
+ break;
154
+ }
96
155
  message = new message_cjs.AssistantMessage({
97
156
  type: event.type,
98
157
  toolCallId: event.toolCallId,
99
158
  toolName: event.toolName,
100
159
  input: event.input
101
- });
160
+ }, {}, event.toolCallId);
161
+ yield new chat_cjs.ChatModelOutput([
162
+ message
163
+ ]);
102
164
  break;
165
+ }
103
166
  case "error":
104
167
  throw new errors_cjs$1.ChatModelError("Unhandled error", [
105
168
  event.error
106
169
  ]);
107
170
  case "tool-result":
171
+ streamEmpty = false;
108
172
  message = new message_cjs.ToolMessage({
109
173
  type: event.type,
110
174
  toolCallId: event.toolCallId,
111
175
  toolName: event.toolName,
112
176
  output: event.output
113
- });
114
- break;
115
- case "abort":
116
- message = new message_cjs.AssistantMessage([]);
177
+ }, {}, `tool_result_${event.toolCallId}`);
178
+ yield new chat_cjs.ChatModelOutput([
179
+ message
180
+ ]);
117
181
  break;
118
- default:
119
- continue;
120
182
  }
121
- lastChunk = new chat_cjs.ChatModelOutput([
122
- message
123
- ]);
124
- yield lastChunk;
125
183
  }
126
- if (!lastChunk) {
184
+ if (streamEmpty) {
127
185
  throw new errors_cjs$1.ChatModelError("No chunks have been received!");
128
186
  }
129
187
  try {
@@ -132,8 +190,10 @@ class VercelChatModel extends chat_cjs.ChatModel {
132
190
  finishReasonPromise,
133
191
  responsePromise
134
192
  ]);
193
+ const lastChunk = new chat_cjs.ChatModelOutput([]);
135
194
  lastChunk.usage = utils_cjs.extractTokenUsage(usage);
136
195
  lastChunk.finishReason = finishReason;
196
+ yield lastChunk;
137
197
  } catch (e) {
138
198
  if (!run.signal.aborted) {
139
199
  throw e;
@@ -1 +1 @@
1
- {"version":3,"sources":["../../../../src/adapters/vercel/backend/chat.ts"],"names":["VercelChatModel","ChatModel","emitter","supportsToolStreaming","model","modelId","ValueError","Emitter","root","child","namespace","providerId","creator","provider","split","toCamelCase","_create","input","run","responseFormat","ZodSchema","schema","output","_createStructure","finishReason","usage","response","messages","generateText","transformInput","ChatModelOutput","transformMessages","extractTokenUsage","generateObject","temperature","abortSignal","signal","_input","ZodArray","ZodEnum","jsonSchema","z","any","schemaName","name","schemaDescription","description","object","AssistantMessage","JSON","stringify","_createStream","isEmpty","tools","fullStream","usagePromise","finishReasonPromise","responsePromise","streamText","lastChunk","event","message","type","text","toolCallId","toolName","ChatModelError","error","ToolMessage","_","Promise","all","e","aborted","map","tool","inputSchema","getInputJsonSchema","msg","CustomMessage","encodeCustomMessage","role","content","UserMessage","SystemMessage","part","join","toolChoice","Tool","toolChoiceSupport","includes","logger","warn","parameters","mapToObj","flatMap","providerOptions","createSnapshot","loadSnapshot","snapshot","instance","fromName","Error","destroy","Object","assign"],"mappings":";;;;;;;;;;;;;;;AA6CO,MAAeA,wBAEZC,kBAAAA,CAAAA;EA/CV;;;;AAgDkBC,EAAAA,OAAAA;EACAC,qBAAAA,GAAiC,IAAA;AAEjD,EAAA,WAAA,CAA6BC,KAAAA,EAAU;AACrC,IAAA,KAAA,EAAK,EAAA,KADsBA,KAAAA,GAAAA,KAAAA;AAE3B,IAAA,IAAI,CAAC,KAAKC,OAAAA,EAAS;AACjB,MAAA,MAAM,IAAIC,sBAAW,+BAAA,CAAA;AACvB;AACA,IAAA,IAAA,CAAKJ,OAAAA,GAAUK,mBAAAA,CAAQC,IAAAA,CAAKC,KAAAA,CAAM;MAChCC,SAAAA,EAAW;AAAC,QAAA,SAAA;QAAW,IAAA,CAAKC,UAAAA;AAAY,QAAA;;MACxCC,OAAAA,EAAS;KACX,CAAA;AACF;AAEA,EAAA,IAAIP,OAAAA,GAAkB;AACpB,IAAA,OAAO,KAAKD,KAAAA,CAAMC,OAAAA;AACpB;AAEA,EAAA,IAAIM,UAAAA,GAAqB;AACvB,IAAA,MAAME,QAAAA,GAAW,IAAA,CAAKT,KAAAA,CAAMS,QAAAA,CAASC,KAAAA,CAAM,GAAA,CAAA,CAAK,CAAA,CAAA,CAAGA,KAAAA,CAAM,GAAA,CAAA,CAAK,CAAA,CAAA;AAC9D,IAAA,OAAOC,mBAAYF,QAAAA,CAAAA;AACrB;EAEA,MAAgBG,OAAAA,CAAQC,OAAuBC,GAAAA,EAA0B;AACvE,IAAA,MAAMC,iBAAiBF,KAAAA,CAAME,cAAAA;AAC7B,IAAA,IAAIA,cAAAA,KAAmBA,cAAAA,YAA0BC,aAAAA,IAAaD,cAAAA,CAAeE,MAAAA,CAAAA,EAAS;AACpF,MAAA,MAAM,EAAEC,MAAAA,EAAM,GAAK,MAAM,KAAKC,gBAAAA,CAC5B;QACE,GAAGN,KAAAA;QACHI,MAAAA,EAAQF;AACV,OAAA,EACAD,GAAAA,CAAAA;AAEF,MAAA,OAAOI,MAAAA;AACT;AAEA,IAAA,MAAM,EACJE,YAAAA,EACAC,KAAAA,EACAC,QAAAA,EAAU,EAAEC,QAAAA,EAAQ,EAAE,GACpB,MAAMC,eAAAA,CAAa,MAAM,IAAA,CAAKC,cAAAA,CAAeZ,KAAAA,CAAAA,CAAAA;AAEjD,IAAA,OAAO,IAAIa,yBACT,IAAA,CAAKC,iBAAAA,CAAkBJ,QAAAA,CAAAA,EACvBK,2BAAAA,CAAkBP,KAAAA,CAAAA,EAClBD,YAAAA,CAAAA;AAEJ;AAEA,EAAA,MAAgBD,iBACd,EAAEF,MAAAA,EAAQ,GAAGJ,KAAAA,IACbC,GAAAA,EACmC;AACnC,IAAA,MAAMQ,QAAAA,GAAW,MAAMO,iBAAAA,CAAe;MACpCC,WAAAA,EAAa,CAAA;MACb,GAAI,MAAM,IAAA,CAAKL,cAAAA,CAAeZ,KAAAA,CAAAA;AAC9BkB,MAAAA,WAAAA,EAAajB,GAAAA,CAAIkB,MAAAA;AACjBhC,MAAAA,KAAAA,EAAO,IAAA,CAAKA,KAAAA;AACZ,MAAA,GAAIiB,kBAAkBD,aAAAA,GAClB;AACEC,QAAAA,MAAAA;QACAC,MAAAA,EAAAA,CAAUD,MAAAA,CAAOgB,UAAUhB,MAAAA,aAAmBiB,YAAAA,GAC1C,WACCjB,MAAAA,CAAOgB,MAAAA,IAAUhB,MAAAA,aAAmBkB,WAAAA,GACnC,MAAA,GACA;OACR,GACA;AACElB,QAAAA,MAAAA,EAAQA,OAAOA,MAAAA,GAASmB,aAAAA,CAAcnB,OAAOA,MAAM,CAAA,GAAIoB,MAAEC,GAAAA,EAAG;AAC5DC,QAAAA,UAAAA,EAAYtB,MAAAA,CAAOuB,IAAAA;AACnBC,QAAAA,iBAAAA,EAAmBxB,MAAAA,CAAOyB;AAC5B;KACN,CAAA;AAEA,IAAA,OAAO;AACLC,MAAAA,MAAAA,EAAQrB,QAAAA,CAASqB,MAAAA;AACjBzB,MAAAA,MAAAA,EAAQ,IAAIQ,wBAAAA,CACV;AAAC,QAAA,IAAIkB,6BAAiBC,IAAAA,CAAKC,SAAAA,CAAUxB,SAASqB,MAAAA,EAAQ,IAAA,EAAM,CAAA,CAAA;AAC5Df,OAAAA,EAAAA,2BAAAA,CAAkBN,QAAAA,CAASD,KAAK,CAAA,EAChCC,QAAAA,CAASF,YAAY;AAEzB,KAAA;AACF;EAEA,OAAO2B,aAAAA,CAAclC,OAAuBC,GAAAA,EAA0B;AACpE,IAAA,IAAI,CAAC,KAAKf,qBAAAA,IAAyB,CAACiD,eAAQnC,KAAAA,CAAMoC,KAAAA,IAAS,EAAE,CAAA,EAAG;AAC9D,MAAA,MAAM3B,QAAAA,GAAW,MAAM,IAAA,CAAKV,OAAAA,CAAQC,OAAOC,GAAAA,CAAAA;AAC3C,MAAA,MAAMQ,QAAAA;AACN,MAAA;AACF;AAEA,IAAA,MAAM,EACJ4B,YACA7B,KAAAA,EAAO8B,YAAAA,EACP/B,cAAcgC,mBAAAA,EACd9B,QAAAA,EAAU+B,eAAAA,EAAe,GACvBC,aAAAA,CAAW;MACb,GAAI,MAAM,IAAA,CAAK7B,cAAAA,CAAeZ,KAAAA,CAAAA;AAC9BkB,MAAAA,WAAAA,EAAajB,GAAAA,CAAIkB;KACnB,CAAA;AAEA,IAAA,IAAIuB,SAAAA,GAAoC,IAAA;AACxC,IAAA,WAAA,MAAiBC,SAASN,UAAAA,EAAY;AACpC,MAAA,IAAIO,OAAAA;AACJ,MAAA,QAAQD,MAAME,IAAAA;QACZ,KAAK,YAAA;AACHD,UAAAA,OAAAA,GAAU,IAAIb,4BAAAA,CAAiBY,KAAAA,CAAMG,IAAI,CAAA;AACzC,UAAA;QACF,KAAK,WAAA;AACHF,UAAAA,OAAAA,GAAU,IAAIb,4BAAAA,CAAiB;AAC7Bc,YAAAA,IAAAA,EAAMF,KAAAA,CAAME,IAAAA;AACZE,YAAAA,UAAAA,EAAYJ,KAAAA,CAAMI,UAAAA;AAClBC,YAAAA,QAAAA,EAAUL,KAAAA,CAAMK,QAAAA;AAChBhD,YAAAA,KAAAA,EAAO2C,KAAAA,CAAM3C;WACf,CAAA;AACA,UAAA;QACF,KAAK,OAAA;AACH,UAAA,MAAM,IAAIiD,4BAAe,iBAAA,EAAmB;YAACN,KAAAA,CAAMO;AAAe,WAAA,CAAA;QACpE,KAAK,aAAA;AACHN,UAAAA,OAAAA,GAAU,IAAIO,uBAAAA,CAAY;AACxBN,YAAAA,IAAAA,EAAMF,KAAAA,CAAME,IAAAA;AACZE,YAAAA,UAAAA,EAAYJ,KAAAA,CAAMI,UAAAA;AAClBC,YAAAA,QAAAA,EAAUL,KAAAA,CAAMK,QAAAA;AAChB3C,YAAAA,MAAAA,EAAQsC,KAAAA,CAAMtC;WAChB,CAAA;AACA,UAAA;QACF,KAAK,OAAA;AACHuC,UAAAA,OAAAA,GAAU,IAAIb,4BAAAA,CAAiB,EAAE,CAAA;AACjC,UAAA;AACF,QAAA;AACE,UAAA;AACJ;AACAW,MAAAA,SAAAA,GAAY,IAAI7B,wBAAAA,CAAgB;AAAC+B,QAAAA;AAAQ,OAAA,CAAA;AACzC,MAAA,MAAMF,SAAAA;AACR;AAEA,IAAA,IAAI,CAACA,SAAAA,EAAW;AACd,MAAA,MAAM,IAAIO,4BAAe,+BAAA,CAAA;AAC3B;AAEA,IAAA,IAAI;AACF,MAAA,MAAM,CAACzC,KAAAA,EAAOD,YAAAA,EAAc6C,CAAAA,CAAAA,GAAK,MAAMC,QAAQC,GAAAA,CAAI;AACjDhB,QAAAA,YAAAA;AACAC,QAAAA,mBAAAA;AACAC,QAAAA;AACD,OAAA,CAAA;AACDE,MAAAA,SAAAA,CAAUlC,KAAAA,GAAQO,4BAAkBP,KAAAA,CAAAA;AACpCkC,MAAAA,SAAAA,CAAUnC,YAAAA,GAAeA,YAAAA;AAC3B,KAAA,CAAA,OAASgD,CAAAA,EAAG;AACV,MAAA,IAAI,CAACtD,GAAAA,CAAIkB,MAAAA,CAAOqC,OAAAA,EAAS;AACvB,QAAA,MAAMD,CAAAA;AACR;AACF;AACF;AAEA,EAAA,MAAgB3C,eACdZ,KAAAA,EACkE;AAClE,IAAA,MAAMoC,KAAAA,GAAQ,MAAMiB,OAAAA,CAAQC,GAAAA,CAAAA,CACzBtD,KAAAA,CAAMoC,SAAS,EAAA,EAAIqB,GAAAA,CAAI,OAAOC,IAAAA,MAAU;AACvC/B,MAAAA,IAAAA,EAAM+B,IAAAA,CAAK/B,IAAAA;AACXE,MAAAA,WAAAA,EAAa6B,IAAAA,CAAK7B,WAAAA;AAClB8B,MAAAA,WAAAA,EAAapC,aAAAA,CAAW,MAAMmC,IAAAA,CAAKE,kBAAAA,EAAkB;AACvD,KAAA,CAAA,CAAA,CAAA;AAGF,IAAA,MAAMlD,QAAAA,GAAWV,KAAAA,CAAMU,QAAAA,CAAS+C,GAAAA,CAAI,CAACI,GAAAA,KAAAA;AACnC,MAAA,IAAIA,eAAeC,yBAAAA,EAAe;AAChCD,QAAAA,GAAAA,GAAME,8BAAoBF,GAAAA,CAAAA;AAC5B;AAEA,MAAA,IAAIA,eAAe9B,4BAAAA,EAAkB;AACnC,QAAA,OAAO;UAAEiC,IAAAA,EAAM,WAAA;AAAaC,UAAAA,OAAAA,EAASJ,GAAAA,CAAII;AAAQ,SAAA;AACnD,OAAA,MAAA,IAAWJ,eAAeV,uBAAAA,EAAa;AACrC,QAAA,OAAO;UAAEa,IAAAA,EAAM,MAAA;AAAQC,UAAAA,OAAAA,EAASJ,GAAAA,CAAII;AAAQ,SAAA;AAC9C,OAAA,MAAA,IAAWJ,eAAeK,uBAAAA,EAAa;AACrC,QAAA,OAAO;UAAEF,IAAAA,EAAM,MAAA;AAAQC,UAAAA,OAAAA,EAASJ,GAAAA,CAAII;AAAQ,SAAA;AAC9C,OAAA,MAAA,IAAWJ,eAAeM,yBAAAA,EAAe;AACvC,QAAA,OAAO;UAAEH,IAAAA,EAAM,QAAA;UAAUC,OAAAA,EAASJ,GAAAA,CAAII,QAAQR,GAAAA,CAAI,CAACW,SAASA,IAAAA,CAAKtB,IAAI,CAAA,CAAEuB,IAAAA,CAAK,IAAA;AAAM,SAAA;AACpF;AACA,MAAA,OAAO;AAAEL,QAAAA,IAAAA,EAAMH,GAAAA,CAAIG,IAAAA;AAAMC,QAAAA,OAAAA,EAASJ,GAAAA,CAAII;AAAQ,OAAA;KAChD,CAAA;AAEA,IAAA,IAAIK,UAAAA;AACJ,IAAA,IAAItE,KAAAA,CAAMsE,UAAAA,IAActE,KAAAA,CAAMsE,UAAAA,YAAsBC,aAAAA,EAAM;AACxD,MAAA,IAAI,IAAA,CAAKC,iBAAAA,CAAkBC,QAAAA,CAAS,QAAA,CAAA,EAAW;AAC7CH,QAAAA,UAAAA,GAAa;UACXzB,IAAAA,EAAM,MAAA;AACNG,UAAAA,QAAAA,EAAUhD,MAAMsE,UAAAA,CAAW3C;AAC7B,SAAA;OACF,MAAO;AACL,QAAA,IAAA,CAAK+C,MAAAA,CAAOC,KAAK,CAAA,wCAAA,CAA0C,CAAA;AAC7D;AACF,KAAA,MAAA,IAAW3E,MAAMsE,UAAAA,EAAY;AAC3B,MAAA,IAAI,IAAA,CAAKE,iBAAAA,CAAkBC,QAAAA,CAASzE,KAAAA,CAAMsE,UAAU,CAAA,EAAG;AACrDA,QAAAA,UAAAA,GAAatE,KAAAA,CAAMsE,UAAAA;OACrB,MAAO;AACL,QAAA,IAAA,CAAKI,MAAAA,CAAOC,IAAAA,CAAK,CAAA,iCAAA,EAAoC3E,KAAAA,CAAMsE,UAAU,CAAA,mBAAA,CAAqB,CAAA;AAC5F;AACF;AAEA,IAAA,OAAO;AACL,MAAA,GAAG,IAAA,CAAKM,UAAAA;MACR,GAAG5E,KAAAA;AACHsE,MAAAA,UAAAA;AACAnF,MAAAA,KAAAA,EAAO,IAAA,CAAKA,KAAAA;AACZiD,MAAAA,KAAAA,EAAOyC,gBAASzC,KAAAA,EAAO,CAAC,EAAET,IAAAA,EAAM,GAAG+B,MAAAA,KAAW;AAAC/B,QAAAA,IAAAA;AAAM+B,QAAAA;AAAK,OAAA,CAAA;AAC1DhD,MAAAA;AACF,KAAA;AACF;AAEUI,EAAAA,iBAAAA,CAAkBJ,QAAAA,EAAiE;AAC3F,IAAA,OAAOA,QAAAA,CAASoE,OAAAA,CAAQ,CAACjB,GAAAA,KAAAA;AACvB,MAAA,IAAIA,GAAAA,CAAIG,SAAS,MAAA,EAAQ;AACvB,QAAA,OAAO,IAAIb,uBAAAA,CAAYU,GAAAA,CAAII,OAAAA,EAASJ,IAAIkB,eAAe,CAAA;AACzD;AACA,MAAA,OAAO,IAAIhD,4BAAAA,CACT8B,GAAAA,CAAII,OAAAA,EACJJ,IAAIkB,eAAe,CAAA;KAEvB,CAAA;AACF;EAEAC,cAAAA,GAAiB;AACf,IAAA,OAAO;AACL,MAAA,GAAG,MAAMA,cAAAA,EAAAA;AACTtF,MAAAA,UAAAA,EAAY,IAAA,CAAKA,UAAAA;AACjBN,MAAAA,OAAAA,EAAS,IAAA,CAAKA,OAAAA;AACdF,MAAAA,qBAAAA,EAAuB,IAAA,CAAKA;AAC9B,KAAA;AACF;AAEA,EAAA,MAAM+F,aAAa,EAAEvF,UAAAA,EAAYN,OAAAA,EAAS,GAAG8F,UAAAA,EAAoD;AAC/F,IAAA,MAAMC,QAAAA,GAAW,MAAMnG,kBAAAA,CAAUoG,QAAAA,CAAS,GAAG1F,UAAAA,CAAAA,CAAAA,EAAcN,OAAAA,CAAAA,CAAS,CAAA;AACpE,IAAA,IAAI,EAAE+F,oBAAoBpG,eAAAA,CAAAA,EAAkB;AAC1C,MAAA,MAAM,IAAIsG,MAAM,4BAAA,CAAA;AAClB;AACAF,IAAAA,QAAAA,CAASG,OAAAA,EAAO;AAChBC,IAAAA,MAAAA,CAAOC,OAAO,IAAA,EAAM;MAClB,GAAGN,QAAAA;AACH/F,MAAAA,KAAAA,EAAOgG,QAAAA,CAAShG;KAClB,CAAA;AACF;AACF","file":"chat.cjs","sourcesContent":["/**\n * Copyright 2025 © BeeAI a Series of LF Projects, LLC\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport {\n ChatModelInput,\n ChatModel,\n ChatModelOutput,\n ChatModelEvents,\n ChatModelObjectInput,\n ChatModelObjectOutput,\n} from \"@/backend/chat.js\";\nimport {\n CoreAssistantMessage,\n ModelMessage,\n CoreToolMessage,\n generateObject,\n generateText,\n jsonSchema,\n LanguageModel as _LanguageModel,\n streamText,\n TextPart,\n ToolCallPart,\n ToolChoice,\n} from \"ai\";\ntype LanguageModelV2 = Exclude<_LanguageModel, string>;\nimport { Emitter } from \"@/emitter/emitter.js\";\nimport {\n AssistantMessage,\n CustomMessage,\n Message,\n SystemMessage,\n ToolMessage,\n UserMessage,\n} from \"@/backend/message.js\";\nimport { GetRunContext } from \"@/context.js\";\nimport { ValueError } from \"@/errors.js\";\nimport { isEmpty, mapToObj, toCamelCase } from \"remeda\";\nimport { FullModelName } from \"@/backend/utils.js\";\nimport { ChatModelError } from \"@/backend/errors.js\";\nimport { z, ZodArray, ZodEnum, ZodSchema } from \"zod\";\nimport { Tool } from \"@/tools/base.js\";\nimport { encodeCustomMessage, extractTokenUsage } from \"@/adapters/vercel/backend/utils.js\";\n\nexport abstract class VercelChatModel<\n M extends LanguageModelV2 = LanguageModelV2,\n> extends ChatModel {\n public readonly emitter: Emitter<ChatModelEvents>;\n public readonly supportsToolStreaming: boolean = true;\n\n constructor(private readonly model: M) {\n super();\n if (!this.modelId) {\n throw new ValueError(\"No modelId has been provided!\");\n }\n this.emitter = Emitter.root.child({\n namespace: [\"backend\", this.providerId, \"chat\"],\n creator: this,\n });\n }\n\n get modelId(): string {\n return this.model.modelId;\n }\n\n get providerId(): string {\n const provider = this.model.provider.split(\".\")[0].split(\"-\")[0];\n return toCamelCase(provider);\n }\n\n protected async _create(input: ChatModelInput, run: GetRunContext<this>) {\n const responseFormat = input.responseFormat;\n if (responseFormat && (responseFormat instanceof ZodSchema || responseFormat.schema)) {\n const { output } = await this._createStructure(\n {\n ...input,\n schema: responseFormat,\n },\n run,\n );\n return output;\n }\n\n const {\n finishReason,\n usage,\n response: { messages },\n } = await generateText(await this.transformInput(input));\n\n return new ChatModelOutput(\n this.transformMessages(messages),\n extractTokenUsage(usage),\n finishReason,\n );\n }\n\n protected async _createStructure<T>(\n { schema, ...input }: ChatModelObjectInput<T>,\n run: GetRunContext<this>,\n ): Promise<ChatModelObjectOutput<T>> {\n const response = await generateObject({\n temperature: 0,\n ...(await this.transformInput(input)),\n abortSignal: run.signal,\n model: this.model,\n ...(schema instanceof ZodSchema\n ? {\n schema,\n output: ((schema._input || schema) instanceof ZodArray\n ? \"array\"\n : (schema._input || schema) instanceof ZodEnum\n ? \"enum\"\n : \"object\") as any,\n }\n : {\n schema: schema.schema ? jsonSchema<T>(schema.schema) : z.any(),\n schemaName: schema.name,\n schemaDescription: schema.description,\n }),\n });\n\n return {\n object: response.object as T,\n output: new ChatModelOutput(\n [new AssistantMessage(JSON.stringify(response.object, null, 2))],\n extractTokenUsage(response.usage),\n response.finishReason,\n ),\n };\n }\n\n async *_createStream(input: ChatModelInput, run: GetRunContext<this>) {\n if (!this.supportsToolStreaming && !isEmpty(input.tools ?? [])) {\n const response = await this._create(input, run);\n yield response;\n return;\n }\n\n const {\n fullStream,\n usage: usagePromise,\n finishReason: finishReasonPromise,\n response: responsePromise,\n } = streamText({\n ...(await this.transformInput(input)),\n abortSignal: run.signal,\n });\n\n let lastChunk: ChatModelOutput | null = null;\n for await (const event of fullStream) {\n let message: Message;\n switch (event.type) {\n case \"text-delta\":\n message = new AssistantMessage(event.text);\n break;\n case \"tool-call\":\n message = new AssistantMessage({\n type: event.type,\n toolCallId: event.toolCallId,\n toolName: event.toolName,\n input: event.input,\n });\n break;\n case \"error\":\n throw new ChatModelError(\"Unhandled error\", [event.error as Error]);\n case \"tool-result\":\n message = new ToolMessage({\n type: event.type,\n toolCallId: event.toolCallId,\n toolName: event.toolName,\n output: event.output as any,\n });\n break;\n case \"abort\":\n message = new AssistantMessage([]);\n break;\n default:\n continue;\n }\n lastChunk = new ChatModelOutput([message]);\n yield lastChunk;\n }\n\n if (!lastChunk) {\n throw new ChatModelError(\"No chunks have been received!\");\n }\n\n try {\n const [usage, finishReason, _] = await Promise.all([\n usagePromise,\n finishReasonPromise,\n responsePromise,\n ]);\n lastChunk.usage = extractTokenUsage(usage);\n lastChunk.finishReason = finishReason;\n } catch (e) {\n if (!run.signal.aborted) {\n throw e;\n }\n }\n }\n\n protected async transformInput(\n input: ChatModelInput,\n ): Promise<Parameters<typeof generateText<Record<string, any>>>[0]> {\n const tools = await Promise.all(\n (input.tools ?? []).map(async (tool) => ({\n name: tool.name,\n description: tool.description,\n inputSchema: jsonSchema(await tool.getInputJsonSchema()),\n })),\n );\n\n const messages = input.messages.map((msg): ModelMessage => {\n if (msg instanceof CustomMessage) {\n msg = encodeCustomMessage(msg);\n }\n\n if (msg instanceof AssistantMessage) {\n return { role: \"assistant\", content: msg.content };\n } else if (msg instanceof ToolMessage) {\n return { role: \"tool\", content: msg.content };\n } else if (msg instanceof UserMessage) {\n return { role: \"user\", content: msg.content };\n } else if (msg instanceof SystemMessage) {\n return { role: \"system\", content: msg.content.map((part) => part.text).join(\"\\n\") };\n }\n return { role: msg.role, content: msg.content } as ModelMessage;\n });\n\n let toolChoice: ToolChoice<Record<string, any>> | undefined;\n if (input.toolChoice && input.toolChoice instanceof Tool) {\n if (this.toolChoiceSupport.includes(\"single\")) {\n toolChoice = {\n type: \"tool\",\n toolName: input.toolChoice.name,\n };\n } else {\n this.logger.warn(`The single tool choice is not supported.`);\n }\n } else if (input.toolChoice) {\n if (this.toolChoiceSupport.includes(input.toolChoice)) {\n toolChoice = input.toolChoice;\n } else {\n this.logger.warn(`The following tool choice value '${input.toolChoice}' is not supported.`);\n }\n }\n\n return {\n ...this.parameters,\n ...input,\n toolChoice,\n model: this.model,\n tools: mapToObj(tools, ({ name, ...tool }) => [name, tool]),\n messages,\n };\n }\n\n protected transformMessages(messages: (CoreAssistantMessage | CoreToolMessage)[]): Message[] {\n return messages.flatMap((msg) => {\n if (msg.role === \"tool\") {\n return new ToolMessage(msg.content, msg.providerOptions);\n }\n return new AssistantMessage(\n msg.content as TextPart | ToolCallPart | string,\n msg.providerOptions,\n );\n });\n }\n\n createSnapshot() {\n return {\n ...super.createSnapshot(),\n providerId: this.providerId,\n modelId: this.modelId,\n supportsToolStreaming: this.supportsToolStreaming,\n };\n }\n\n async loadSnapshot({ providerId, modelId, ...snapshot }: ReturnType<typeof this.createSnapshot>) {\n const instance = await ChatModel.fromName(`${providerId}:${modelId}` as FullModelName);\n if (!(instance instanceof VercelChatModel)) {\n throw new Error(\"Incorrect deserialization!\");\n }\n instance.destroy();\n Object.assign(this, {\n ...snapshot,\n model: instance.model,\n });\n }\n}\n"]}
1
+ {"version":3,"sources":["../../../../src/adapters/vercel/backend/chat.ts"],"names":["VercelChatModel","ChatModel","emitter","supportsToolStreaming","model","modelId","ValueError","Emitter","root","child","namespace","providerId","creator","provider","split","toCamelCase","_create","input","run","responseFormat","ZodSchema","schema","output","_createStructure","finishReason","usage","response","messages","generateText","temperature","transformInput","abortSignal","signal","ChatModelOutput","transformMessages","extractTokenUsage","generateObject","_input","ZodArray","ZodEnum","jsonSchema","z","any","schemaName","name","schemaDescription","description","object","AssistantMessage","JSON","stringify","_createStream","isEmpty","tools","fullStream","usagePromise","finishReasonPromise","responsePromise","streamText","streamEmpty","streamedToolCalls","Map","event","message","type","text","id","streamPartialToolCalls","chunk","toolName","toolCallId","set","delta","get","existingToolCall","delete","ChatModelError","error","ToolMessage","_","Promise","all","lastChunk","e","aborted","map","tool","inputSchema","getInputJsonSchema","msg","CustomMessage","encodeCustomMessage","role","content","UserMessage","SystemMessage","part","join","toolChoice","Tool","toolChoiceSupport","includes","logger","warn","parameters","mapToObj","flatMap","providerOptions","createSnapshot","loadSnapshot","snapshot","instance","fromName","Error","destroy","Object","assign"],"mappings":";;;;;;;;;;;;;;;AA6CO,MAAeA,wBAEZC,kBAAAA,CAAAA;EA/CV;;;;AAgDkBC,EAAAA,OAAAA;EACAC,qBAAAA,GAAiC,IAAA;AAEjD,EAAA,WAAA,CAA6BC,KAAAA,EAAU;AACrC,IAAA,KAAA,EAAK,EAAA,KADsBA,KAAAA,GAAAA,KAAAA;AAE3B,IAAA,IAAI,CAAC,KAAKC,OAAAA,EAAS;AACjB,MAAA,MAAM,IAAIC,sBAAW,+BAAA,CAAA;AACvB;AACA,IAAA,IAAA,CAAKJ,OAAAA,GAAUK,mBAAAA,CAAQC,IAAAA,CAAKC,KAAAA,CAAM;MAChCC,SAAAA,EAAW;AAAC,QAAA,SAAA;QAAW,IAAA,CAAKC,UAAAA;AAAY,QAAA;;MACxCC,OAAAA,EAAS;KACX,CAAA;AACF;AAEA,EAAA,IAAIP,OAAAA,GAAkB;AACpB,IAAA,OAAO,KAAKD,KAAAA,CAAMC,OAAAA;AACpB;AAEA,EAAA,IAAIM,UAAAA,GAAqB;AACvB,IAAA,MAAME,QAAAA,GAAW,IAAA,CAAKT,KAAAA,CAAMS,QAAAA,CAASC,KAAAA,CAAM,GAAA,CAAA,CAAK,CAAA,CAAA,CAAGA,KAAAA,CAAM,GAAA,CAAA,CAAK,CAAA,CAAA;AAC9D,IAAA,OAAOC,mBAAYF,QAAAA,CAAAA;AACrB;EAEA,MAAgBG,OAAAA,CAAQC,OAAuBC,GAAAA,EAA0B;AACvE,IAAA,MAAMC,iBAAiBF,KAAAA,CAAME,cAAAA;AAC7B,IAAA,IAAIA,cAAAA,KAAmBA,cAAAA,YAA0BC,aAAAA,IAAaD,cAAAA,CAAeE,MAAAA,CAAAA,EAAS;AACpF,MAAA,MAAM,EAAEC,MAAAA,EAAM,GAAK,MAAM,KAAKC,gBAAAA,CAC5B;QACE,GAAGN,KAAAA;QACHI,MAAAA,EAAQF;AACV,OAAA,EACAD,GAAAA,CAAAA;AAEF,MAAA,OAAOI,MAAAA;AACT;AAEA,IAAA,MAAM,EACJE,cACAC,KAAAA,EACAC,QAAAA,EAAU,EAAEC,QAAAA,EAAQ,EAAE,GACpB,MAAMC,eAAAA,CAAa;MACrBC,WAAAA,EAAa,CAAA;MACb,GAAI,MAAM,IAAA,CAAKC,cAAAA,CAAeb,KAAAA,CAAAA;AAC9Bc,MAAAA,WAAAA,EAAab,GAAAA,CAAIc;KACnB,CAAA;AAEA,IAAA,OAAO,IAAIC,yBACT,IAAA,CAAKC,iBAAAA,CAAkBP,QAAAA,CAAAA,EACvBQ,2BAAAA,CAAkBV,KAAAA,CAAAA,EAClBD,YAAAA,CAAAA;AAEJ;AAEA,EAAA,MAAgBD,iBACd,EAAEF,MAAAA,EAAQ,GAAGJ,KAAAA,IACbC,GAAAA,EACmC;AACnC,IAAA,MAAMQ,QAAAA,GAAW,MAAMU,iBAAAA,CAAe;MACpCP,WAAAA,EAAa,CAAA;MACb,GAAI,MAAM,IAAA,CAAKC,cAAAA,CAAeb,KAAAA,CAAAA;AAC9Bc,MAAAA,WAAAA,EAAab,GAAAA,CAAIc,MAAAA;AACjB,MAAA,GAAIX,kBAAkBD,aAAAA,GAClB;AACEC,QAAAA,MAAAA;QACAC,MAAAA,EAAAA,CAAUD,MAAAA,CAAOgB,UAAUhB,MAAAA,aAAmBiB,YAAAA,GAC1C,WACCjB,MAAAA,CAAOgB,MAAAA,IAAUhB,MAAAA,aAAmBkB,WAAAA,GACnC,MAAA,GACA;OACR,GACA;AACElB,QAAAA,MAAAA,EAAQA,OAAOA,MAAAA,GAASmB,aAAAA,CAAcnB,OAAOA,MAAM,CAAA,GAAIoB,MAAEC,GAAAA,EAAG;AAC5DC,QAAAA,UAAAA,EAAYtB,MAAAA,CAAOuB,IAAAA;AACnBC,QAAAA,iBAAAA,EAAmBxB,MAAAA,CAAOyB;AAC5B;KACN,CAAA;AAEA,IAAA,OAAO;AACLC,MAAAA,MAAAA,EAAQrB,QAAAA,CAASqB,MAAAA;AACjBzB,MAAAA,MAAAA,EAAQ,IAAIW,wBAAAA,CACV;AAAC,QAAA,IAAIe,6BAAiBC,IAAAA,CAAKC,SAAAA,CAAUxB,SAASqB,MAAAA,EAAQ,IAAA,EAAM,CAAA,CAAA;AAC5DZ,OAAAA,EAAAA,2BAAAA,CAAkBT,QAAAA,CAASD,KAAK,CAAA,EAChCC,QAAAA,CAASF,YAAY;AAEzB,KAAA;AACF;EAEA,OAAO2B,aAAAA,CAAclC,OAAuBC,GAAAA,EAA0B;AACpE,IAAA,MAAMC,iBAAiBF,KAAAA,CAAME,cAAAA;AAC7B,IAAA,IAAIA,cAAAA,KAAmBA,cAAAA,YAA0BC,aAAAA,IAAaD,cAAAA,CAAeE,MAAAA,CAAAA,EAAS;AACpF,MAAA,MAAM,EAAEC,MAAAA,EAAM,GAAK,MAAM,KAAKC,gBAAAA,CAC5B;QACE,GAAGN,KAAAA;QACHI,MAAAA,EAAQF;AACV,OAAA,EACAD,GAAAA,CAAAA;AAEF,MAAA,MAAMI,MAAAA;AACN,MAAA;AACF;AAEA,IAAA,IAAI,CAAC,KAAKnB,qBAAAA,IAAyB,CAACiD,eAAQnC,KAAAA,CAAMoC,KAAAA,IAAS,EAAE,CAAA,EAAG;AAC9D,MAAA,MAAM3B,QAAAA,GAAW,MAAM,IAAA,CAAKV,OAAAA,CAAQC,OAAOC,GAAAA,CAAAA;AAC3C,MAAA,MAAMQ,QAAAA;AACN,MAAA;AACF;AAEA,IAAA,MAAM,EACJ4B,YACA7B,KAAAA,EAAO8B,YAAAA,EACP/B,cAAcgC,mBAAAA,EACd9B,QAAAA,EAAU+B,eAAAA,EAAe,GACvBC,aAAAA,CAAW;MACb,GAAI,MAAM,IAAA,CAAK5B,cAAAA,CAAeb,KAAAA,CAAAA;AAC9Bc,MAAAA,WAAAA,EAAab,GAAAA,CAAIc;KACnB,CAAA;AAEA,IAAA,IAAI2B,WAAAA,GAAc,IAAA;AAClB,IAAA,MAAMC,iBAAAA,uBAAwBC,GAAAA,EAAAA;AAC9B,IAAA,WAAA,MAAiBC,SAASR,UAAAA,EAAY;AACpC,MAAA,IAAIS,OAAAA;AACJ,MAAA,QAAQD,MAAME,IAAAA;QACZ,KAAK,YAAA;AACHL,UAAAA,WAAAA,GAAc,KAAA;AACdI,UAAAA,OAAAA,GAAU,IAAIf,4BAAAA,CAAiBc,KAAAA,CAAMG,MAAM,EAAC,EAAGH,MAAMI,EAAE,CAAA;AACvD,UAAA,MAAM,IAAIjC,wBAAAA,CAAgB;AAAC8B,YAAAA;AAAQ,WAAA,CAAA;AACnC,UAAA;QACF,KAAK,UAAA;AACHJ,UAAAA,WAAAA,GAAc,KAAA;AACd,UAAA;AACF,QAAA,KAAK,kBAAA,EAAoB;AACvB,UAAA,IAAI,CAAC1C,MAAMkD,sBAAAA,EAAwB;AACjC,YAAA;AACF;AAEA,UAAA,MAAMC,KAAAA,GAAsB;YAC1BJ,IAAAA,EAAM,WAAA;AACNK,YAAAA,QAAAA,EAAUP,KAAAA,CAAMO,QAAAA;AAChBC,YAAAA,UAAAA,EAAYR,KAAAA,CAAMI,EAAAA;YAClBjD,KAAAA,EAAO;AACT,WAAA;AACA2C,UAAAA,iBAAAA,CAAkBW,GAAAA,CAAIT,KAAAA,CAAMI,EAAAA,EAAIE,KAAAA,CAAAA;AAChC,UAAA,MAAML,WAAU,IAAIf,4BAAAA,CAAiBoB,OAAO,EAAC,EAAGN,MAAMI,EAAE,CAAA;AACxD,UAAA,MAAM,IAAIjC,wBAAAA,CAAgB;AAAC8B,YAAAA;AAAQ,WAAA,CAAA;AACnC,UAAA;AACF;AACA,QAAA,KAAK,kBAAA,EAAoB;AACvB,UAAA,IAAI,CAAC9C,MAAMkD,sBAAAA,EAAwB;AACjC,YAAA;AACF;AAEA,UAAA,IAAIL,MAAMU,KAAAA,EAAO;AACf,YAAA,MAAMJ,KAAAA,GAAQR,iBAAAA,CAAkBa,GAAAA,CAAIX,KAAAA,CAAMI,EAAE,CAAA;AAC5C,YAAA,MAAMH,QAAAA,GAAU,IAAIf,4BAAAA,CAAiB;cAAE,GAAGoB,KAAAA;AAAOnD,cAAAA,KAAAA,EAAO6C,KAAAA,CAAMU;aAAM,EAAG,EAAC,EAAGV,KAAAA,CAAMI,EAAE,CAAA;AACnF,YAAA,MAAM,IAAIjC,wBAAAA,CAAgB;AAAC8B,cAAAA;AAAQ,aAAA,CAAA;AACrC;AACA,UAAA;AACF;AACA,QAAA,KAAK,WAAA,EAAa;AAChBJ,UAAAA,WAAAA,GAAc,KAAA;AACd,UAAA,MAAMe,gBAAAA,GAAmBd,iBAAAA,CAAkBa,GAAAA,CAAIX,KAAAA,CAAMQ,UAAU,CAAA;AAC/D,UAAA,IAAII,gBAAAA,EAAkB;AACpBd,YAAAA,iBAAAA,CAAkBe,MAAAA,CAAOb,MAAMQ,UAAU,CAAA;AACzC,YAAA;AACF;AACAP,UAAAA,OAAAA,GAAU,IAAIf,4BAAAA,CACZ;AACEgB,YAAAA,IAAAA,EAAMF,KAAAA,CAAME,IAAAA;AACZM,YAAAA,UAAAA,EAAYR,KAAAA,CAAMQ,UAAAA;AAClBD,YAAAA,QAAAA,EAAUP,KAAAA,CAAMO,QAAAA;AAChBpD,YAAAA,KAAAA,EAAO6C,KAAAA,CAAM7C;WACf,EACA,EAAC,EACD6C,KAAAA,CAAMQ,UAAU,CAAA;AAElB,UAAA,MAAM,IAAIrC,wBAAAA,CAAgB;AAAC8B,YAAAA;AAAQ,WAAA,CAAA;AACnC,UAAA;AACF;QACA,KAAK,OAAA;AACH,UAAA,MAAM,IAAIa,4BAAe,iBAAA,EAAmB;YAACd,KAAAA,CAAMe;AAAe,WAAA,CAAA;QACpE,KAAK,aAAA;AACHlB,UAAAA,WAAAA,GAAc,KAAA;AACdI,UAAAA,OAAAA,GAAU,IAAIe,uBAAAA,CACZ;AACEd,YAAAA,IAAAA,EAAMF,KAAAA,CAAME,IAAAA;AACZM,YAAAA,UAAAA,EAAYR,KAAAA,CAAMQ,UAAAA;AAClBD,YAAAA,QAAAA,EAAUP,KAAAA,CAAMO,QAAAA;AAChB/C,YAAAA,MAAAA,EAAQwC,KAAAA,CAAMxC;AAChB,WAAA,EACA,EAAC,EACD,CAAA,YAAA,EAAewC,KAAAA,CAAMQ,UAAU,CAAA,CAAE,CAAA;AAEnC,UAAA,MAAM,IAAIrC,wBAAAA,CAAgB;AAAC8B,YAAAA;AAAQ,WAAA,CAAA;AACnC,UAAA;AAKJ;AACF;AAEA,IAAA,IAAIJ,WAAAA,EAAa;AACf,MAAA,MAAM,IAAIiB,4BAAe,+BAAA,CAAA;AAC3B;AAEA,IAAA,IAAI;AACF,MAAA,MAAM,CAACnD,KAAAA,EAAOD,YAAAA,EAAcuD,CAAAA,CAAAA,GAAK,MAAMC,QAAQC,GAAAA,CAAI;AACjD1B,QAAAA,YAAAA;AACAC,QAAAA,mBAAAA;AACAC,QAAAA;AACD,OAAA,CAAA;AACD,MAAA,MAAMyB,SAAAA,GAAY,IAAIjD,wBAAAA,CAAgB,EAAE,CAAA;AACxCiD,MAAAA,SAAAA,CAAUzD,KAAAA,GAAQU,4BAAkBV,KAAAA,CAAAA;AACpCyD,MAAAA,SAAAA,CAAU1D,YAAAA,GAAeA,YAAAA;AACzB,MAAA,MAAM0D,SAAAA;AACR,KAAA,CAAA,OAASC,CAAAA,EAAG;AACV,MAAA,IAAI,CAACjE,GAAAA,CAAIc,MAAAA,CAAOoD,OAAAA,EAAS;AACvB,QAAA,MAAMD,CAAAA;AACR;AACF;AACF;AAEA,EAAA,MAAgBrD,eACdb,KAAAA,EACkE;AAClE,IAAA,MAAMoC,KAAAA,GAAQ,MAAM2B,OAAAA,CAAQC,GAAAA,CAAAA,CACzBhE,KAAAA,CAAMoC,SAAS,EAAA,EAAIgC,GAAAA,CAAI,OAAOC,IAAAA,MAAU;AACvC1C,MAAAA,IAAAA,EAAM0C,IAAAA,CAAK1C,IAAAA;AACXE,MAAAA,WAAAA,EAAawC,IAAAA,CAAKxC,WAAAA;AAClByC,MAAAA,WAAAA,EAAa/C,aAAAA,CAAW,MAAM8C,IAAAA,CAAKE,kBAAAA,EAAkB;AACvD,KAAA,CAAA,CAAA,CAAA;AAGF,IAAA,MAAM7D,QAAAA,GAAWV,KAAAA,CAAMU,QAAAA,CAAS0D,GAAAA,CAAI,CAACI,GAAAA,KAAAA;AACnC,MAAA,IAAIA,eAAeC,yBAAAA,EAAe;AAChCD,QAAAA,GAAAA,GAAME,8BAAoBF,GAAAA,CAAAA;AAC5B;AAEA,MAAA,IAAIA,eAAezC,4BAAAA,EAAkB;AACnC,QAAA,OAAO;UAAE4C,IAAAA,EAAM,WAAA;AAAaC,UAAAA,OAAAA,EAASJ,GAAAA,CAAII;AAAQ,SAAA;AACnD,OAAA,MAAA,IAAWJ,eAAeX,uBAAAA,EAAa;AACrC,QAAA,OAAO;UAAEc,IAAAA,EAAM,MAAA;AAAQC,UAAAA,OAAAA,EAASJ,GAAAA,CAAII;AAAQ,SAAA;AAC9C,OAAA,MAAA,IAAWJ,eAAeK,uBAAAA,EAAa;AACrC,QAAA,OAAO;UAAEF,IAAAA,EAAM,MAAA;AAAQC,UAAAA,OAAAA,EAASJ,GAAAA,CAAII;AAAQ,SAAA;AAC9C,OAAA,MAAA,IAAWJ,eAAeM,yBAAAA,EAAe;AACvC,QAAA,OAAO;UAAEH,IAAAA,EAAM,QAAA;UAAUC,OAAAA,EAASJ,GAAAA,CAAII,QAAQR,GAAAA,CAAI,CAACW,SAASA,IAAAA,CAAK/B,IAAI,CAAA,CAAEgC,IAAAA,CAAK,IAAA;AAAM,SAAA;AACpF;AACA,MAAA,OAAO;AAAEL,QAAAA,IAAAA,EAAMH,GAAAA,CAAIG,IAAAA;AAAMC,QAAAA,OAAAA,EAASJ,GAAAA,CAAII;AAAQ,OAAA;KAChD,CAAA;AAEA,IAAA,IAAIK,UAAAA;AACJ,IAAA,IAAIjF,KAAAA,CAAMiF,UAAAA,IAAcjF,KAAAA,CAAMiF,UAAAA,YAAsBC,aAAAA,EAAM;AACxD,MAAA,IAAI,IAAA,CAAKC,iBAAAA,CAAkBC,QAAAA,CAAS,QAAA,CAAA,EAAW;AAC7CH,QAAAA,UAAAA,GAAa;UACXlC,IAAAA,EAAM,MAAA;AACNK,UAAAA,QAAAA,EAAUpD,MAAMiF,UAAAA,CAAWtD;AAC7B,SAAA;OACF,MAAO;AACL,QAAA,IAAA,CAAK0D,MAAAA,CAAOC,KAAK,CAAA,wCAAA,CAA0C,CAAA;AAC7D;AACF,KAAA,MAAA,IAAWtF,MAAMiF,UAAAA,EAAY;AAC3B,MAAA,IAAI,IAAA,CAAKE,iBAAAA,CAAkBC,QAAAA,CAASpF,KAAAA,CAAMiF,UAAU,CAAA,EAAG;AACrDA,QAAAA,UAAAA,GAAajF,KAAAA,CAAMiF,UAAAA;OACrB,MAAO;AACL,QAAA,IAAA,CAAKI,MAAAA,CAAOC,IAAAA,CAAK,CAAA,iCAAA,EAAoCtF,KAAAA,CAAMiF,UAAU,CAAA,mBAAA,CAAqB,CAAA;AAC5F;AACF;AAEA,IAAA,OAAO;AACL,MAAA,GAAG,IAAA,CAAKM,UAAAA;MACR,GAAGvF,KAAAA;AACHiF,MAAAA,UAAAA;AACA9F,MAAAA,KAAAA,EAAO,IAAA,CAAKA,KAAAA;AACZiD,MAAAA,KAAAA,EAAOoD,gBAASpD,KAAAA,EAAO,CAAC,EAAET,IAAAA,EAAM,GAAG0C,MAAAA,KAAW;AAAC1C,QAAAA,IAAAA;AAAM0C,QAAAA;AAAK,OAAA,CAAA;AAC1D3D,MAAAA;AACF,KAAA;AACF;AAEUO,EAAAA,iBAAAA,CAAkBP,QAAAA,EAAiE;AAC3F,IAAA,OAAOA,QAAAA,CAAS+E,OAAAA,CAAQ,CAACjB,GAAAA,KAAAA;AACvB,MAAA,IAAIA,GAAAA,CAAIG,SAAS,MAAA,EAAQ;AACvB,QAAA,OAAO,IAAId,uBAAAA,CAAYW,GAAAA,CAAII,OAAAA,EAASJ,IAAIkB,eAAe,CAAA;AACzD;AACA,MAAA,OAAO,IAAI3D,4BAAAA,CACTyC,GAAAA,CAAII,OAAAA,EACJJ,IAAIkB,eAAe,CAAA;KAEvB,CAAA;AACF;EAEAC,cAAAA,GAAiB;AACf,IAAA,OAAO;AACL,MAAA,GAAG,MAAMA,cAAAA,EAAAA;AACTjG,MAAAA,UAAAA,EAAY,IAAA,CAAKA,UAAAA;AACjBN,MAAAA,OAAAA,EAAS,IAAA,CAAKA,OAAAA;AACdF,MAAAA,qBAAAA,EAAuB,IAAA,CAAKA;AAC9B,KAAA;AACF;AAEA,EAAA,MAAM0G,aAAa,EAAElG,UAAAA,EAAYN,OAAAA,EAAS,GAAGyG,UAAAA,EAAoD;AAC/F,IAAA,MAAMC,QAAAA,GAAW,MAAM9G,kBAAAA,CAAU+G,QAAAA,CAAS,GAAGrG,UAAAA,CAAAA,CAAAA,EAAcN,OAAAA,CAAAA,CAAS,CAAA;AACpE,IAAA,IAAI,EAAE0G,oBAAoB/G,eAAAA,CAAAA,EAAkB;AAC1C,MAAA,MAAM,IAAIiH,MAAM,4BAAA,CAAA;AAClB;AACAF,IAAAA,QAAAA,CAASG,OAAAA,EAAO;AAChBC,IAAAA,MAAAA,CAAOC,OAAO,IAAA,EAAM;MAClB,GAAGN,QAAAA;AACH1G,MAAAA,KAAAA,EAAO2G,QAAAA,CAAS3G;KAClB,CAAA;AACF;AACF","file":"chat.cjs","sourcesContent":["/**\n * Copyright 2025 © BeeAI a Series of LF Projects, LLC\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport {\n ChatModelInput,\n ChatModel,\n ChatModelOutput,\n ChatModelEvents,\n ChatModelObjectInput,\n ChatModelObjectOutput,\n} from \"@/backend/chat.js\";\nimport {\n CoreAssistantMessage,\n ModelMessage,\n CoreToolMessage,\n generateObject,\n generateText,\n jsonSchema,\n LanguageModel as _LanguageModel,\n streamText,\n TextPart,\n ToolCallPart,\n ToolChoice,\n} from \"ai\";\ntype LanguageModelV2 = Exclude<_LanguageModel, string>;\nimport { Emitter } from \"@/emitter/emitter.js\";\nimport {\n AssistantMessage,\n CustomMessage,\n Message,\n SystemMessage,\n ToolMessage,\n UserMessage,\n} from \"@/backend/message.js\";\nimport { GetRunContext } from \"@/context.js\";\nimport { ValueError } from \"@/errors.js\";\nimport { isEmpty, mapToObj, toCamelCase } from \"remeda\";\nimport { FullModelName } from \"@/backend/utils.js\";\nimport { ChatModelError } from \"@/backend/errors.js\";\nimport { z, ZodArray, ZodEnum, ZodSchema } from \"zod\";\nimport { Tool } from \"@/tools/base.js\";\nimport { encodeCustomMessage, extractTokenUsage } from \"@/adapters/vercel/backend/utils.js\";\n\nexport abstract class VercelChatModel<\n M extends LanguageModelV2 = LanguageModelV2,\n> extends ChatModel {\n public readonly emitter: Emitter<ChatModelEvents>;\n public readonly supportsToolStreaming: boolean = true;\n\n constructor(private readonly model: M) {\n super();\n if (!this.modelId) {\n throw new ValueError(\"No modelId has been provided!\");\n }\n this.emitter = Emitter.root.child({\n namespace: [\"backend\", this.providerId, \"chat\"],\n creator: this,\n });\n }\n\n get modelId(): string {\n return this.model.modelId;\n }\n\n get providerId(): string {\n const provider = this.model.provider.split(\".\")[0].split(\"-\")[0];\n return toCamelCase(provider);\n }\n\n protected async _create(input: ChatModelInput, run: GetRunContext<this>) {\n const responseFormat = input.responseFormat;\n if (responseFormat && (responseFormat instanceof ZodSchema || responseFormat.schema)) {\n const { output } = await this._createStructure(\n {\n ...input,\n schema: responseFormat,\n },\n run,\n );\n return output;\n }\n\n const {\n finishReason,\n usage,\n response: { messages },\n } = await generateText({\n temperature: 0,\n ...(await this.transformInput(input)),\n abortSignal: run.signal,\n });\n\n return new ChatModelOutput(\n this.transformMessages(messages),\n extractTokenUsage(usage),\n finishReason,\n );\n }\n\n protected async _createStructure<T>(\n { schema, ...input }: ChatModelObjectInput<T>,\n run: GetRunContext<this>,\n ): Promise<ChatModelObjectOutput<T>> {\n const response = await generateObject({\n temperature: 0,\n ...(await this.transformInput(input)),\n abortSignal: run.signal,\n ...(schema instanceof ZodSchema\n ? {\n schema,\n output: ((schema._input || schema) instanceof ZodArray\n ? \"array\"\n : (schema._input || schema) instanceof ZodEnum\n ? \"enum\"\n : \"object\") as any,\n }\n : {\n schema: schema.schema ? jsonSchema<T>(schema.schema) : z.any(),\n schemaName: schema.name,\n schemaDescription: schema.description,\n }),\n });\n\n return {\n object: response.object as T,\n output: new ChatModelOutput(\n [new AssistantMessage(JSON.stringify(response.object, null, 2))],\n extractTokenUsage(response.usage),\n response.finishReason,\n ),\n };\n }\n\n async *_createStream(input: ChatModelInput, run: GetRunContext<this>) {\n const responseFormat = input.responseFormat;\n if (responseFormat && (responseFormat instanceof ZodSchema || responseFormat.schema)) {\n const { output } = await this._createStructure(\n {\n ...input,\n schema: responseFormat,\n },\n run,\n );\n yield output;\n return;\n }\n\n if (!this.supportsToolStreaming && !isEmpty(input.tools ?? [])) {\n const response = await this._create(input, run);\n yield response;\n return;\n }\n\n const {\n fullStream,\n usage: usagePromise,\n finishReason: finishReasonPromise,\n response: responsePromise,\n } = streamText({\n ...(await this.transformInput(input)),\n abortSignal: run.signal,\n });\n\n let streamEmpty = true;\n const streamedToolCalls = new Map<string, ToolCallPart>();\n for await (const event of fullStream) {\n let message: Message;\n switch (event.type) {\n case \"text-delta\":\n streamEmpty = false;\n message = new AssistantMessage(event.text, {}, event.id);\n yield new ChatModelOutput([message]);\n break;\n case \"text-end\":\n streamEmpty = false;\n break;\n case \"tool-input-start\": {\n if (!input.streamPartialToolCalls) {\n break;\n }\n\n const chunk: ToolCallPart = {\n type: \"tool-call\",\n toolName: event.toolName,\n toolCallId: event.id,\n input: \"\",\n };\n streamedToolCalls.set(event.id, chunk);\n const message = new AssistantMessage(chunk, {}, event.id);\n yield new ChatModelOutput([message]);\n break;\n }\n case \"tool-input-delta\": {\n if (!input.streamPartialToolCalls) {\n break;\n }\n\n if (event.delta) {\n const chunk = streamedToolCalls.get(event.id)!;\n const message = new AssistantMessage({ ...chunk, input: event.delta }, {}, event.id);\n yield new ChatModelOutput([message]);\n }\n break;\n }\n case \"tool-call\": {\n streamEmpty = false;\n const existingToolCall = streamedToolCalls.get(event.toolCallId);\n if (existingToolCall) {\n streamedToolCalls.delete(event.toolCallId);\n break;\n }\n message = new AssistantMessage(\n {\n type: event.type,\n toolCallId: event.toolCallId,\n toolName: event.toolName,\n input: event.input,\n },\n {},\n event.toolCallId,\n );\n yield new ChatModelOutput([message]);\n break;\n }\n case \"error\":\n throw new ChatModelError(\"Unhandled error\", [event.error as Error]);\n case \"tool-result\":\n streamEmpty = false;\n message = new ToolMessage(\n {\n type: event.type,\n toolCallId: event.toolCallId,\n toolName: event.toolName,\n output: event.output as any,\n },\n {},\n `tool_result_${event.toolCallId}`,\n );\n yield new ChatModelOutput([message]);\n break;\n case \"abort\":\n break;\n default:\n break;\n }\n }\n\n if (streamEmpty) {\n throw new ChatModelError(\"No chunks have been received!\");\n }\n\n try {\n const [usage, finishReason, _] = await Promise.all([\n usagePromise,\n finishReasonPromise,\n responsePromise,\n ]);\n const lastChunk = new ChatModelOutput([]);\n lastChunk.usage = extractTokenUsage(usage);\n lastChunk.finishReason = finishReason;\n yield lastChunk;\n } catch (e) {\n if (!run.signal.aborted) {\n throw e;\n }\n }\n }\n\n protected async transformInput(\n input: ChatModelInput,\n ): Promise<Parameters<typeof generateText<Record<string, any>>>[0]> {\n const tools = await Promise.all(\n (input.tools ?? []).map(async (tool) => ({\n name: tool.name,\n description: tool.description,\n inputSchema: jsonSchema(await tool.getInputJsonSchema()),\n })),\n );\n\n const messages = input.messages.map((msg): ModelMessage => {\n if (msg instanceof CustomMessage) {\n msg = encodeCustomMessage(msg);\n }\n\n if (msg instanceof AssistantMessage) {\n return { role: \"assistant\", content: msg.content };\n } else if (msg instanceof ToolMessage) {\n return { role: \"tool\", content: msg.content };\n } else if (msg instanceof UserMessage) {\n return { role: \"user\", content: msg.content };\n } else if (msg instanceof SystemMessage) {\n return { role: \"system\", content: msg.content.map((part) => part.text).join(\"\\n\") };\n }\n return { role: msg.role, content: msg.content } as ModelMessage;\n });\n\n let toolChoice: ToolChoice<Record<string, any>> | undefined;\n if (input.toolChoice && input.toolChoice instanceof Tool) {\n if (this.toolChoiceSupport.includes(\"single\")) {\n toolChoice = {\n type: \"tool\",\n toolName: input.toolChoice.name,\n };\n } else {\n this.logger.warn(`The single tool choice is not supported.`);\n }\n } else if (input.toolChoice) {\n if (this.toolChoiceSupport.includes(input.toolChoice)) {\n toolChoice = input.toolChoice;\n } else {\n this.logger.warn(`The following tool choice value '${input.toolChoice}' is not supported.`);\n }\n }\n\n return {\n ...this.parameters,\n ...input,\n toolChoice,\n model: this.model,\n tools: mapToObj(tools, ({ name, ...tool }) => [name, tool]),\n messages,\n };\n }\n\n protected transformMessages(messages: (CoreAssistantMessage | CoreToolMessage)[]): Message[] {\n return messages.flatMap((msg) => {\n if (msg.role === \"tool\") {\n return new ToolMessage(msg.content, msg.providerOptions);\n }\n return new AssistantMessage(\n msg.content as TextPart | ToolCallPart | string,\n msg.providerOptions,\n );\n });\n }\n\n createSnapshot() {\n return {\n ...super.createSnapshot(),\n providerId: this.providerId,\n modelId: this.modelId,\n supportsToolStreaming: this.supportsToolStreaming,\n };\n }\n\n async loadSnapshot({ providerId, modelId, ...snapshot }: ReturnType<typeof this.createSnapshot>) {\n const instance = await ChatModel.fromName(`${providerId}:${modelId}` as FullModelName);\n if (!(instance instanceof VercelChatModel)) {\n throw new Error(\"Incorrect deserialization!\");\n }\n instance.destroy();\n Object.assign(this, {\n ...snapshot,\n model: instance.model,\n });\n }\n}\n"]}
@@ -1,7 +1,7 @@
1
1
  import { Logger } from '../../../logger/logger.cjs';
2
- import { m as ChatModel, g as ChatModelEvents, d as ChatModelInput, n as ChatModelOutput, a as ChatModelObjectInput, b as ChatModelObjectOutput, i as ChatModelCache, C as ChatModelParameters, l as ChatModelToolChoiceSupport } from '../../../chat-DB_cf15S.cjs';
2
+ import { m as ChatModel, g as ChatModelEvents, d as ChatModelInput, n as ChatModelOutput, a as ChatModelObjectInput, b as ChatModelObjectOutput, i as ChatModelCache, C as ChatModelParameters, l as ChatModelToolChoiceSupport } from '../../../chat-BOb9n7AJ.cjs';
3
3
  import { LanguageModel, generateText, CoreAssistantMessage, CoreToolMessage } from 'ai';
4
- import { E as Emitter } from '../../../emitter-CGKd_-kK.cjs';
4
+ import { E as Emitter } from '../../../emitter-jN55XZZq.cjs';
5
5
  import { Message } from '../../../backend/message.cjs';
6
6
  import { GetRunContext } from '../../../context.cjs';
7
7
  import 'pino';
@@ -1,7 +1,7 @@
1
1
  import { Logger } from '../../../logger/logger.js';
2
- import { m as ChatModel, g as ChatModelEvents, d as ChatModelInput, n as ChatModelOutput, a as ChatModelObjectInput, b as ChatModelObjectOutput, i as ChatModelCache, C as ChatModelParameters, l as ChatModelToolChoiceSupport } from '../../../chat-BbCYMAdE.js';
2
+ import { m as ChatModel, g as ChatModelEvents, d as ChatModelInput, n as ChatModelOutput, a as ChatModelObjectInput, b as ChatModelObjectOutput, i as ChatModelCache, C as ChatModelParameters, l as ChatModelToolChoiceSupport } from '../../../chat-DrqwUgD8.js';
3
3
  import { LanguageModel, generateText, CoreAssistantMessage, CoreToolMessage } from 'ai';
4
- import { E as Emitter } from '../../../emitter-CN7I4uSV.js';
4
+ import { E as Emitter } from '../../../emitter-36-9MnvA.js';
5
5
  import { Message } from '../../../backend/message.js';
6
6
  import { GetRunContext } from '../../../context.js';
7
7
  import 'pino';