ai 6.0.33 → 6.0.35

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (357) hide show
  1. package/CHANGELOG.md +16 -0
  2. package/dist/index.d.mts +50 -21
  3. package/dist/index.d.ts +50 -21
  4. package/dist/index.js +348 -286
  5. package/dist/index.js.map +1 -1
  6. package/dist/index.mjs +280 -219
  7. package/dist/index.mjs.map +1 -1
  8. package/dist/internal/index.js +1 -1
  9. package/dist/internal/index.mjs +1 -1
  10. package/docs/02-foundations/03-prompts.mdx +2 -2
  11. package/docs/03-ai-sdk-core/15-tools-and-tool-calling.mdx +1 -1
  12. package/docs/07-reference/01-ai-sdk-core/28-output.mdx +1 -1
  13. package/docs/07-reference/05-ai-sdk-errors/ai-ui-message-stream-error.mdx +67 -0
  14. package/package.json +6 -4
  15. package/src/agent/agent.ts +116 -0
  16. package/src/agent/create-agent-ui-stream-response.test.ts +258 -0
  17. package/src/agent/create-agent-ui-stream-response.ts +50 -0
  18. package/src/agent/create-agent-ui-stream.ts +73 -0
  19. package/src/agent/index.ts +33 -0
  20. package/src/agent/infer-agent-tools.ts +7 -0
  21. package/src/agent/infer-agent-ui-message.test-d.ts +54 -0
  22. package/src/agent/infer-agent-ui-message.ts +11 -0
  23. package/src/agent/pipe-agent-ui-stream-to-response.ts +52 -0
  24. package/src/agent/tool-loop-agent-on-finish-callback.ts +31 -0
  25. package/src/agent/tool-loop-agent-on-step-finish-callback.ts +11 -0
  26. package/src/agent/tool-loop-agent-settings.ts +182 -0
  27. package/src/agent/tool-loop-agent.test-d.ts +114 -0
  28. package/src/agent/tool-loop-agent.test.ts +442 -0
  29. package/src/agent/tool-loop-agent.ts +114 -0
  30. package/src/embed/__snapshots__/embed-many.test.ts.snap +191 -0
  31. package/src/embed/__snapshots__/embed.test.ts.snap +81 -0
  32. package/src/embed/embed-many-result.ts +53 -0
  33. package/src/embed/embed-many.test.ts +653 -0
  34. package/src/embed/embed-many.ts +378 -0
  35. package/src/embed/embed-result.ts +50 -0
  36. package/src/embed/embed.test.ts +298 -0
  37. package/src/embed/embed.ts +211 -0
  38. package/src/embed/index.ts +4 -0
  39. package/src/error/index.ts +35 -0
  40. package/src/error/invalid-argument-error.ts +34 -0
  41. package/src/error/invalid-stream-part-error.ts +28 -0
  42. package/src/error/invalid-tool-approval-error.ts +26 -0
  43. package/src/error/invalid-tool-input-error.ts +33 -0
  44. package/src/error/no-image-generated-error.ts +39 -0
  45. package/src/error/no-object-generated-error.ts +70 -0
  46. package/src/error/no-output-generated-error.ts +26 -0
  47. package/src/error/no-speech-generated-error.ts +18 -0
  48. package/src/error/no-such-tool-error.ts +35 -0
  49. package/src/error/no-transcript-generated-error.ts +20 -0
  50. package/src/error/tool-call-not-found-for-approval-error.ts +32 -0
  51. package/src/error/tool-call-repair-error.ts +30 -0
  52. package/src/error/ui-message-stream-error.ts +48 -0
  53. package/src/error/unsupported-model-version-error.ts +23 -0
  54. package/src/error/verify-no-object-generated-error.ts +27 -0
  55. package/src/generate-image/generate-image-result.ts +42 -0
  56. package/src/generate-image/generate-image.test.ts +1420 -0
  57. package/src/generate-image/generate-image.ts +360 -0
  58. package/src/generate-image/index.ts +18 -0
  59. package/src/generate-object/__snapshots__/generate-object.test.ts.snap +133 -0
  60. package/src/generate-object/__snapshots__/stream-object.test.ts.snap +297 -0
  61. package/src/generate-object/generate-object-result.ts +67 -0
  62. package/src/generate-object/generate-object.test-d.ts +49 -0
  63. package/src/generate-object/generate-object.test.ts +1191 -0
  64. package/src/generate-object/generate-object.ts +518 -0
  65. package/src/generate-object/index.ts +9 -0
  66. package/src/generate-object/inject-json-instruction.test.ts +181 -0
  67. package/src/generate-object/inject-json-instruction.ts +30 -0
  68. package/src/generate-object/output-strategy.ts +415 -0
  69. package/src/generate-object/parse-and-validate-object-result.ts +111 -0
  70. package/src/generate-object/repair-text.ts +12 -0
  71. package/src/generate-object/stream-object-result.ts +120 -0
  72. package/src/generate-object/stream-object.test-d.ts +74 -0
  73. package/src/generate-object/stream-object.test.ts +1950 -0
  74. package/src/generate-object/stream-object.ts +986 -0
  75. package/src/generate-object/validate-object-generation-input.ts +144 -0
  76. package/src/generate-speech/generate-speech-result.ts +30 -0
  77. package/src/generate-speech/generate-speech.test.ts +300 -0
  78. package/src/generate-speech/generate-speech.ts +190 -0
  79. package/src/generate-speech/generated-audio-file.ts +65 -0
  80. package/src/generate-speech/index.ts +3 -0
  81. package/src/generate-text/__snapshots__/generate-text.test.ts.snap +1872 -0
  82. package/src/generate-text/__snapshots__/stream-text.test.ts.snap +1255 -0
  83. package/src/generate-text/collect-tool-approvals.test.ts +553 -0
  84. package/src/generate-text/collect-tool-approvals.ts +116 -0
  85. package/src/generate-text/content-part.ts +25 -0
  86. package/src/generate-text/execute-tool-call.ts +129 -0
  87. package/src/generate-text/extract-reasoning-content.ts +17 -0
  88. package/src/generate-text/extract-text-content.ts +15 -0
  89. package/src/generate-text/generate-text-result.ts +168 -0
  90. package/src/generate-text/generate-text.test-d.ts +68 -0
  91. package/src/generate-text/generate-text.test.ts +7011 -0
  92. package/src/generate-text/generate-text.ts +1223 -0
  93. package/src/generate-text/generated-file.ts +70 -0
  94. package/src/generate-text/index.ts +57 -0
  95. package/src/generate-text/is-approval-needed.ts +29 -0
  96. package/src/generate-text/output-utils.ts +23 -0
  97. package/src/generate-text/output.test.ts +698 -0
  98. package/src/generate-text/output.ts +590 -0
  99. package/src/generate-text/parse-tool-call.test.ts +570 -0
  100. package/src/generate-text/parse-tool-call.ts +188 -0
  101. package/src/generate-text/prepare-step.ts +103 -0
  102. package/src/generate-text/prune-messages.test.ts +720 -0
  103. package/src/generate-text/prune-messages.ts +167 -0
  104. package/src/generate-text/reasoning-output.ts +20 -0
  105. package/src/generate-text/reasoning.ts +8 -0
  106. package/src/generate-text/response-message.ts +10 -0
  107. package/src/generate-text/run-tools-transformation.test.ts +1143 -0
  108. package/src/generate-text/run-tools-transformation.ts +420 -0
  109. package/src/generate-text/smooth-stream.test.ts +2101 -0
  110. package/src/generate-text/smooth-stream.ts +162 -0
  111. package/src/generate-text/step-result.ts +238 -0
  112. package/src/generate-text/stop-condition.ts +29 -0
  113. package/src/generate-text/stream-text-result.ts +463 -0
  114. package/src/generate-text/stream-text.test-d.ts +200 -0
  115. package/src/generate-text/stream-text.test.ts +19979 -0
  116. package/src/generate-text/stream-text.ts +2505 -0
  117. package/src/generate-text/to-response-messages.test.ts +922 -0
  118. package/src/generate-text/to-response-messages.ts +163 -0
  119. package/src/generate-text/tool-approval-request-output.ts +21 -0
  120. package/src/generate-text/tool-call-repair-function.ts +27 -0
  121. package/src/generate-text/tool-call.ts +47 -0
  122. package/src/generate-text/tool-error.ts +34 -0
  123. package/src/generate-text/tool-output-denied.ts +21 -0
  124. package/src/generate-text/tool-output.ts +7 -0
  125. package/src/generate-text/tool-result.ts +36 -0
  126. package/src/generate-text/tool-set.ts +14 -0
  127. package/src/global.ts +24 -0
  128. package/src/index.ts +50 -0
  129. package/src/logger/index.ts +6 -0
  130. package/src/logger/log-warnings.test.ts +351 -0
  131. package/src/logger/log-warnings.ts +119 -0
  132. package/src/middleware/__snapshots__/simulate-streaming-middleware.test.ts.snap +64 -0
  133. package/src/middleware/add-tool-input-examples-middleware.test.ts +476 -0
  134. package/src/middleware/add-tool-input-examples-middleware.ts +90 -0
  135. package/src/middleware/default-embedding-settings-middleware.test.ts +126 -0
  136. package/src/middleware/default-embedding-settings-middleware.ts +22 -0
  137. package/src/middleware/default-settings-middleware.test.ts +388 -0
  138. package/src/middleware/default-settings-middleware.ts +33 -0
  139. package/src/middleware/extract-json-middleware.test.ts +827 -0
  140. package/src/middleware/extract-json-middleware.ts +197 -0
  141. package/src/middleware/extract-reasoning-middleware.test.ts +1028 -0
  142. package/src/middleware/extract-reasoning-middleware.ts +238 -0
  143. package/src/middleware/index.ts +10 -0
  144. package/src/middleware/simulate-streaming-middleware.test.ts +911 -0
  145. package/src/middleware/simulate-streaming-middleware.ts +79 -0
  146. package/src/middleware/wrap-embedding-model.test.ts +358 -0
  147. package/src/middleware/wrap-embedding-model.ts +86 -0
  148. package/src/middleware/wrap-image-model.test.ts +423 -0
  149. package/src/middleware/wrap-image-model.ts +85 -0
  150. package/src/middleware/wrap-language-model.test.ts +518 -0
  151. package/src/middleware/wrap-language-model.ts +104 -0
  152. package/src/middleware/wrap-provider.test.ts +120 -0
  153. package/src/middleware/wrap-provider.ts +51 -0
  154. package/src/model/as-embedding-model-v3.test.ts +319 -0
  155. package/src/model/as-embedding-model-v3.ts +24 -0
  156. package/src/model/as-image-model-v3.test.ts +409 -0
  157. package/src/model/as-image-model-v3.ts +24 -0
  158. package/src/model/as-language-model-v3.test.ts +508 -0
  159. package/src/model/as-language-model-v3.ts +103 -0
  160. package/src/model/as-provider-v3.ts +36 -0
  161. package/src/model/as-speech-model-v3.test.ts +356 -0
  162. package/src/model/as-speech-model-v3.ts +24 -0
  163. package/src/model/as-transcription-model-v3.test.ts +529 -0
  164. package/src/model/as-transcription-model-v3.ts +24 -0
  165. package/src/model/resolve-model.test.ts +244 -0
  166. package/src/model/resolve-model.ts +126 -0
  167. package/src/prompt/call-settings.ts +148 -0
  168. package/src/prompt/content-part.ts +209 -0
  169. package/src/prompt/convert-to-language-model-prompt.test.ts +2018 -0
  170. package/src/prompt/convert-to-language-model-prompt.ts +442 -0
  171. package/src/prompt/create-tool-model-output.test.ts +508 -0
  172. package/src/prompt/create-tool-model-output.ts +34 -0
  173. package/src/prompt/data-content.test.ts +15 -0
  174. package/src/prompt/data-content.ts +134 -0
  175. package/src/prompt/index.ts +27 -0
  176. package/src/prompt/invalid-data-content-error.ts +29 -0
  177. package/src/prompt/invalid-message-role-error.ts +27 -0
  178. package/src/prompt/message-conversion-error.ts +28 -0
  179. package/src/prompt/message.ts +68 -0
  180. package/src/prompt/prepare-call-settings.test.ts +159 -0
  181. package/src/prompt/prepare-call-settings.ts +108 -0
  182. package/src/prompt/prepare-tools-and-tool-choice.test.ts +461 -0
  183. package/src/prompt/prepare-tools-and-tool-choice.ts +86 -0
  184. package/src/prompt/prompt.ts +43 -0
  185. package/src/prompt/split-data-url.ts +17 -0
  186. package/src/prompt/standardize-prompt.test.ts +82 -0
  187. package/src/prompt/standardize-prompt.ts +99 -0
  188. package/src/prompt/wrap-gateway-error.ts +29 -0
  189. package/src/registry/custom-provider.test.ts +211 -0
  190. package/src/registry/custom-provider.ts +155 -0
  191. package/src/registry/index.ts +7 -0
  192. package/src/registry/no-such-provider-error.ts +41 -0
  193. package/src/registry/provider-registry.test.ts +691 -0
  194. package/src/registry/provider-registry.ts +328 -0
  195. package/src/rerank/index.ts +2 -0
  196. package/src/rerank/rerank-result.ts +70 -0
  197. package/src/rerank/rerank.test.ts +516 -0
  198. package/src/rerank/rerank.ts +237 -0
  199. package/src/telemetry/assemble-operation-name.ts +21 -0
  200. package/src/telemetry/get-base-telemetry-attributes.ts +53 -0
  201. package/src/telemetry/get-tracer.ts +20 -0
  202. package/src/telemetry/noop-tracer.ts +69 -0
  203. package/src/telemetry/record-span.ts +63 -0
  204. package/src/telemetry/select-telemetry-attributes.ts +78 -0
  205. package/src/telemetry/select-temetry-attributes.test.ts +114 -0
  206. package/src/telemetry/stringify-for-telemetry.test.ts +114 -0
  207. package/src/telemetry/stringify-for-telemetry.ts +33 -0
  208. package/src/telemetry/telemetry-settings.ts +44 -0
  209. package/src/test/mock-embedding-model-v2.ts +35 -0
  210. package/src/test/mock-embedding-model-v3.ts +48 -0
  211. package/src/test/mock-image-model-v2.ts +28 -0
  212. package/src/test/mock-image-model-v3.ts +28 -0
  213. package/src/test/mock-language-model-v2.ts +72 -0
  214. package/src/test/mock-language-model-v3.ts +77 -0
  215. package/src/test/mock-provider-v2.ts +68 -0
  216. package/src/test/mock-provider-v3.ts +80 -0
  217. package/src/test/mock-reranking-model-v3.ts +25 -0
  218. package/src/test/mock-server-response.ts +69 -0
  219. package/src/test/mock-speech-model-v2.ts +24 -0
  220. package/src/test/mock-speech-model-v3.ts +24 -0
  221. package/src/test/mock-tracer.ts +156 -0
  222. package/src/test/mock-transcription-model-v2.ts +24 -0
  223. package/src/test/mock-transcription-model-v3.ts +24 -0
  224. package/src/test/mock-values.ts +4 -0
  225. package/src/test/not-implemented.ts +3 -0
  226. package/src/text-stream/create-text-stream-response.test.ts +38 -0
  227. package/src/text-stream/create-text-stream-response.ts +18 -0
  228. package/src/text-stream/index.ts +2 -0
  229. package/src/text-stream/pipe-text-stream-to-response.test.ts +38 -0
  230. package/src/text-stream/pipe-text-stream-to-response.ts +26 -0
  231. package/src/transcribe/index.ts +2 -0
  232. package/src/transcribe/transcribe-result.ts +60 -0
  233. package/src/transcribe/transcribe.test.ts +313 -0
  234. package/src/transcribe/transcribe.ts +173 -0
  235. package/src/types/embedding-model-middleware.ts +3 -0
  236. package/src/types/embedding-model.ts +18 -0
  237. package/src/types/image-model-middleware.ts +3 -0
  238. package/src/types/image-model-response-metadata.ts +16 -0
  239. package/src/types/image-model.ts +19 -0
  240. package/src/types/index.ts +29 -0
  241. package/src/types/json-value.ts +15 -0
  242. package/src/types/language-model-middleware.ts +3 -0
  243. package/src/types/language-model-request-metadata.ts +6 -0
  244. package/src/types/language-model-response-metadata.ts +21 -0
  245. package/src/types/language-model.ts +104 -0
  246. package/src/types/provider-metadata.ts +16 -0
  247. package/src/types/provider.ts +55 -0
  248. package/src/types/reranking-model.ts +6 -0
  249. package/src/types/speech-model-response-metadata.ts +21 -0
  250. package/src/types/speech-model.ts +6 -0
  251. package/src/types/transcription-model-response-metadata.ts +16 -0
  252. package/src/types/transcription-model.ts +9 -0
  253. package/src/types/usage.ts +200 -0
  254. package/src/types/warning.ts +7 -0
  255. package/src/ui/__snapshots__/append-response-messages.test.ts.snap +416 -0
  256. package/src/ui/__snapshots__/convert-to-model-messages.test.ts.snap +419 -0
  257. package/src/ui/__snapshots__/process-chat-text-response.test.ts.snap +142 -0
  258. package/src/ui/call-completion-api.ts +157 -0
  259. package/src/ui/chat-transport.ts +83 -0
  260. package/src/ui/chat.test-d.ts +233 -0
  261. package/src/ui/chat.test.ts +2695 -0
  262. package/src/ui/chat.ts +716 -0
  263. package/src/ui/convert-file-list-to-file-ui-parts.ts +36 -0
  264. package/src/ui/convert-to-model-messages.test.ts +2775 -0
  265. package/src/ui/convert-to-model-messages.ts +373 -0
  266. package/src/ui/default-chat-transport.ts +36 -0
  267. package/src/ui/direct-chat-transport.test.ts +446 -0
  268. package/src/ui/direct-chat-transport.ts +118 -0
  269. package/src/ui/http-chat-transport.test.ts +185 -0
  270. package/src/ui/http-chat-transport.ts +292 -0
  271. package/src/ui/index.ts +71 -0
  272. package/src/ui/last-assistant-message-is-complete-with-approval-responses.ts +44 -0
  273. package/src/ui/last-assistant-message-is-complete-with-tool-calls.test.ts +371 -0
  274. package/src/ui/last-assistant-message-is-complete-with-tool-calls.ts +39 -0
  275. package/src/ui/process-text-stream.test.ts +38 -0
  276. package/src/ui/process-text-stream.ts +16 -0
  277. package/src/ui/process-ui-message-stream.test.ts +8294 -0
  278. package/src/ui/process-ui-message-stream.ts +761 -0
  279. package/src/ui/text-stream-chat-transport.ts +23 -0
  280. package/src/ui/transform-text-to-ui-message-stream.test.ts +124 -0
  281. package/src/ui/transform-text-to-ui-message-stream.ts +27 -0
  282. package/src/ui/ui-messages.test.ts +48 -0
  283. package/src/ui/ui-messages.ts +534 -0
  284. package/src/ui/use-completion.ts +84 -0
  285. package/src/ui/validate-ui-messages.test.ts +1428 -0
  286. package/src/ui/validate-ui-messages.ts +476 -0
  287. package/src/ui-message-stream/create-ui-message-stream-response.test.ts +266 -0
  288. package/src/ui-message-stream/create-ui-message-stream-response.ts +32 -0
  289. package/src/ui-message-stream/create-ui-message-stream.test.ts +639 -0
  290. package/src/ui-message-stream/create-ui-message-stream.ts +124 -0
  291. package/src/ui-message-stream/get-response-ui-message-id.test.ts +55 -0
  292. package/src/ui-message-stream/get-response-ui-message-id.ts +24 -0
  293. package/src/ui-message-stream/handle-ui-message-stream-finish.test.ts +429 -0
  294. package/src/ui-message-stream/handle-ui-message-stream-finish.ts +135 -0
  295. package/src/ui-message-stream/index.ts +13 -0
  296. package/src/ui-message-stream/json-to-sse-transform-stream.ts +12 -0
  297. package/src/ui-message-stream/pipe-ui-message-stream-to-response.test.ts +90 -0
  298. package/src/ui-message-stream/pipe-ui-message-stream-to-response.ts +40 -0
  299. package/src/ui-message-stream/read-ui-message-stream.test.ts +122 -0
  300. package/src/ui-message-stream/read-ui-message-stream.ts +87 -0
  301. package/src/ui-message-stream/ui-message-chunks.test-d.ts +18 -0
  302. package/src/ui-message-stream/ui-message-chunks.ts +344 -0
  303. package/src/ui-message-stream/ui-message-stream-headers.ts +7 -0
  304. package/src/ui-message-stream/ui-message-stream-on-finish-callback.ts +32 -0
  305. package/src/ui-message-stream/ui-message-stream-response-init.ts +5 -0
  306. package/src/ui-message-stream/ui-message-stream-writer.ts +24 -0
  307. package/src/util/as-array.ts +3 -0
  308. package/src/util/async-iterable-stream.test.ts +241 -0
  309. package/src/util/async-iterable-stream.ts +94 -0
  310. package/src/util/consume-stream.ts +29 -0
  311. package/src/util/cosine-similarity.test.ts +57 -0
  312. package/src/util/cosine-similarity.ts +47 -0
  313. package/src/util/create-resolvable-promise.ts +30 -0
  314. package/src/util/create-stitchable-stream.test.ts +239 -0
  315. package/src/util/create-stitchable-stream.ts +112 -0
  316. package/src/util/data-url.ts +17 -0
  317. package/src/util/deep-partial.ts +84 -0
  318. package/src/util/detect-media-type.test.ts +670 -0
  319. package/src/util/detect-media-type.ts +184 -0
  320. package/src/util/download/download-function.ts +45 -0
  321. package/src/util/download/download.test.ts +69 -0
  322. package/src/util/download/download.ts +46 -0
  323. package/src/util/error-handler.ts +1 -0
  324. package/src/util/fix-json.test.ts +279 -0
  325. package/src/util/fix-json.ts +401 -0
  326. package/src/util/get-potential-start-index.test.ts +34 -0
  327. package/src/util/get-potential-start-index.ts +30 -0
  328. package/src/util/index.ts +11 -0
  329. package/src/util/is-deep-equal-data.test.ts +119 -0
  330. package/src/util/is-deep-equal-data.ts +48 -0
  331. package/src/util/is-non-empty-object.ts +5 -0
  332. package/src/util/job.ts +1 -0
  333. package/src/util/log-v2-compatibility-warning.ts +21 -0
  334. package/src/util/merge-abort-signals.test.ts +155 -0
  335. package/src/util/merge-abort-signals.ts +43 -0
  336. package/src/util/merge-objects.test.ts +118 -0
  337. package/src/util/merge-objects.ts +79 -0
  338. package/src/util/now.ts +4 -0
  339. package/src/util/parse-partial-json.test.ts +80 -0
  340. package/src/util/parse-partial-json.ts +30 -0
  341. package/src/util/prepare-headers.test.ts +51 -0
  342. package/src/util/prepare-headers.ts +14 -0
  343. package/src/util/prepare-retries.test.ts +10 -0
  344. package/src/util/prepare-retries.ts +47 -0
  345. package/src/util/retry-error.ts +41 -0
  346. package/src/util/retry-with-exponential-backoff.test.ts +446 -0
  347. package/src/util/retry-with-exponential-backoff.ts +154 -0
  348. package/src/util/serial-job-executor.test.ts +162 -0
  349. package/src/util/serial-job-executor.ts +36 -0
  350. package/src/util/simulate-readable-stream.test.ts +98 -0
  351. package/src/util/simulate-readable-stream.ts +39 -0
  352. package/src/util/split-array.test.ts +60 -0
  353. package/src/util/split-array.ts +20 -0
  354. package/src/util/value-of.ts +65 -0
  355. package/src/util/write-to-server-response.test.ts +266 -0
  356. package/src/util/write-to-server-response.ts +49 -0
  357. package/src/version.ts +5 -0
@@ -0,0 +1,1028 @@
1
+ import { LanguageModelV3Usage } from '@ai-sdk/provider';
2
+ import {
3
+ convertArrayToReadableStream,
4
+ convertAsyncIterableToArray,
5
+ } from '@ai-sdk/provider-utils/test';
6
+ import { describe, expect, it } from 'vitest';
7
+ import { generateText, streamText } from '../generate-text';
8
+ import { wrapLanguageModel } from '../middleware/wrap-language-model';
9
+ import { MockLanguageModelV3 } from '../test/mock-language-model-v3';
10
+ import { extractReasoningMiddleware } from './extract-reasoning-middleware';
11
+
12
+ const testUsage: LanguageModelV3Usage = {
13
+ inputTokens: {
14
+ total: 5,
15
+ noCache: 5,
16
+ cacheRead: 0,
17
+ cacheWrite: 0,
18
+ },
19
+ outputTokens: {
20
+ total: 10,
21
+ text: 10,
22
+ reasoning: 3,
23
+ },
24
+ };
25
+
26
+ describe('extractReasoningMiddleware', () => {
27
+ describe('wrapGenerate', () => {
28
+ it('should extract reasoning from <think> tags', async () => {
29
+ const mockModel = new MockLanguageModelV3({
30
+ async doGenerate() {
31
+ return {
32
+ content: [
33
+ {
34
+ type: 'text',
35
+ text: '<think>analyzing the request</think>Here is the response',
36
+ },
37
+ ],
38
+ finishReason: { unified: 'stop', raw: 'stop' },
39
+ usage: testUsage,
40
+ warnings: [],
41
+ };
42
+ },
43
+ });
44
+
45
+ const result = await generateText({
46
+ model: wrapLanguageModel({
47
+ model: mockModel,
48
+ middleware: extractReasoningMiddleware({ tagName: 'think' }),
49
+ }),
50
+ prompt: 'Hello, how can I help?',
51
+ });
52
+
53
+ expect(result.content).toMatchInlineSnapshot(`
54
+ [
55
+ {
56
+ "text": "analyzing the request",
57
+ "type": "reasoning",
58
+ },
59
+ {
60
+ "text": "Here is the response",
61
+ "type": "text",
62
+ },
63
+ ]
64
+ `);
65
+ });
66
+
67
+ it('should extract reasoning from <think> tags when there is no text', async () => {
68
+ const mockModel = new MockLanguageModelV3({
69
+ async doGenerate() {
70
+ return {
71
+ content: [
72
+ {
73
+ type: 'text',
74
+ text: '<think>analyzing the request\n</think>',
75
+ },
76
+ ],
77
+ finishReason: { unified: 'stop', raw: 'stop' },
78
+ usage: testUsage,
79
+ warnings: [],
80
+ };
81
+ },
82
+ });
83
+
84
+ const result = await generateText({
85
+ model: wrapLanguageModel({
86
+ model: mockModel,
87
+ middleware: extractReasoningMiddleware({ tagName: 'think' }),
88
+ }),
89
+ prompt: 'Hello, how can I help?',
90
+ });
91
+
92
+ expect(result.content).toMatchInlineSnapshot(`
93
+ [
94
+ {
95
+ "text": "analyzing the request
96
+ ",
97
+ "type": "reasoning",
98
+ },
99
+ {
100
+ "text": "",
101
+ "type": "text",
102
+ },
103
+ ]
104
+ `);
105
+ });
106
+
107
+ it('should extract reasoning from multiple <think> tags', async () => {
108
+ const mockModel = new MockLanguageModelV3({
109
+ async doGenerate() {
110
+ return {
111
+ content: [
112
+ {
113
+ type: 'text',
114
+ text: '<think>analyzing the request</think>Here is the response<think>thinking about the response</think>more',
115
+ },
116
+ ],
117
+ finishReason: { unified: 'stop', raw: 'stop' },
118
+ usage: testUsage,
119
+ warnings: [],
120
+ };
121
+ },
122
+ });
123
+
124
+ const result = await generateText({
125
+ model: wrapLanguageModel({
126
+ model: mockModel,
127
+ middleware: extractReasoningMiddleware({ tagName: 'think' }),
128
+ }),
129
+ prompt: 'Hello, how can I help?',
130
+ });
131
+
132
+ expect(result.content).toMatchInlineSnapshot(`
133
+ [
134
+ {
135
+ "text": "analyzing the request
136
+ thinking about the response",
137
+ "type": "reasoning",
138
+ },
139
+ {
140
+ "text": "Here is the response
141
+ more",
142
+ "type": "text",
143
+ },
144
+ ]
145
+ `);
146
+ });
147
+
148
+ it('should prepend <think> tag IFF startWithReasoning is true', async () => {
149
+ const mockModel = new MockLanguageModelV3({
150
+ async doGenerate() {
151
+ return {
152
+ content: [
153
+ {
154
+ type: 'text',
155
+ text: 'analyzing the request</think>Here is the response',
156
+ },
157
+ ],
158
+ finishReason: { unified: 'stop', raw: 'stop' },
159
+ usage: testUsage,
160
+ warnings: [],
161
+ };
162
+ },
163
+ });
164
+
165
+ const resultTrue = await generateText({
166
+ model: wrapLanguageModel({
167
+ model: mockModel,
168
+ middleware: extractReasoningMiddleware({
169
+ tagName: 'think',
170
+ startWithReasoning: true,
171
+ }),
172
+ }),
173
+ prompt: 'Hello, how can I help?',
174
+ });
175
+
176
+ const resultFalse = await generateText({
177
+ model: wrapLanguageModel({
178
+ model: mockModel,
179
+ middleware: extractReasoningMiddleware({
180
+ tagName: 'think',
181
+ }),
182
+ }),
183
+ prompt: 'Hello, how can I help?',
184
+ });
185
+
186
+ expect(resultTrue.content).toMatchInlineSnapshot(`
187
+ [
188
+ {
189
+ "text": "analyzing the request",
190
+ "type": "reasoning",
191
+ },
192
+ {
193
+ "text": "Here is the response",
194
+ "type": "text",
195
+ },
196
+ ]
197
+ `);
198
+
199
+ expect(resultFalse.content).toMatchInlineSnapshot(`
200
+ [
201
+ {
202
+ "text": "analyzing the request</think>Here is the response",
203
+ "type": "text",
204
+ },
205
+ ]
206
+ `);
207
+ });
208
+
209
+ it('should preserve reasoning property even when rest contains other properties', async () => {
210
+ const mockModel = new MockLanguageModelV3({
211
+ async doGenerate() {
212
+ return {
213
+ content: [
214
+ {
215
+ type: 'text',
216
+ text: '<think>analyzing the request</think>Here is the response',
217
+ },
218
+ ],
219
+ finishReason: { unified: 'stop', raw: 'stop' },
220
+ usage: testUsage,
221
+ reasoning: undefined,
222
+ warnings: [],
223
+ };
224
+ },
225
+ });
226
+
227
+ const result = await generateText({
228
+ model: wrapLanguageModel({
229
+ model: mockModel,
230
+ middleware: extractReasoningMiddleware({ tagName: 'think' }),
231
+ }),
232
+ prompt: 'Hello, how can I help?',
233
+ });
234
+
235
+ expect(result.content).toMatchInlineSnapshot(`
236
+ [
237
+ {
238
+ "text": "analyzing the request",
239
+ "type": "reasoning",
240
+ },
241
+ {
242
+ "text": "Here is the response",
243
+ "type": "text",
244
+ },
245
+ ]
246
+ `);
247
+ });
248
+ });
249
+
250
+ describe('wrapStream', () => {
251
+ it('should extract reasoning from split <think> tags', async () => {
252
+ const mockModel = new MockLanguageModelV3({
253
+ async doStream() {
254
+ return {
255
+ stream: convertArrayToReadableStream([
256
+ {
257
+ type: 'response-metadata',
258
+ id: 'id-0',
259
+ modelId: 'mock-model-id',
260
+ timestamp: new Date(0),
261
+ },
262
+ { type: 'text-start', id: '1' },
263
+ { type: 'text-delta', id: '1', delta: '<think>' },
264
+ { type: 'text-delta', id: '1', delta: 'ana' },
265
+ { type: 'text-delta', id: '1', delta: 'lyzing the request' },
266
+ { type: 'text-delta', id: '1', delta: '</think>' },
267
+ { type: 'text-delta', id: '1', delta: 'Here' },
268
+ { type: 'text-delta', id: '1', delta: ' is the response' },
269
+ { type: 'text-end', id: '1' },
270
+ {
271
+ type: 'finish',
272
+ finishReason: { unified: 'stop', raw: 'stop' },
273
+ usage: testUsage,
274
+ },
275
+ ]),
276
+ };
277
+ },
278
+ });
279
+
280
+ const result = streamText({
281
+ model: wrapLanguageModel({
282
+ model: mockModel,
283
+ middleware: extractReasoningMiddleware({ tagName: 'think' }),
284
+ }),
285
+ prompt: 'Hello, how can I help?',
286
+ });
287
+
288
+ expect(await convertAsyncIterableToArray(result.fullStream))
289
+ .toMatchInlineSnapshot(`
290
+ [
291
+ {
292
+ "type": "start",
293
+ },
294
+ {
295
+ "request": {},
296
+ "type": "start-step",
297
+ "warnings": [],
298
+ },
299
+ {
300
+ "id": "reasoning-0",
301
+ "type": "reasoning-start",
302
+ },
303
+ {
304
+ "id": "reasoning-0",
305
+ "providerMetadata": undefined,
306
+ "text": "ana",
307
+ "type": "reasoning-delta",
308
+ },
309
+ {
310
+ "id": "reasoning-0",
311
+ "providerMetadata": undefined,
312
+ "text": "lyzing the request",
313
+ "type": "reasoning-delta",
314
+ },
315
+ {
316
+ "id": "reasoning-0",
317
+ "type": "reasoning-end",
318
+ },
319
+ {
320
+ "id": "1",
321
+ "type": "text-start",
322
+ },
323
+ {
324
+ "id": "1",
325
+ "providerMetadata": undefined,
326
+ "text": "Here",
327
+ "type": "text-delta",
328
+ },
329
+ {
330
+ "id": "1",
331
+ "providerMetadata": undefined,
332
+ "text": " is the response",
333
+ "type": "text-delta",
334
+ },
335
+ {
336
+ "id": "1",
337
+ "type": "text-end",
338
+ },
339
+ {
340
+ "finishReason": "stop",
341
+ "providerMetadata": undefined,
342
+ "rawFinishReason": "stop",
343
+ "response": {
344
+ "headers": undefined,
345
+ "id": "id-0",
346
+ "modelId": "mock-model-id",
347
+ "timestamp": 1970-01-01T00:00:00.000Z,
348
+ },
349
+ "type": "finish-step",
350
+ "usage": {
351
+ "cachedInputTokens": 0,
352
+ "inputTokenDetails": {
353
+ "cacheReadTokens": 0,
354
+ "cacheWriteTokens": 0,
355
+ "noCacheTokens": 5,
356
+ },
357
+ "inputTokens": 5,
358
+ "outputTokenDetails": {
359
+ "reasoningTokens": 3,
360
+ "textTokens": 10,
361
+ },
362
+ "outputTokens": 10,
363
+ "raw": undefined,
364
+ "reasoningTokens": 3,
365
+ "totalTokens": 15,
366
+ },
367
+ },
368
+ {
369
+ "finishReason": "stop",
370
+ "rawFinishReason": "stop",
371
+ "totalUsage": {
372
+ "cachedInputTokens": 0,
373
+ "inputTokenDetails": {
374
+ "cacheReadTokens": 0,
375
+ "cacheWriteTokens": 0,
376
+ "noCacheTokens": 5,
377
+ },
378
+ "inputTokens": 5,
379
+ "outputTokenDetails": {
380
+ "reasoningTokens": 3,
381
+ "textTokens": 10,
382
+ },
383
+ "outputTokens": 10,
384
+ "reasoningTokens": 3,
385
+ "totalTokens": 15,
386
+ },
387
+ "type": "finish",
388
+ },
389
+ ]
390
+ `);
391
+ });
392
+
393
+ it('should extract reasoning from single chunk with multiple <think> tags', async () => {
394
+ const mockModel = new MockLanguageModelV3({
395
+ async doStream() {
396
+ return {
397
+ stream: convertArrayToReadableStream([
398
+ {
399
+ type: 'response-metadata',
400
+ id: 'id-0',
401
+ modelId: 'mock-model-id',
402
+ timestamp: new Date(0),
403
+ },
404
+ { type: 'text-start', id: '1' },
405
+ {
406
+ type: 'text-delta',
407
+ id: '1',
408
+ delta:
409
+ '<think>analyzing the request</think>Here is the response<think>thinking about the response</think>more',
410
+ },
411
+ { type: 'text-end', id: '1' },
412
+ {
413
+ type: 'finish',
414
+ finishReason: { unified: 'stop', raw: 'stop' },
415
+ usage: testUsage,
416
+ },
417
+ ]),
418
+ };
419
+ },
420
+ });
421
+
422
+ const result = streamText({
423
+ model: wrapLanguageModel({
424
+ model: mockModel,
425
+ middleware: extractReasoningMiddleware({ tagName: 'think' }),
426
+ }),
427
+ prompt: 'Hello, how can I help?',
428
+ });
429
+
430
+ expect(await convertAsyncIterableToArray(result.fullStream))
431
+ .toMatchInlineSnapshot(`
432
+ [
433
+ {
434
+ "type": "start",
435
+ },
436
+ {
437
+ "request": {},
438
+ "type": "start-step",
439
+ "warnings": [],
440
+ },
441
+ {
442
+ "id": "reasoning-0",
443
+ "type": "reasoning-start",
444
+ },
445
+ {
446
+ "id": "reasoning-0",
447
+ "providerMetadata": undefined,
448
+ "text": "analyzing the request",
449
+ "type": "reasoning-delta",
450
+ },
451
+ {
452
+ "id": "reasoning-0",
453
+ "type": "reasoning-end",
454
+ },
455
+ {
456
+ "id": "1",
457
+ "type": "text-start",
458
+ },
459
+ {
460
+ "id": "1",
461
+ "providerMetadata": undefined,
462
+ "text": "Here is the response",
463
+ "type": "text-delta",
464
+ },
465
+ {
466
+ "id": "reasoning-1",
467
+ "type": "reasoning-start",
468
+ },
469
+ {
470
+ "id": "reasoning-1",
471
+ "providerMetadata": undefined,
472
+ "text": "
473
+ thinking about the response",
474
+ "type": "reasoning-delta",
475
+ },
476
+ {
477
+ "id": "reasoning-1",
478
+ "type": "reasoning-end",
479
+ },
480
+ {
481
+ "id": "1",
482
+ "providerMetadata": undefined,
483
+ "text": "
484
+ more",
485
+ "type": "text-delta",
486
+ },
487
+ {
488
+ "id": "1",
489
+ "type": "text-end",
490
+ },
491
+ {
492
+ "finishReason": "stop",
493
+ "providerMetadata": undefined,
494
+ "rawFinishReason": "stop",
495
+ "response": {
496
+ "headers": undefined,
497
+ "id": "id-0",
498
+ "modelId": "mock-model-id",
499
+ "timestamp": 1970-01-01T00:00:00.000Z,
500
+ },
501
+ "type": "finish-step",
502
+ "usage": {
503
+ "cachedInputTokens": 0,
504
+ "inputTokenDetails": {
505
+ "cacheReadTokens": 0,
506
+ "cacheWriteTokens": 0,
507
+ "noCacheTokens": 5,
508
+ },
509
+ "inputTokens": 5,
510
+ "outputTokenDetails": {
511
+ "reasoningTokens": 3,
512
+ "textTokens": 10,
513
+ },
514
+ "outputTokens": 10,
515
+ "raw": undefined,
516
+ "reasoningTokens": 3,
517
+ "totalTokens": 15,
518
+ },
519
+ },
520
+ {
521
+ "finishReason": "stop",
522
+ "rawFinishReason": "stop",
523
+ "totalUsage": {
524
+ "cachedInputTokens": 0,
525
+ "inputTokenDetails": {
526
+ "cacheReadTokens": 0,
527
+ "cacheWriteTokens": 0,
528
+ "noCacheTokens": 5,
529
+ },
530
+ "inputTokens": 5,
531
+ "outputTokenDetails": {
532
+ "reasoningTokens": 3,
533
+ "textTokens": 10,
534
+ },
535
+ "outputTokens": 10,
536
+ "reasoningTokens": 3,
537
+ "totalTokens": 15,
538
+ },
539
+ "type": "finish",
540
+ },
541
+ ]
542
+ `);
543
+ });
544
+
545
+ it('should extract reasoning from <think> when there is no text', async () => {
546
+ const mockModel = new MockLanguageModelV3({
547
+ async doStream() {
548
+ return {
549
+ stream: convertArrayToReadableStream([
550
+ {
551
+ type: 'response-metadata',
552
+ id: 'id-0',
553
+ modelId: 'mock-model-id',
554
+ timestamp: new Date(0),
555
+ },
556
+ { type: 'text-start', id: '1' },
557
+ { type: 'text-delta', id: '1', delta: '<think>' },
558
+ { type: 'text-delta', id: '1', delta: 'ana' },
559
+ { type: 'text-delta', id: '1', delta: 'lyzing the request\n' },
560
+ { type: 'text-delta', id: '1', delta: '</think>' },
561
+ { type: 'text-end', id: '1' },
562
+ {
563
+ type: 'finish',
564
+ finishReason: { unified: 'stop', raw: 'stop' },
565
+ usage: testUsage,
566
+ },
567
+ ]),
568
+ };
569
+ },
570
+ });
571
+
572
+ const result = streamText({
573
+ model: wrapLanguageModel({
574
+ model: mockModel,
575
+ middleware: extractReasoningMiddleware({ tagName: 'think' }),
576
+ }),
577
+ prompt: 'Hello, how can I help?',
578
+ });
579
+
580
+ expect(await convertAsyncIterableToArray(result.fullStream))
581
+ .toMatchInlineSnapshot(`
582
+ [
583
+ {
584
+ "type": "start",
585
+ },
586
+ {
587
+ "request": {},
588
+ "type": "start-step",
589
+ "warnings": [],
590
+ },
591
+ {
592
+ "id": "reasoning-0",
593
+ "type": "reasoning-start",
594
+ },
595
+ {
596
+ "id": "reasoning-0",
597
+ "providerMetadata": undefined,
598
+ "text": "ana",
599
+ "type": "reasoning-delta",
600
+ },
601
+ {
602
+ "id": "reasoning-0",
603
+ "providerMetadata": undefined,
604
+ "text": "lyzing the request
605
+ ",
606
+ "type": "reasoning-delta",
607
+ },
608
+ {
609
+ "id": "reasoning-0",
610
+ "type": "reasoning-end",
611
+ },
612
+ {
613
+ "id": "1",
614
+ "type": "text-start",
615
+ },
616
+ {
617
+ "id": "1",
618
+ "type": "text-end",
619
+ },
620
+ {
621
+ "finishReason": "stop",
622
+ "providerMetadata": undefined,
623
+ "rawFinishReason": "stop",
624
+ "response": {
625
+ "headers": undefined,
626
+ "id": "id-0",
627
+ "modelId": "mock-model-id",
628
+ "timestamp": 1970-01-01T00:00:00.000Z,
629
+ },
630
+ "type": "finish-step",
631
+ "usage": {
632
+ "cachedInputTokens": 0,
633
+ "inputTokenDetails": {
634
+ "cacheReadTokens": 0,
635
+ "cacheWriteTokens": 0,
636
+ "noCacheTokens": 5,
637
+ },
638
+ "inputTokens": 5,
639
+ "outputTokenDetails": {
640
+ "reasoningTokens": 3,
641
+ "textTokens": 10,
642
+ },
643
+ "outputTokens": 10,
644
+ "raw": undefined,
645
+ "reasoningTokens": 3,
646
+ "totalTokens": 15,
647
+ },
648
+ },
649
+ {
650
+ "finishReason": "stop",
651
+ "rawFinishReason": "stop",
652
+ "totalUsage": {
653
+ "cachedInputTokens": 0,
654
+ "inputTokenDetails": {
655
+ "cacheReadTokens": 0,
656
+ "cacheWriteTokens": 0,
657
+ "noCacheTokens": 5,
658
+ },
659
+ "inputTokens": 5,
660
+ "outputTokenDetails": {
661
+ "reasoningTokens": 3,
662
+ "textTokens": 10,
663
+ },
664
+ "outputTokens": 10,
665
+ "reasoningTokens": 3,
666
+ "totalTokens": 15,
667
+ },
668
+ "type": "finish",
669
+ },
670
+ ]
671
+ `);
672
+ });
673
+
674
+ it('should prepend <think> tag if startWithReasoning is true', async () => {
675
+ const mockModel = new MockLanguageModelV3({
676
+ async doStream() {
677
+ return {
678
+ stream: convertArrayToReadableStream([
679
+ {
680
+ type: 'response-metadata',
681
+ id: 'id-0',
682
+ modelId: 'mock-model-id',
683
+ timestamp: new Date(0),
684
+ },
685
+ { type: 'text-start', id: '1' },
686
+ { type: 'text-delta', id: '1', delta: 'ana' },
687
+ { type: 'text-delta', id: '1', delta: 'lyzing the request\n' },
688
+ { type: 'text-delta', id: '1', delta: '</think>' },
689
+ { type: 'text-delta', id: '1', delta: 'this is the response' },
690
+ { type: 'text-end', id: '1' },
691
+ {
692
+ type: 'finish',
693
+ finishReason: { unified: 'stop', raw: 'stop' },
694
+ usage: testUsage,
695
+ },
696
+ ]),
697
+ };
698
+ },
699
+ });
700
+
701
+ const resultTrue = streamText({
702
+ model: wrapLanguageModel({
703
+ model: mockModel,
704
+ middleware: extractReasoningMiddleware({
705
+ tagName: 'think',
706
+ startWithReasoning: true,
707
+ }),
708
+ }),
709
+ prompt: 'Hello, how can I help?',
710
+ });
711
+
712
+ const resultFalse = streamText({
713
+ model: wrapLanguageModel({
714
+ model: mockModel,
715
+ middleware: extractReasoningMiddleware({ tagName: 'think' }),
716
+ }),
717
+ prompt: 'Hello, how can I help?',
718
+ });
719
+
720
+ expect(await convertAsyncIterableToArray(resultTrue.fullStream))
721
+ .toMatchInlineSnapshot(`
722
+ [
723
+ {
724
+ "type": "start",
725
+ },
726
+ {
727
+ "request": {},
728
+ "type": "start-step",
729
+ "warnings": [],
730
+ },
731
+ {
732
+ "id": "reasoning-0",
733
+ "type": "reasoning-start",
734
+ },
735
+ {
736
+ "id": "reasoning-0",
737
+ "providerMetadata": undefined,
738
+ "text": "ana",
739
+ "type": "reasoning-delta",
740
+ },
741
+ {
742
+ "id": "reasoning-0",
743
+ "providerMetadata": undefined,
744
+ "text": "lyzing the request
745
+ ",
746
+ "type": "reasoning-delta",
747
+ },
748
+ {
749
+ "id": "reasoning-0",
750
+ "type": "reasoning-end",
751
+ },
752
+ {
753
+ "id": "1",
754
+ "type": "text-start",
755
+ },
756
+ {
757
+ "id": "1",
758
+ "providerMetadata": undefined,
759
+ "text": "this is the response",
760
+ "type": "text-delta",
761
+ },
762
+ {
763
+ "id": "1",
764
+ "type": "text-end",
765
+ },
766
+ {
767
+ "finishReason": "stop",
768
+ "providerMetadata": undefined,
769
+ "rawFinishReason": "stop",
770
+ "response": {
771
+ "headers": undefined,
772
+ "id": "id-0",
773
+ "modelId": "mock-model-id",
774
+ "timestamp": 1970-01-01T00:00:00.000Z,
775
+ },
776
+ "type": "finish-step",
777
+ "usage": {
778
+ "cachedInputTokens": 0,
779
+ "inputTokenDetails": {
780
+ "cacheReadTokens": 0,
781
+ "cacheWriteTokens": 0,
782
+ "noCacheTokens": 5,
783
+ },
784
+ "inputTokens": 5,
785
+ "outputTokenDetails": {
786
+ "reasoningTokens": 3,
787
+ "textTokens": 10,
788
+ },
789
+ "outputTokens": 10,
790
+ "raw": undefined,
791
+ "reasoningTokens": 3,
792
+ "totalTokens": 15,
793
+ },
794
+ },
795
+ {
796
+ "finishReason": "stop",
797
+ "rawFinishReason": "stop",
798
+ "totalUsage": {
799
+ "cachedInputTokens": 0,
800
+ "inputTokenDetails": {
801
+ "cacheReadTokens": 0,
802
+ "cacheWriteTokens": 0,
803
+ "noCacheTokens": 5,
804
+ },
805
+ "inputTokens": 5,
806
+ "outputTokenDetails": {
807
+ "reasoningTokens": 3,
808
+ "textTokens": 10,
809
+ },
810
+ "outputTokens": 10,
811
+ "reasoningTokens": 3,
812
+ "totalTokens": 15,
813
+ },
814
+ "type": "finish",
815
+ },
816
+ ]
817
+ `);
818
+
819
+ expect(await convertAsyncIterableToArray(resultFalse.fullStream))
820
+ .toMatchInlineSnapshot(`
821
+ [
822
+ {
823
+ "type": "start",
824
+ },
825
+ {
826
+ "request": {},
827
+ "type": "start-step",
828
+ "warnings": [],
829
+ },
830
+ {
831
+ "id": "1",
832
+ "type": "text-start",
833
+ },
834
+ {
835
+ "id": "1",
836
+ "providerMetadata": undefined,
837
+ "text": "ana",
838
+ "type": "text-delta",
839
+ },
840
+ {
841
+ "id": "1",
842
+ "providerMetadata": undefined,
843
+ "text": "lyzing the request
844
+ ",
845
+ "type": "text-delta",
846
+ },
847
+ {
848
+ "id": "1",
849
+ "providerMetadata": undefined,
850
+ "text": "</think>",
851
+ "type": "text-delta",
852
+ },
853
+ {
854
+ "id": "1",
855
+ "providerMetadata": undefined,
856
+ "text": "this is the response",
857
+ "type": "text-delta",
858
+ },
859
+ {
860
+ "id": "1",
861
+ "type": "text-end",
862
+ },
863
+ {
864
+ "finishReason": "stop",
865
+ "providerMetadata": undefined,
866
+ "rawFinishReason": "stop",
867
+ "response": {
868
+ "headers": undefined,
869
+ "id": "id-0",
870
+ "modelId": "mock-model-id",
871
+ "timestamp": 1970-01-01T00:00:00.000Z,
872
+ },
873
+ "type": "finish-step",
874
+ "usage": {
875
+ "cachedInputTokens": 0,
876
+ "inputTokenDetails": {
877
+ "cacheReadTokens": 0,
878
+ "cacheWriteTokens": 0,
879
+ "noCacheTokens": 5,
880
+ },
881
+ "inputTokens": 5,
882
+ "outputTokenDetails": {
883
+ "reasoningTokens": 3,
884
+ "textTokens": 10,
885
+ },
886
+ "outputTokens": 10,
887
+ "raw": undefined,
888
+ "reasoningTokens": 3,
889
+ "totalTokens": 15,
890
+ },
891
+ },
892
+ {
893
+ "finishReason": "stop",
894
+ "rawFinishReason": "stop",
895
+ "totalUsage": {
896
+ "cachedInputTokens": 0,
897
+ "inputTokenDetails": {
898
+ "cacheReadTokens": 0,
899
+ "cacheWriteTokens": 0,
900
+ "noCacheTokens": 5,
901
+ },
902
+ "inputTokens": 5,
903
+ "outputTokenDetails": {
904
+ "reasoningTokens": 3,
905
+ "textTokens": 10,
906
+ },
907
+ "outputTokens": 10,
908
+ "reasoningTokens": 3,
909
+ "totalTokens": 15,
910
+ },
911
+ "type": "finish",
912
+ },
913
+ ]
914
+ `);
915
+ });
916
+
917
+ it('should keep original text when <think> tag is not present', async () => {
918
+ const mockModel = new MockLanguageModelV3({
919
+ async doStream() {
920
+ return {
921
+ stream: convertArrayToReadableStream([
922
+ {
923
+ type: 'response-metadata',
924
+ id: 'id-0',
925
+ modelId: 'mock-model-id',
926
+ timestamp: new Date(0),
927
+ },
928
+ { type: 'text-start', id: '1' },
929
+ { type: 'text-delta', id: '1', delta: 'this is the response' },
930
+ { type: 'text-end', id: '1' },
931
+ {
932
+ type: 'finish',
933
+ finishReason: { unified: 'stop', raw: 'stop' },
934
+ usage: testUsage,
935
+ },
936
+ ]),
937
+ };
938
+ },
939
+ });
940
+
941
+ const result = streamText({
942
+ model: wrapLanguageModel({
943
+ model: mockModel,
944
+ middleware: extractReasoningMiddleware({ tagName: 'think' }),
945
+ }),
946
+ prompt: 'Hello, how can I help?',
947
+ });
948
+
949
+ expect(await convertAsyncIterableToArray(result.fullStream))
950
+ .toMatchInlineSnapshot(`
951
+ [
952
+ {
953
+ "type": "start",
954
+ },
955
+ {
956
+ "request": {},
957
+ "type": "start-step",
958
+ "warnings": [],
959
+ },
960
+ {
961
+ "id": "1",
962
+ "type": "text-start",
963
+ },
964
+ {
965
+ "id": "1",
966
+ "providerMetadata": undefined,
967
+ "text": "this is the response",
968
+ "type": "text-delta",
969
+ },
970
+ {
971
+ "id": "1",
972
+ "type": "text-end",
973
+ },
974
+ {
975
+ "finishReason": "stop",
976
+ "providerMetadata": undefined,
977
+ "rawFinishReason": "stop",
978
+ "response": {
979
+ "headers": undefined,
980
+ "id": "id-0",
981
+ "modelId": "mock-model-id",
982
+ "timestamp": 1970-01-01T00:00:00.000Z,
983
+ },
984
+ "type": "finish-step",
985
+ "usage": {
986
+ "cachedInputTokens": 0,
987
+ "inputTokenDetails": {
988
+ "cacheReadTokens": 0,
989
+ "cacheWriteTokens": 0,
990
+ "noCacheTokens": 5,
991
+ },
992
+ "inputTokens": 5,
993
+ "outputTokenDetails": {
994
+ "reasoningTokens": 3,
995
+ "textTokens": 10,
996
+ },
997
+ "outputTokens": 10,
998
+ "raw": undefined,
999
+ "reasoningTokens": 3,
1000
+ "totalTokens": 15,
1001
+ },
1002
+ },
1003
+ {
1004
+ "finishReason": "stop",
1005
+ "rawFinishReason": "stop",
1006
+ "totalUsage": {
1007
+ "cachedInputTokens": 0,
1008
+ "inputTokenDetails": {
1009
+ "cacheReadTokens": 0,
1010
+ "cacheWriteTokens": 0,
1011
+ "noCacheTokens": 5,
1012
+ },
1013
+ "inputTokens": 5,
1014
+ "outputTokenDetails": {
1015
+ "reasoningTokens": 3,
1016
+ "textTokens": 10,
1017
+ },
1018
+ "outputTokens": 10,
1019
+ "reasoningTokens": 3,
1020
+ "totalTokens": 15,
1021
+ },
1022
+ "type": "finish",
1023
+ },
1024
+ ]
1025
+ `);
1026
+ });
1027
+ });
1028
+ });