@livekit/agents 1.0.48 → 1.1.0-dev.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (373) hide show
  1. package/dist/constants.cjs +27 -0
  2. package/dist/constants.cjs.map +1 -1
  3. package/dist/constants.d.cts +9 -0
  4. package/dist/constants.d.ts +9 -0
  5. package/dist/constants.d.ts.map +1 -1
  6. package/dist/constants.js +18 -0
  7. package/dist/constants.js.map +1 -1
  8. package/dist/inference/api_protos.d.cts +71 -71
  9. package/dist/inference/api_protos.d.ts +71 -71
  10. package/dist/inference/interruption/defaults.cjs +81 -0
  11. package/dist/inference/interruption/defaults.cjs.map +1 -0
  12. package/dist/inference/interruption/defaults.d.cts +19 -0
  13. package/dist/inference/interruption/defaults.d.ts +19 -0
  14. package/dist/inference/interruption/defaults.d.ts.map +1 -0
  15. package/dist/inference/interruption/defaults.js +46 -0
  16. package/dist/inference/interruption/defaults.js.map +1 -0
  17. package/dist/inference/interruption/errors.cjs +44 -0
  18. package/dist/inference/interruption/errors.cjs.map +1 -0
  19. package/dist/inference/interruption/errors.d.cts +12 -0
  20. package/dist/inference/interruption/errors.d.ts +12 -0
  21. package/dist/inference/interruption/errors.d.ts.map +1 -0
  22. package/dist/inference/interruption/errors.js +20 -0
  23. package/dist/inference/interruption/errors.js.map +1 -0
  24. package/dist/inference/interruption/http_transport.cjs +147 -0
  25. package/dist/inference/interruption/http_transport.cjs.map +1 -0
  26. package/dist/inference/interruption/http_transport.d.cts +63 -0
  27. package/dist/inference/interruption/http_transport.d.ts +63 -0
  28. package/dist/inference/interruption/http_transport.d.ts.map +1 -0
  29. package/dist/inference/interruption/http_transport.js +121 -0
  30. package/dist/inference/interruption/http_transport.js.map +1 -0
  31. package/dist/inference/interruption/interruption_cache_entry.cjs +58 -0
  32. package/dist/inference/interruption/interruption_cache_entry.cjs.map +1 -0
  33. package/dist/inference/interruption/interruption_cache_entry.d.cts +30 -0
  34. package/dist/inference/interruption/interruption_cache_entry.d.ts +30 -0
  35. package/dist/inference/interruption/interruption_cache_entry.d.ts.map +1 -0
  36. package/dist/inference/interruption/interruption_cache_entry.js +34 -0
  37. package/dist/inference/interruption/interruption_cache_entry.js.map +1 -0
  38. package/dist/inference/interruption/interruption_detector.cjs +181 -0
  39. package/dist/inference/interruption/interruption_detector.cjs.map +1 -0
  40. package/dist/inference/interruption/interruption_detector.d.cts +59 -0
  41. package/dist/inference/interruption/interruption_detector.d.ts +59 -0
  42. package/dist/inference/interruption/interruption_detector.d.ts.map +1 -0
  43. package/dist/inference/interruption/interruption_detector.js +147 -0
  44. package/dist/inference/interruption/interruption_detector.js.map +1 -0
  45. package/dist/inference/interruption/interruption_stream.cjs +368 -0
  46. package/dist/inference/interruption/interruption_stream.cjs.map +1 -0
  47. package/dist/inference/interruption/interruption_stream.d.cts +46 -0
  48. package/dist/inference/interruption/interruption_stream.d.ts +46 -0
  49. package/dist/inference/interruption/interruption_stream.d.ts.map +1 -0
  50. package/dist/inference/interruption/interruption_stream.js +344 -0
  51. package/dist/inference/interruption/interruption_stream.js.map +1 -0
  52. package/dist/inference/interruption/types.cjs +17 -0
  53. package/dist/inference/interruption/types.cjs.map +1 -0
  54. package/dist/inference/interruption/types.d.cts +66 -0
  55. package/dist/inference/interruption/types.d.ts +66 -0
  56. package/dist/inference/interruption/types.d.ts.map +1 -0
  57. package/dist/inference/interruption/types.js +1 -0
  58. package/dist/inference/interruption/types.js.map +1 -0
  59. package/dist/inference/interruption/utils.cjs +130 -0
  60. package/dist/inference/interruption/utils.cjs.map +1 -0
  61. package/dist/inference/interruption/utils.d.cts +41 -0
  62. package/dist/inference/interruption/utils.d.ts +41 -0
  63. package/dist/inference/interruption/utils.d.ts.map +1 -0
  64. package/dist/inference/interruption/utils.js +105 -0
  65. package/dist/inference/interruption/utils.js.map +1 -0
  66. package/dist/inference/interruption/utils.test.cjs +105 -0
  67. package/dist/inference/interruption/utils.test.cjs.map +1 -0
  68. package/dist/inference/interruption/utils.test.js +104 -0
  69. package/dist/inference/interruption/utils.test.js.map +1 -0
  70. package/dist/inference/interruption/ws_transport.cjs +329 -0
  71. package/dist/inference/interruption/ws_transport.cjs.map +1 -0
  72. package/dist/inference/interruption/ws_transport.d.cts +33 -0
  73. package/dist/inference/interruption/ws_transport.d.ts +33 -0
  74. package/dist/inference/interruption/ws_transport.d.ts.map +1 -0
  75. package/dist/inference/interruption/ws_transport.js +295 -0
  76. package/dist/inference/interruption/ws_transport.js.map +1 -0
  77. package/dist/inference/llm.cjs +14 -10
  78. package/dist/inference/llm.cjs.map +1 -1
  79. package/dist/inference/llm.d.cts +2 -1
  80. package/dist/inference/llm.d.ts +2 -1
  81. package/dist/inference/llm.d.ts.map +1 -1
  82. package/dist/inference/llm.js +8 -10
  83. package/dist/inference/llm.js.map +1 -1
  84. package/dist/inference/stt.cjs +7 -2
  85. package/dist/inference/stt.cjs.map +1 -1
  86. package/dist/inference/stt.d.cts +2 -0
  87. package/dist/inference/stt.d.ts +2 -0
  88. package/dist/inference/stt.d.ts.map +1 -1
  89. package/dist/inference/stt.js +8 -3
  90. package/dist/inference/stt.js.map +1 -1
  91. package/dist/inference/tts.cjs +7 -2
  92. package/dist/inference/tts.cjs.map +1 -1
  93. package/dist/inference/tts.d.cts +2 -0
  94. package/dist/inference/tts.d.ts +2 -0
  95. package/dist/inference/tts.d.ts.map +1 -1
  96. package/dist/inference/tts.js +8 -3
  97. package/dist/inference/tts.js.map +1 -1
  98. package/dist/inference/utils.cjs +26 -7
  99. package/dist/inference/utils.cjs.map +1 -1
  100. package/dist/inference/utils.d.cts +13 -0
  101. package/dist/inference/utils.d.ts +13 -0
  102. package/dist/inference/utils.d.ts.map +1 -1
  103. package/dist/inference/utils.js +18 -2
  104. package/dist/inference/utils.js.map +1 -1
  105. package/dist/llm/chat_context.cjs +20 -2
  106. package/dist/llm/chat_context.cjs.map +1 -1
  107. package/dist/llm/chat_context.d.cts +19 -1
  108. package/dist/llm/chat_context.d.ts +19 -1
  109. package/dist/llm/chat_context.d.ts.map +1 -1
  110. package/dist/llm/chat_context.js +20 -2
  111. package/dist/llm/chat_context.js.map +1 -1
  112. package/dist/llm/index.cjs.map +1 -1
  113. package/dist/llm/index.d.cts +1 -1
  114. package/dist/llm/index.d.ts +1 -1
  115. package/dist/llm/index.d.ts.map +1 -1
  116. package/dist/llm/index.js.map +1 -1
  117. package/dist/llm/llm.cjs +16 -1
  118. package/dist/llm/llm.cjs.map +1 -1
  119. package/dist/llm/llm.d.cts +9 -0
  120. package/dist/llm/llm.d.ts +9 -0
  121. package/dist/llm/llm.d.ts.map +1 -1
  122. package/dist/llm/llm.js +16 -1
  123. package/dist/llm/llm.js.map +1 -1
  124. package/dist/llm/realtime.cjs +3 -0
  125. package/dist/llm/realtime.cjs.map +1 -1
  126. package/dist/llm/realtime.d.cts +1 -0
  127. package/dist/llm/realtime.d.ts +1 -0
  128. package/dist/llm/realtime.d.ts.map +1 -1
  129. package/dist/llm/realtime.js +3 -0
  130. package/dist/llm/realtime.js.map +1 -1
  131. package/dist/metrics/base.cjs.map +1 -1
  132. package/dist/metrics/base.d.cts +45 -1
  133. package/dist/metrics/base.d.ts +45 -1
  134. package/dist/metrics/base.d.ts.map +1 -1
  135. package/dist/metrics/index.cjs +5 -0
  136. package/dist/metrics/index.cjs.map +1 -1
  137. package/dist/metrics/index.d.cts +2 -1
  138. package/dist/metrics/index.d.ts +2 -1
  139. package/dist/metrics/index.d.ts.map +1 -1
  140. package/dist/metrics/index.js +6 -0
  141. package/dist/metrics/index.js.map +1 -1
  142. package/dist/metrics/model_usage.cjs +189 -0
  143. package/dist/metrics/model_usage.cjs.map +1 -0
  144. package/dist/metrics/model_usage.d.cts +92 -0
  145. package/dist/metrics/model_usage.d.ts +92 -0
  146. package/dist/metrics/model_usage.d.ts.map +1 -0
  147. package/dist/metrics/model_usage.js +164 -0
  148. package/dist/metrics/model_usage.js.map +1 -0
  149. package/dist/metrics/model_usage.test.cjs +474 -0
  150. package/dist/metrics/model_usage.test.cjs.map +1 -0
  151. package/dist/metrics/model_usage.test.js +476 -0
  152. package/dist/metrics/model_usage.test.js.map +1 -0
  153. package/dist/metrics/usage_collector.cjs +3 -0
  154. package/dist/metrics/usage_collector.cjs.map +1 -1
  155. package/dist/metrics/usage_collector.d.cts +9 -0
  156. package/dist/metrics/usage_collector.d.ts +9 -0
  157. package/dist/metrics/usage_collector.d.ts.map +1 -1
  158. package/dist/metrics/usage_collector.js +3 -0
  159. package/dist/metrics/usage_collector.js.map +1 -1
  160. package/dist/metrics/utils.cjs +9 -0
  161. package/dist/metrics/utils.cjs.map +1 -1
  162. package/dist/metrics/utils.d.ts.map +1 -1
  163. package/dist/metrics/utils.js +9 -0
  164. package/dist/metrics/utils.js.map +1 -1
  165. package/dist/stream/multi_input_stream.test.cjs +4 -0
  166. package/dist/stream/multi_input_stream.test.cjs.map +1 -1
  167. package/dist/stream/multi_input_stream.test.js +5 -1
  168. package/dist/stream/multi_input_stream.test.js.map +1 -1
  169. package/dist/stream/stream_channel.cjs +31 -0
  170. package/dist/stream/stream_channel.cjs.map +1 -1
  171. package/dist/stream/stream_channel.d.cts +4 -2
  172. package/dist/stream/stream_channel.d.ts +4 -2
  173. package/dist/stream/stream_channel.d.ts.map +1 -1
  174. package/dist/stream/stream_channel.js +31 -0
  175. package/dist/stream/stream_channel.js.map +1 -1
  176. package/dist/stt/stt.cjs +34 -2
  177. package/dist/stt/stt.cjs.map +1 -1
  178. package/dist/stt/stt.d.cts +22 -0
  179. package/dist/stt/stt.d.ts +22 -0
  180. package/dist/stt/stt.d.ts.map +1 -1
  181. package/dist/stt/stt.js +34 -2
  182. package/dist/stt/stt.js.map +1 -1
  183. package/dist/telemetry/otel_http_exporter.cjs +24 -5
  184. package/dist/telemetry/otel_http_exporter.cjs.map +1 -1
  185. package/dist/telemetry/otel_http_exporter.d.cts +1 -0
  186. package/dist/telemetry/otel_http_exporter.d.ts +1 -0
  187. package/dist/telemetry/otel_http_exporter.d.ts.map +1 -1
  188. package/dist/telemetry/otel_http_exporter.js +24 -5
  189. package/dist/telemetry/otel_http_exporter.js.map +1 -1
  190. package/dist/telemetry/trace_types.cjs +5 -5
  191. package/dist/telemetry/trace_types.cjs.map +1 -1
  192. package/dist/telemetry/trace_types.d.cts +9 -5
  193. package/dist/telemetry/trace_types.d.ts +9 -5
  194. package/dist/telemetry/trace_types.d.ts.map +1 -1
  195. package/dist/telemetry/trace_types.js +5 -5
  196. package/dist/telemetry/trace_types.js.map +1 -1
  197. package/dist/telemetry/traces.cjs +47 -8
  198. package/dist/telemetry/traces.cjs.map +1 -1
  199. package/dist/telemetry/traces.d.ts.map +1 -1
  200. package/dist/telemetry/traces.js +47 -8
  201. package/dist/telemetry/traces.js.map +1 -1
  202. package/dist/tts/tts.cjs +64 -2
  203. package/dist/tts/tts.cjs.map +1 -1
  204. package/dist/tts/tts.d.cts +34 -0
  205. package/dist/tts/tts.d.ts +34 -0
  206. package/dist/tts/tts.d.ts.map +1 -1
  207. package/dist/tts/tts.js +64 -2
  208. package/dist/tts/tts.js.map +1 -1
  209. package/dist/version.cjs +1 -1
  210. package/dist/version.js +1 -1
  211. package/dist/voice/agent.cjs +25 -4
  212. package/dist/voice/agent.cjs.map +1 -1
  213. package/dist/voice/agent.d.cts +10 -2
  214. package/dist/voice/agent.d.ts +10 -2
  215. package/dist/voice/agent.d.ts.map +1 -1
  216. package/dist/voice/agent.js +25 -4
  217. package/dist/voice/agent.js.map +1 -1
  218. package/dist/voice/agent_activity.cjs +261 -36
  219. package/dist/voice/agent_activity.cjs.map +1 -1
  220. package/dist/voice/agent_activity.d.cts +20 -6
  221. package/dist/voice/agent_activity.d.ts +20 -6
  222. package/dist/voice/agent_activity.d.ts.map +1 -1
  223. package/dist/voice/agent_activity.js +262 -37
  224. package/dist/voice/agent_activity.js.map +1 -1
  225. package/dist/voice/agent_session.cjs +105 -48
  226. package/dist/voice/agent_session.cjs.map +1 -1
  227. package/dist/voice/agent_session.d.cts +90 -20
  228. package/dist/voice/agent_session.d.ts +90 -20
  229. package/dist/voice/agent_session.d.ts.map +1 -1
  230. package/dist/voice/agent_session.js +105 -46
  231. package/dist/voice/agent_session.js.map +1 -1
  232. package/dist/voice/audio_recognition.cjs +287 -6
  233. package/dist/voice/audio_recognition.cjs.map +1 -1
  234. package/dist/voice/audio_recognition.d.cts +42 -3
  235. package/dist/voice/audio_recognition.d.ts +42 -3
  236. package/dist/voice/audio_recognition.d.ts.map +1 -1
  237. package/dist/voice/audio_recognition.js +289 -7
  238. package/dist/voice/audio_recognition.js.map +1 -1
  239. package/dist/voice/client_events.cjs +554 -0
  240. package/dist/voice/client_events.cjs.map +1 -0
  241. package/dist/voice/client_events.d.cts +195 -0
  242. package/dist/voice/client_events.d.ts +195 -0
  243. package/dist/voice/client_events.d.ts.map +1 -0
  244. package/dist/voice/client_events.js +548 -0
  245. package/dist/voice/client_events.js.map +1 -0
  246. package/dist/voice/events.cjs +1 -0
  247. package/dist/voice/events.cjs.map +1 -1
  248. package/dist/voice/events.d.cts +8 -5
  249. package/dist/voice/events.d.ts +8 -5
  250. package/dist/voice/events.d.ts.map +1 -1
  251. package/dist/voice/events.js +1 -0
  252. package/dist/voice/events.js.map +1 -1
  253. package/dist/voice/generation.cjs +43 -8
  254. package/dist/voice/generation.cjs.map +1 -1
  255. package/dist/voice/generation.d.cts +3 -3
  256. package/dist/voice/generation.d.ts +3 -3
  257. package/dist/voice/generation.d.ts.map +1 -1
  258. package/dist/voice/generation.js +43 -8
  259. package/dist/voice/generation.js.map +1 -1
  260. package/dist/voice/index.cjs +1 -0
  261. package/dist/voice/index.cjs.map +1 -1
  262. package/dist/voice/index.d.cts +1 -0
  263. package/dist/voice/index.d.ts +1 -0
  264. package/dist/voice/index.d.ts.map +1 -1
  265. package/dist/voice/index.js +1 -0
  266. package/dist/voice/index.js.map +1 -1
  267. package/dist/voice/report.cjs +20 -8
  268. package/dist/voice/report.cjs.map +1 -1
  269. package/dist/voice/report.d.cts +5 -0
  270. package/dist/voice/report.d.ts +5 -0
  271. package/dist/voice/report.d.ts.map +1 -1
  272. package/dist/voice/report.js +20 -8
  273. package/dist/voice/report.js.map +1 -1
  274. package/dist/voice/report.test.cjs +106 -0
  275. package/dist/voice/report.test.cjs.map +1 -0
  276. package/dist/voice/report.test.js +105 -0
  277. package/dist/voice/report.test.js.map +1 -0
  278. package/dist/voice/room_io/room_io.cjs +5 -39
  279. package/dist/voice/room_io/room_io.cjs.map +1 -1
  280. package/dist/voice/room_io/room_io.d.cts +4 -9
  281. package/dist/voice/room_io/room_io.d.ts +4 -9
  282. package/dist/voice/room_io/room_io.d.ts.map +1 -1
  283. package/dist/voice/room_io/room_io.js +5 -40
  284. package/dist/voice/room_io/room_io.js.map +1 -1
  285. package/dist/voice/turn_config/endpointing.cjs +33 -0
  286. package/dist/voice/turn_config/endpointing.cjs.map +1 -0
  287. package/dist/voice/turn_config/endpointing.d.cts +30 -0
  288. package/dist/voice/turn_config/endpointing.d.ts +30 -0
  289. package/dist/voice/turn_config/endpointing.d.ts.map +1 -0
  290. package/dist/voice/turn_config/endpointing.js +9 -0
  291. package/dist/voice/turn_config/endpointing.js.map +1 -0
  292. package/dist/voice/turn_config/interruption.cjs +37 -0
  293. package/dist/voice/turn_config/interruption.cjs.map +1 -0
  294. package/dist/voice/turn_config/interruption.d.cts +53 -0
  295. package/dist/voice/turn_config/interruption.d.ts +53 -0
  296. package/dist/voice/turn_config/interruption.d.ts.map +1 -0
  297. package/dist/voice/turn_config/interruption.js +13 -0
  298. package/dist/voice/turn_config/interruption.js.map +1 -0
  299. package/dist/voice/turn_config/turn_handling.cjs +35 -0
  300. package/dist/voice/turn_config/turn_handling.cjs.map +1 -0
  301. package/dist/voice/turn_config/turn_handling.d.cts +36 -0
  302. package/dist/voice/turn_config/turn_handling.d.ts +36 -0
  303. package/dist/voice/turn_config/turn_handling.d.ts.map +1 -0
  304. package/dist/voice/turn_config/turn_handling.js +11 -0
  305. package/dist/voice/turn_config/turn_handling.js.map +1 -0
  306. package/dist/voice/turn_config/utils.cjs +97 -0
  307. package/dist/voice/turn_config/utils.cjs.map +1 -0
  308. package/dist/voice/turn_config/utils.d.cts +25 -0
  309. package/dist/voice/turn_config/utils.d.ts +25 -0
  310. package/dist/voice/turn_config/utils.d.ts.map +1 -0
  311. package/dist/voice/turn_config/utils.js +73 -0
  312. package/dist/voice/turn_config/utils.js.map +1 -0
  313. package/dist/voice/turn_config/utils.test.cjs +86 -0
  314. package/dist/voice/turn_config/utils.test.cjs.map +1 -0
  315. package/dist/voice/turn_config/utils.test.js +85 -0
  316. package/dist/voice/turn_config/utils.test.js.map +1 -0
  317. package/dist/voice/wire_format.cjs +798 -0
  318. package/dist/voice/wire_format.cjs.map +1 -0
  319. package/dist/voice/wire_format.d.cts +5503 -0
  320. package/dist/voice/wire_format.d.ts +5503 -0
  321. package/dist/voice/wire_format.d.ts.map +1 -0
  322. package/dist/voice/wire_format.js +728 -0
  323. package/dist/voice/wire_format.js.map +1 -0
  324. package/package.json +2 -1
  325. package/src/constants.ts +13 -0
  326. package/src/inference/interruption/defaults.ts +51 -0
  327. package/src/inference/interruption/errors.ts +25 -0
  328. package/src/inference/interruption/http_transport.ts +187 -0
  329. package/src/inference/interruption/interruption_cache_entry.ts +50 -0
  330. package/src/inference/interruption/interruption_detector.ts +188 -0
  331. package/src/inference/interruption/interruption_stream.ts +467 -0
  332. package/src/inference/interruption/types.ts +84 -0
  333. package/src/inference/interruption/utils.test.ts +132 -0
  334. package/src/inference/interruption/utils.ts +137 -0
  335. package/src/inference/interruption/ws_transport.ts +402 -0
  336. package/src/inference/llm.ts +9 -12
  337. package/src/inference/stt.ts +10 -3
  338. package/src/inference/tts.ts +10 -3
  339. package/src/inference/utils.ts +29 -1
  340. package/src/llm/chat_context.ts +40 -2
  341. package/src/llm/index.ts +1 -0
  342. package/src/llm/llm.ts +16 -0
  343. package/src/llm/realtime.ts +4 -0
  344. package/src/metrics/base.ts +48 -1
  345. package/src/metrics/index.ts +11 -0
  346. package/src/metrics/model_usage.test.ts +545 -0
  347. package/src/metrics/model_usage.ts +262 -0
  348. package/src/metrics/usage_collector.ts +11 -0
  349. package/src/metrics/utils.ts +11 -0
  350. package/src/stream/multi_input_stream.test.ts +6 -1
  351. package/src/stream/stream_channel.ts +34 -2
  352. package/src/stt/stt.ts +38 -0
  353. package/src/telemetry/otel_http_exporter.ts +28 -5
  354. package/src/telemetry/trace_types.ts +11 -8
  355. package/src/telemetry/traces.ts +111 -54
  356. package/src/tts/tts.ts +69 -1
  357. package/src/voice/agent.ts +30 -3
  358. package/src/voice/agent_activity.ts +327 -28
  359. package/src/voice/agent_session.ts +207 -59
  360. package/src/voice/audio_recognition.ts +385 -9
  361. package/src/voice/client_events.ts +838 -0
  362. package/src/voice/events.ts +14 -4
  363. package/src/voice/generation.ts +52 -9
  364. package/src/voice/index.ts +1 -0
  365. package/src/voice/report.test.ts +117 -0
  366. package/src/voice/report.ts +29 -6
  367. package/src/voice/room_io/room_io.ts +7 -61
  368. package/src/voice/turn_config/endpointing.ts +33 -0
  369. package/src/voice/turn_config/interruption.ts +56 -0
  370. package/src/voice/turn_config/turn_handling.ts +45 -0
  371. package/src/voice/turn_config/utils.test.ts +100 -0
  372. package/src/voice/turn_config/utils.ts +103 -0
  373. package/src/voice/wire_format.ts +827 -0
@@ -0,0 +1,476 @@
1
+ import { beforeEach, describe, expect, it } from "vitest";
2
+ import {
3
+ ModelUsageCollector,
4
+ filterZeroValues
5
+ } from "./model_usage.js";
6
+ describe("model_usage", () => {
7
+ describe("filterZeroValues", () => {
8
+ it("should filter out zero values from LLMModelUsage", () => {
9
+ const usage = {
10
+ type: "llm_usage",
11
+ provider: "openai",
12
+ model: "gpt-4o",
13
+ inputTokens: 100,
14
+ inputCachedTokens: 0,
15
+ inputAudioTokens: 0,
16
+ inputCachedAudioTokens: 0,
17
+ inputTextTokens: 0,
18
+ inputCachedTextTokens: 0,
19
+ inputImageTokens: 0,
20
+ inputCachedImageTokens: 0,
21
+ outputTokens: 50,
22
+ outputAudioTokens: 0,
23
+ outputTextTokens: 0,
24
+ sessionDurationMs: 0
25
+ };
26
+ const filtered = filterZeroValues(usage);
27
+ expect(filtered.type).toBe("llm_usage");
28
+ expect(filtered.provider).toBe("openai");
29
+ expect(filtered.model).toBe("gpt-4o");
30
+ expect(filtered.inputTokens).toBe(100);
31
+ expect(filtered.outputTokens).toBe(50);
32
+ expect(filtered.inputCachedTokens).toBeUndefined();
33
+ expect(filtered.inputAudioTokens).toBeUndefined();
34
+ expect(filtered.sessionDurationMs).toBeUndefined();
35
+ });
36
+ it("should filter out zero values from TTSModelUsage", () => {
37
+ const usage = {
38
+ type: "tts_usage",
39
+ provider: "elevenlabs",
40
+ model: "eleven_turbo_v2",
41
+ inputTokens: 0,
42
+ outputTokens: 0,
43
+ charactersCount: 500,
44
+ audioDurationMs: 3e3
45
+ };
46
+ const filtered = filterZeroValues(usage);
47
+ expect(filtered.type).toBe("tts_usage");
48
+ expect(filtered.provider).toBe("elevenlabs");
49
+ expect(filtered.charactersCount).toBe(500);
50
+ expect(filtered.audioDurationMs).toBe(3e3);
51
+ expect(filtered.inputTokens).toBeUndefined();
52
+ expect(filtered.outputTokens).toBeUndefined();
53
+ });
54
+ it("should keep all values when none are zero", () => {
55
+ const usage = {
56
+ type: "stt_usage",
57
+ provider: "deepgram",
58
+ model: "nova-2",
59
+ inputTokens: 10,
60
+ outputTokens: 20,
61
+ audioDurationMs: 5e3
62
+ };
63
+ const filtered = filterZeroValues(usage);
64
+ expect(Object.keys(filtered)).toHaveLength(6);
65
+ expect(filtered).toEqual(usage);
66
+ });
67
+ });
68
+ describe("ModelUsageCollector", () => {
69
+ let collector;
70
+ beforeEach(() => {
71
+ collector = new ModelUsageCollector();
72
+ });
73
+ describe("collect LLM metrics", () => {
74
+ it("should aggregate LLM metrics by provider and model", () => {
75
+ const metrics1 = {
76
+ type: "llm_metrics",
77
+ label: "test",
78
+ requestId: "req1",
79
+ timestamp: Date.now(),
80
+ durationMs: 100,
81
+ ttftMs: 50,
82
+ cancelled: false,
83
+ completionTokens: 100,
84
+ promptTokens: 200,
85
+ promptCachedTokens: 50,
86
+ totalTokens: 300,
87
+ tokensPerSecond: 10,
88
+ metadata: {
89
+ modelProvider: "openai",
90
+ modelName: "gpt-4o"
91
+ }
92
+ };
93
+ const metrics2 = {
94
+ type: "llm_metrics",
95
+ label: "test",
96
+ requestId: "req2",
97
+ timestamp: Date.now(),
98
+ durationMs: 150,
99
+ ttftMs: 60,
100
+ cancelled: false,
101
+ completionTokens: 150,
102
+ promptTokens: 300,
103
+ promptCachedTokens: 75,
104
+ totalTokens: 450,
105
+ tokensPerSecond: 12,
106
+ metadata: {
107
+ modelProvider: "openai",
108
+ modelName: "gpt-4o"
109
+ }
110
+ };
111
+ collector.collect(metrics1);
112
+ collector.collect(metrics2);
113
+ const usage = collector.flatten();
114
+ expect(usage).toHaveLength(1);
115
+ const llmUsage = usage[0];
116
+ expect(llmUsage.type).toBe("llm_usage");
117
+ expect(llmUsage.provider).toBe("openai");
118
+ expect(llmUsage.model).toBe("gpt-4o");
119
+ expect(llmUsage.inputTokens).toBe(500);
120
+ expect(llmUsage.inputCachedTokens).toBe(125);
121
+ expect(llmUsage.outputTokens).toBe(250);
122
+ });
123
+ it("should separate metrics by different providers", () => {
124
+ const openaiMetrics = {
125
+ type: "llm_metrics",
126
+ label: "test",
127
+ requestId: "req1",
128
+ timestamp: Date.now(),
129
+ durationMs: 100,
130
+ ttftMs: 50,
131
+ cancelled: false,
132
+ completionTokens: 100,
133
+ promptTokens: 200,
134
+ promptCachedTokens: 0,
135
+ totalTokens: 300,
136
+ tokensPerSecond: 10,
137
+ metadata: {
138
+ modelProvider: "openai",
139
+ modelName: "gpt-4o"
140
+ }
141
+ };
142
+ const anthropicMetrics = {
143
+ type: "llm_metrics",
144
+ label: "test",
145
+ requestId: "req2",
146
+ timestamp: Date.now(),
147
+ durationMs: 120,
148
+ ttftMs: 55,
149
+ cancelled: false,
150
+ completionTokens: 80,
151
+ promptTokens: 150,
152
+ promptCachedTokens: 0,
153
+ totalTokens: 230,
154
+ tokensPerSecond: 8,
155
+ metadata: {
156
+ modelProvider: "anthropic",
157
+ modelName: "claude-3-5-sonnet"
158
+ }
159
+ };
160
+ collector.collect(openaiMetrics);
161
+ collector.collect(anthropicMetrics);
162
+ const usage = collector.flatten();
163
+ expect(usage).toHaveLength(2);
164
+ const openaiUsage = usage.find(
165
+ (u) => u.type === "llm_usage" && u.provider === "openai"
166
+ );
167
+ const anthropicUsage = usage.find(
168
+ (u) => u.type === "llm_usage" && u.provider === "anthropic"
169
+ );
170
+ expect(openaiUsage.inputTokens).toBe(200);
171
+ expect(openaiUsage.outputTokens).toBe(100);
172
+ expect(anthropicUsage.inputTokens).toBe(150);
173
+ expect(anthropicUsage.outputTokens).toBe(80);
174
+ });
175
+ });
176
+ describe("collect TTS metrics", () => {
177
+ it("should aggregate TTS metrics by provider and model", () => {
178
+ const metrics1 = {
179
+ type: "tts_metrics",
180
+ label: "test",
181
+ requestId: "req1",
182
+ timestamp: Date.now(),
183
+ ttfbMs: 100,
184
+ durationMs: 500,
185
+ audioDurationMs: 3e3,
186
+ cancelled: false,
187
+ charactersCount: 100,
188
+ inputTokens: 10,
189
+ outputTokens: 20,
190
+ streamed: true,
191
+ metadata: {
192
+ modelProvider: "elevenlabs",
193
+ modelName: "eleven_turbo_v2"
194
+ }
195
+ };
196
+ const metrics2 = {
197
+ type: "tts_metrics",
198
+ label: "test",
199
+ requestId: "req2",
200
+ timestamp: Date.now(),
201
+ ttfbMs: 120,
202
+ durationMs: 600,
203
+ audioDurationMs: 4e3,
204
+ cancelled: false,
205
+ charactersCount: 200,
206
+ inputTokens: 15,
207
+ outputTokens: 25,
208
+ streamed: true,
209
+ metadata: {
210
+ modelProvider: "elevenlabs",
211
+ modelName: "eleven_turbo_v2"
212
+ }
213
+ };
214
+ collector.collect(metrics1);
215
+ collector.collect(metrics2);
216
+ const usage = collector.flatten();
217
+ expect(usage).toHaveLength(1);
218
+ const ttsUsage = usage[0];
219
+ expect(ttsUsage.type).toBe("tts_usage");
220
+ expect(ttsUsage.provider).toBe("elevenlabs");
221
+ expect(ttsUsage.model).toBe("eleven_turbo_v2");
222
+ expect(ttsUsage.charactersCount).toBe(300);
223
+ expect(ttsUsage.audioDurationMs).toBe(7e3);
224
+ expect(ttsUsage.inputTokens).toBe(25);
225
+ expect(ttsUsage.outputTokens).toBe(45);
226
+ });
227
+ });
228
+ describe("collect STT metrics", () => {
229
+ it("should aggregate STT metrics by provider and model", () => {
230
+ const metrics1 = {
231
+ type: "stt_metrics",
232
+ label: "test",
233
+ requestId: "req1",
234
+ timestamp: Date.now(),
235
+ durationMs: 0,
236
+ audioDurationMs: 5e3,
237
+ inputTokens: 50,
238
+ outputTokens: 100,
239
+ streamed: true,
240
+ metadata: {
241
+ modelProvider: "deepgram",
242
+ modelName: "nova-2"
243
+ }
244
+ };
245
+ const metrics2 = {
246
+ type: "stt_metrics",
247
+ label: "test",
248
+ requestId: "req2",
249
+ timestamp: Date.now(),
250
+ durationMs: 0,
251
+ audioDurationMs: 3e3,
252
+ inputTokens: 30,
253
+ outputTokens: 60,
254
+ streamed: true,
255
+ metadata: {
256
+ modelProvider: "deepgram",
257
+ modelName: "nova-2"
258
+ }
259
+ };
260
+ collector.collect(metrics1);
261
+ collector.collect(metrics2);
262
+ const usage = collector.flatten();
263
+ expect(usage).toHaveLength(1);
264
+ const sttUsage = usage[0];
265
+ expect(sttUsage.type).toBe("stt_usage");
266
+ expect(sttUsage.provider).toBe("deepgram");
267
+ expect(sttUsage.model).toBe("nova-2");
268
+ expect(sttUsage.audioDurationMs).toBe(8e3);
269
+ expect(sttUsage.inputTokens).toBe(80);
270
+ expect(sttUsage.outputTokens).toBe(160);
271
+ });
272
+ });
273
+ describe("collect realtime model metrics", () => {
274
+ it("should aggregate realtime model metrics with detailed token breakdown", () => {
275
+ const metrics = {
276
+ type: "realtime_model_metrics",
277
+ label: "test",
278
+ requestId: "req1",
279
+ timestamp: Date.now(),
280
+ durationMs: 1e3,
281
+ ttftMs: 100,
282
+ cancelled: false,
283
+ inputTokens: 500,
284
+ outputTokens: 300,
285
+ totalTokens: 800,
286
+ tokensPerSecond: 10,
287
+ sessionDurationMs: 5e3,
288
+ inputTokenDetails: {
289
+ audioTokens: 200,
290
+ textTokens: 250,
291
+ imageTokens: 50,
292
+ cachedTokens: 100,
293
+ cachedTokensDetails: {
294
+ audioTokens: 30,
295
+ textTokens: 50,
296
+ imageTokens: 20
297
+ }
298
+ },
299
+ outputTokenDetails: {
300
+ textTokens: 200,
301
+ audioTokens: 100,
302
+ imageTokens: 0
303
+ },
304
+ metadata: {
305
+ modelProvider: "openai",
306
+ modelName: "gpt-4o-realtime"
307
+ }
308
+ };
309
+ collector.collect(metrics);
310
+ const usage = collector.flatten();
311
+ expect(usage).toHaveLength(1);
312
+ const llmUsage = usage[0];
313
+ expect(llmUsage.type).toBe("llm_usage");
314
+ expect(llmUsage.provider).toBe("openai");
315
+ expect(llmUsage.model).toBe("gpt-4o-realtime");
316
+ expect(llmUsage.inputTokens).toBe(500);
317
+ expect(llmUsage.inputCachedTokens).toBe(100);
318
+ expect(llmUsage.inputAudioTokens).toBe(200);
319
+ expect(llmUsage.inputCachedAudioTokens).toBe(30);
320
+ expect(llmUsage.inputTextTokens).toBe(250);
321
+ expect(llmUsage.inputCachedTextTokens).toBe(50);
322
+ expect(llmUsage.inputImageTokens).toBe(50);
323
+ expect(llmUsage.inputCachedImageTokens).toBe(20);
324
+ expect(llmUsage.outputTokens).toBe(300);
325
+ expect(llmUsage.outputTextTokens).toBe(200);
326
+ expect(llmUsage.outputAudioTokens).toBe(100);
327
+ expect(llmUsage.sessionDurationMs).toBe(5e3);
328
+ });
329
+ });
330
+ describe("mixed metrics collection", () => {
331
+ it("should collect and separate LLM, TTS, and STT metrics", () => {
332
+ const llmMetrics = {
333
+ type: "llm_metrics",
334
+ label: "test",
335
+ requestId: "req1",
336
+ timestamp: Date.now(),
337
+ durationMs: 100,
338
+ ttftMs: 50,
339
+ cancelled: false,
340
+ completionTokens: 100,
341
+ promptTokens: 200,
342
+ promptCachedTokens: 0,
343
+ totalTokens: 300,
344
+ tokensPerSecond: 10,
345
+ metadata: {
346
+ modelProvider: "openai",
347
+ modelName: "gpt-4o"
348
+ }
349
+ };
350
+ const ttsMetrics = {
351
+ type: "tts_metrics",
352
+ label: "test",
353
+ requestId: "req2",
354
+ timestamp: Date.now(),
355
+ ttfbMs: 100,
356
+ durationMs: 500,
357
+ audioDurationMs: 3e3,
358
+ cancelled: false,
359
+ charactersCount: 100,
360
+ streamed: true,
361
+ metadata: {
362
+ modelProvider: "elevenlabs",
363
+ modelName: "eleven_turbo_v2"
364
+ }
365
+ };
366
+ const sttMetrics = {
367
+ type: "stt_metrics",
368
+ label: "test",
369
+ requestId: "req3",
370
+ timestamp: Date.now(),
371
+ durationMs: 0,
372
+ audioDurationMs: 5e3,
373
+ streamed: true,
374
+ metadata: {
375
+ modelProvider: "deepgram",
376
+ modelName: "nova-2"
377
+ }
378
+ };
379
+ collector.collect(llmMetrics);
380
+ collector.collect(ttsMetrics);
381
+ collector.collect(sttMetrics);
382
+ const usage = collector.flatten();
383
+ expect(usage).toHaveLength(3);
384
+ const llmUsage = usage.find((u) => u.type === "llm_usage");
385
+ const ttsUsage = usage.find((u) => u.type === "tts_usage");
386
+ const sttUsage = usage.find((u) => u.type === "stt_usage");
387
+ expect(llmUsage).toBeDefined();
388
+ expect(ttsUsage).toBeDefined();
389
+ expect(sttUsage).toBeDefined();
390
+ });
391
+ });
392
+ describe("flatten returns copies", () => {
393
+ it("should return deep copies of usage objects", () => {
394
+ const metrics = {
395
+ type: "llm_metrics",
396
+ label: "test",
397
+ requestId: "req1",
398
+ timestamp: Date.now(),
399
+ durationMs: 100,
400
+ ttftMs: 50,
401
+ cancelled: false,
402
+ completionTokens: 100,
403
+ promptTokens: 200,
404
+ promptCachedTokens: 0,
405
+ totalTokens: 300,
406
+ tokensPerSecond: 10,
407
+ metadata: {
408
+ modelProvider: "openai",
409
+ modelName: "gpt-4o"
410
+ }
411
+ };
412
+ collector.collect(metrics);
413
+ const usage1 = collector.flatten();
414
+ const usage2 = collector.flatten();
415
+ expect(usage1[0]).toEqual(usage2[0]);
416
+ expect(usage1[0]).not.toBe(usage2[0]);
417
+ usage1[0].inputTokens = 9999;
418
+ expect(usage2[0].inputTokens).toBe(200);
419
+ });
420
+ });
421
+ describe("handles missing metadata", () => {
422
+ it("should use empty strings when metadata is missing", () => {
423
+ const metrics = {
424
+ type: "llm_metrics",
425
+ label: "test",
426
+ requestId: "req1",
427
+ timestamp: Date.now(),
428
+ durationMs: 100,
429
+ ttftMs: 50,
430
+ cancelled: false,
431
+ completionTokens: 100,
432
+ promptTokens: 200,
433
+ promptCachedTokens: 0,
434
+ totalTokens: 300,
435
+ tokensPerSecond: 10
436
+ // No metadata
437
+ };
438
+ collector.collect(metrics);
439
+ const usage = collector.flatten();
440
+ expect(usage).toHaveLength(1);
441
+ const llmUsage = usage[0];
442
+ expect(llmUsage.provider).toBe("");
443
+ expect(llmUsage.model).toBe("");
444
+ });
445
+ });
446
+ describe("ignores VAD and EOU metrics", () => {
447
+ it("should not collect VAD metrics", () => {
448
+ const vadMetrics = {
449
+ type: "vad_metrics",
450
+ label: "test",
451
+ timestamp: Date.now(),
452
+ idleTimeMs: 100,
453
+ inferenceDurationTotalMs: 50,
454
+ inferenceCount: 10
455
+ };
456
+ collector.collect(vadMetrics);
457
+ const usage = collector.flatten();
458
+ expect(usage).toHaveLength(0);
459
+ });
460
+ it("should not collect EOU metrics", () => {
461
+ const eouMetrics = {
462
+ type: "eou_metrics",
463
+ timestamp: Date.now(),
464
+ endOfUtteranceDelayMs: 100,
465
+ transcriptionDelayMs: 50,
466
+ onUserTurnCompletedDelayMs: 30,
467
+ lastSpeakingTimeMs: Date.now()
468
+ };
469
+ collector.collect(eouMetrics);
470
+ const usage = collector.flatten();
471
+ expect(usage).toHaveLength(0);
472
+ });
473
+ });
474
+ });
475
+ });
476
+ //# sourceMappingURL=model_usage.test.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"sources":["../../src/metrics/model_usage.test.ts"],"sourcesContent":["// SPDX-FileCopyrightText: 2024 LiveKit, Inc.\n//\n// SPDX-License-Identifier: Apache-2.0\nimport { beforeEach, describe, expect, it } from 'vitest';\nimport type { LLMMetrics, RealtimeModelMetrics, STTMetrics, TTSMetrics } from './base.js';\nimport {\n type LLMModelUsage,\n ModelUsageCollector,\n type STTModelUsage,\n type TTSModelUsage,\n filterZeroValues,\n} from './model_usage.js';\n\ndescribe('model_usage', () => {\n describe('filterZeroValues', () => {\n it('should filter out zero values from LLMModelUsage', () => {\n const usage: LLMModelUsage = {\n type: 'llm_usage',\n provider: 'openai',\n model: 'gpt-4o',\n inputTokens: 100,\n inputCachedTokens: 0,\n inputAudioTokens: 0,\n inputCachedAudioTokens: 0,\n inputTextTokens: 0,\n inputCachedTextTokens: 0,\n inputImageTokens: 0,\n inputCachedImageTokens: 0,\n outputTokens: 50,\n outputAudioTokens: 0,\n outputTextTokens: 0,\n sessionDurationMs: 0,\n };\n\n const filtered = filterZeroValues(usage);\n\n expect(filtered.type).toBe('llm_usage');\n expect(filtered.provider).toBe('openai');\n expect(filtered.model).toBe('gpt-4o');\n expect(filtered.inputTokens).toBe(100);\n expect(filtered.outputTokens).toBe(50);\n // Zero values should be filtered out\n expect(filtered.inputCachedTokens).toBeUndefined();\n expect(filtered.inputAudioTokens).toBeUndefined();\n expect(filtered.sessionDurationMs).toBeUndefined();\n });\n\n it('should filter out zero values from TTSModelUsage', () => {\n const usage: TTSModelUsage = {\n type: 'tts_usage',\n provider: 'elevenlabs',\n model: 'eleven_turbo_v2',\n inputTokens: 0,\n outputTokens: 0,\n charactersCount: 500,\n audioDurationMs: 3000,\n };\n\n const filtered = filterZeroValues(usage);\n\n expect(filtered.type).toBe('tts_usage');\n expect(filtered.provider).toBe('elevenlabs');\n expect(filtered.charactersCount).toBe(500);\n expect(filtered.audioDurationMs).toBe(3000);\n expect(filtered.inputTokens).toBeUndefined();\n expect(filtered.outputTokens).toBeUndefined();\n });\n\n it('should keep all values when none are zero', () => {\n const usage: STTModelUsage = {\n type: 'stt_usage',\n provider: 'deepgram',\n model: 'nova-2',\n inputTokens: 10,\n outputTokens: 20,\n audioDurationMs: 5000,\n };\n\n const filtered = filterZeroValues(usage);\n\n expect(Object.keys(filtered)).toHaveLength(6);\n expect(filtered).toEqual(usage);\n });\n });\n\n describe('ModelUsageCollector', () => {\n let collector: ModelUsageCollector;\n\n beforeEach(() => {\n collector = new ModelUsageCollector();\n });\n\n describe('collect LLM metrics', () => {\n it('should aggregate LLM metrics by provider and model', () => {\n const metrics1: LLMMetrics = {\n type: 'llm_metrics',\n label: 'test',\n requestId: 'req1',\n timestamp: Date.now(),\n durationMs: 100,\n ttftMs: 50,\n cancelled: false,\n completionTokens: 100,\n promptTokens: 200,\n promptCachedTokens: 50,\n totalTokens: 300,\n tokensPerSecond: 10,\n metadata: {\n modelProvider: 'openai',\n modelName: 'gpt-4o',\n },\n };\n\n const metrics2: LLMMetrics = {\n type: 'llm_metrics',\n label: 'test',\n requestId: 'req2',\n timestamp: Date.now(),\n durationMs: 150,\n ttftMs: 60,\n cancelled: false,\n completionTokens: 150,\n promptTokens: 300,\n promptCachedTokens: 75,\n totalTokens: 450,\n tokensPerSecond: 12,\n metadata: {\n modelProvider: 'openai',\n modelName: 'gpt-4o',\n },\n };\n\n collector.collect(metrics1);\n collector.collect(metrics2);\n\n const usage = collector.flatten();\n expect(usage).toHaveLength(1);\n\n const llmUsage = usage[0] as LLMModelUsage;\n expect(llmUsage.type).toBe('llm_usage');\n expect(llmUsage.provider).toBe('openai');\n expect(llmUsage.model).toBe('gpt-4o');\n expect(llmUsage.inputTokens).toBe(500); // 200 + 300\n expect(llmUsage.inputCachedTokens).toBe(125); // 50 + 75\n expect(llmUsage.outputTokens).toBe(250); // 100 + 150\n });\n\n it('should separate metrics by different providers', () => {\n const openaiMetrics: LLMMetrics = {\n type: 'llm_metrics',\n label: 'test',\n requestId: 'req1',\n timestamp: Date.now(),\n durationMs: 100,\n ttftMs: 50,\n cancelled: false,\n completionTokens: 100,\n promptTokens: 200,\n promptCachedTokens: 0,\n totalTokens: 300,\n tokensPerSecond: 10,\n metadata: {\n modelProvider: 'openai',\n modelName: 'gpt-4o',\n },\n };\n\n const anthropicMetrics: LLMMetrics = {\n type: 'llm_metrics',\n label: 'test',\n requestId: 'req2',\n timestamp: Date.now(),\n durationMs: 120,\n ttftMs: 55,\n cancelled: false,\n completionTokens: 80,\n promptTokens: 150,\n promptCachedTokens: 0,\n totalTokens: 230,\n tokensPerSecond: 8,\n metadata: {\n modelProvider: 'anthropic',\n modelName: 'claude-3-5-sonnet',\n },\n };\n\n collector.collect(openaiMetrics);\n collector.collect(anthropicMetrics);\n\n const usage = collector.flatten();\n expect(usage).toHaveLength(2);\n\n const openaiUsage = usage.find(\n (u) => u.type === 'llm_usage' && u.provider === 'openai',\n ) as LLMModelUsage;\n const anthropicUsage = usage.find(\n (u) => u.type === 'llm_usage' && u.provider === 'anthropic',\n ) as LLMModelUsage;\n\n expect(openaiUsage.inputTokens).toBe(200);\n expect(openaiUsage.outputTokens).toBe(100);\n expect(anthropicUsage.inputTokens).toBe(150);\n expect(anthropicUsage.outputTokens).toBe(80);\n });\n });\n\n describe('collect TTS metrics', () => {\n it('should aggregate TTS metrics by provider and model', () => {\n const metrics1: TTSMetrics = {\n type: 'tts_metrics',\n label: 'test',\n requestId: 'req1',\n timestamp: Date.now(),\n ttfbMs: 100,\n durationMs: 500,\n audioDurationMs: 3000,\n cancelled: false,\n charactersCount: 100,\n inputTokens: 10,\n outputTokens: 20,\n streamed: true,\n metadata: {\n modelProvider: 'elevenlabs',\n modelName: 'eleven_turbo_v2',\n },\n };\n\n const metrics2: TTSMetrics = {\n type: 'tts_metrics',\n label: 'test',\n requestId: 'req2',\n timestamp: Date.now(),\n ttfbMs: 120,\n durationMs: 600,\n audioDurationMs: 4000,\n cancelled: false,\n charactersCount: 200,\n inputTokens: 15,\n outputTokens: 25,\n streamed: true,\n metadata: {\n modelProvider: 'elevenlabs',\n modelName: 'eleven_turbo_v2',\n },\n };\n\n collector.collect(metrics1);\n collector.collect(metrics2);\n\n const usage = collector.flatten();\n expect(usage).toHaveLength(1);\n\n const ttsUsage = usage[0] as TTSModelUsage;\n expect(ttsUsage.type).toBe('tts_usage');\n expect(ttsUsage.provider).toBe('elevenlabs');\n expect(ttsUsage.model).toBe('eleven_turbo_v2');\n expect(ttsUsage.charactersCount).toBe(300); // 100 + 200\n expect(ttsUsage.audioDurationMs).toBe(7000); // 3000 + 4000\n expect(ttsUsage.inputTokens).toBe(25); // 10 + 15\n expect(ttsUsage.outputTokens).toBe(45); // 20 + 25\n });\n });\n\n describe('collect STT metrics', () => {\n it('should aggregate STT metrics by provider and model', () => {\n const metrics1: STTMetrics = {\n type: 'stt_metrics',\n label: 'test',\n requestId: 'req1',\n timestamp: Date.now(),\n durationMs: 0,\n audioDurationMs: 5000,\n inputTokens: 50,\n outputTokens: 100,\n streamed: true,\n metadata: {\n modelProvider: 'deepgram',\n modelName: 'nova-2',\n },\n };\n\n const metrics2: STTMetrics = {\n type: 'stt_metrics',\n label: 'test',\n requestId: 'req2',\n timestamp: Date.now(),\n durationMs: 0,\n audioDurationMs: 3000,\n inputTokens: 30,\n outputTokens: 60,\n streamed: true,\n metadata: {\n modelProvider: 'deepgram',\n modelName: 'nova-2',\n },\n };\n\n collector.collect(metrics1);\n collector.collect(metrics2);\n\n const usage = collector.flatten();\n expect(usage).toHaveLength(1);\n\n const sttUsage = usage[0] as STTModelUsage;\n expect(sttUsage.type).toBe('stt_usage');\n expect(sttUsage.provider).toBe('deepgram');\n expect(sttUsage.model).toBe('nova-2');\n expect(sttUsage.audioDurationMs).toBe(8000); // 5000 + 3000\n expect(sttUsage.inputTokens).toBe(80); // 50 + 30\n expect(sttUsage.outputTokens).toBe(160); // 100 + 60\n });\n });\n\n describe('collect realtime model metrics', () => {\n it('should aggregate realtime model metrics with detailed token breakdown', () => {\n const metrics: RealtimeModelMetrics = {\n type: 'realtime_model_metrics',\n label: 'test',\n requestId: 'req1',\n timestamp: Date.now(),\n durationMs: 1000,\n ttftMs: 100,\n cancelled: false,\n inputTokens: 500,\n outputTokens: 300,\n totalTokens: 800,\n tokensPerSecond: 10,\n sessionDurationMs: 5000,\n inputTokenDetails: {\n audioTokens: 200,\n textTokens: 250,\n imageTokens: 50,\n cachedTokens: 100,\n cachedTokensDetails: {\n audioTokens: 30,\n textTokens: 50,\n imageTokens: 20,\n },\n },\n outputTokenDetails: {\n textTokens: 200,\n audioTokens: 100,\n imageTokens: 0,\n },\n metadata: {\n modelProvider: 'openai',\n modelName: 'gpt-4o-realtime',\n },\n };\n\n collector.collect(metrics);\n\n const usage = collector.flatten();\n expect(usage).toHaveLength(1);\n\n const llmUsage = usage[0] as LLMModelUsage;\n expect(llmUsage.type).toBe('llm_usage');\n expect(llmUsage.provider).toBe('openai');\n expect(llmUsage.model).toBe('gpt-4o-realtime');\n expect(llmUsage.inputTokens).toBe(500);\n expect(llmUsage.inputCachedTokens).toBe(100);\n expect(llmUsage.inputAudioTokens).toBe(200);\n expect(llmUsage.inputCachedAudioTokens).toBe(30);\n expect(llmUsage.inputTextTokens).toBe(250);\n expect(llmUsage.inputCachedTextTokens).toBe(50);\n expect(llmUsage.inputImageTokens).toBe(50);\n expect(llmUsage.inputCachedImageTokens).toBe(20);\n expect(llmUsage.outputTokens).toBe(300);\n expect(llmUsage.outputTextTokens).toBe(200);\n expect(llmUsage.outputAudioTokens).toBe(100);\n expect(llmUsage.sessionDurationMs).toBe(5000);\n });\n });\n\n describe('mixed metrics collection', () => {\n it('should collect and separate LLM, TTS, and STT metrics', () => {\n const llmMetrics: LLMMetrics = {\n type: 'llm_metrics',\n label: 'test',\n requestId: 'req1',\n timestamp: Date.now(),\n durationMs: 100,\n ttftMs: 50,\n cancelled: false,\n completionTokens: 100,\n promptTokens: 200,\n promptCachedTokens: 0,\n totalTokens: 300,\n tokensPerSecond: 10,\n metadata: {\n modelProvider: 'openai',\n modelName: 'gpt-4o',\n },\n };\n\n const ttsMetrics: TTSMetrics = {\n type: 'tts_metrics',\n label: 'test',\n requestId: 'req2',\n timestamp: Date.now(),\n ttfbMs: 100,\n durationMs: 500,\n audioDurationMs: 3000,\n cancelled: false,\n charactersCount: 100,\n streamed: true,\n metadata: {\n modelProvider: 'elevenlabs',\n modelName: 'eleven_turbo_v2',\n },\n };\n\n const sttMetrics: STTMetrics = {\n type: 'stt_metrics',\n label: 'test',\n requestId: 'req3',\n timestamp: Date.now(),\n durationMs: 0,\n audioDurationMs: 5000,\n streamed: true,\n metadata: {\n modelProvider: 'deepgram',\n modelName: 'nova-2',\n },\n };\n\n collector.collect(llmMetrics);\n collector.collect(ttsMetrics);\n collector.collect(sttMetrics);\n\n const usage = collector.flatten();\n expect(usage).toHaveLength(3);\n\n const llmUsage = usage.find((u) => u.type === 'llm_usage');\n const ttsUsage = usage.find((u) => u.type === 'tts_usage');\n const sttUsage = usage.find((u) => u.type === 'stt_usage');\n\n expect(llmUsage).toBeDefined();\n expect(ttsUsage).toBeDefined();\n expect(sttUsage).toBeDefined();\n });\n });\n\n describe('flatten returns copies', () => {\n it('should return deep copies of usage objects', () => {\n const metrics: LLMMetrics = {\n type: 'llm_metrics',\n label: 'test',\n requestId: 'req1',\n timestamp: Date.now(),\n durationMs: 100,\n ttftMs: 50,\n cancelled: false,\n completionTokens: 100,\n promptTokens: 200,\n promptCachedTokens: 0,\n totalTokens: 300,\n tokensPerSecond: 10,\n metadata: {\n modelProvider: 'openai',\n modelName: 'gpt-4o',\n },\n };\n\n collector.collect(metrics);\n\n const usage1 = collector.flatten();\n const usage2 = collector.flatten();\n\n // Should be equal values\n expect(usage1[0]).toEqual(usage2[0]);\n\n // But not the same object reference\n expect(usage1[0]).not.toBe(usage2[0]);\n\n // Modifying one shouldn't affect the other\n (usage1[0] as LLMModelUsage).inputTokens = 9999;\n expect((usage2[0] as LLMModelUsage).inputTokens).toBe(200);\n });\n });\n\n describe('handles missing metadata', () => {\n it('should use empty strings when metadata is missing', () => {\n const metrics: LLMMetrics = {\n type: 'llm_metrics',\n label: 'test',\n requestId: 'req1',\n timestamp: Date.now(),\n durationMs: 100,\n ttftMs: 50,\n cancelled: false,\n completionTokens: 100,\n promptTokens: 200,\n promptCachedTokens: 0,\n totalTokens: 300,\n tokensPerSecond: 10,\n // No metadata\n };\n\n collector.collect(metrics);\n\n const usage = collector.flatten();\n expect(usage).toHaveLength(1);\n\n const llmUsage = usage[0] as LLMModelUsage;\n expect(llmUsage.provider).toBe('');\n expect(llmUsage.model).toBe('');\n });\n });\n\n describe('ignores VAD and EOU metrics', () => {\n it('should not collect VAD metrics', () => {\n const vadMetrics = {\n type: 'vad_metrics' as const,\n label: 'test',\n timestamp: Date.now(),\n idleTimeMs: 100,\n inferenceDurationTotalMs: 50,\n inferenceCount: 10,\n };\n\n collector.collect(vadMetrics);\n\n const usage = collector.flatten();\n expect(usage).toHaveLength(0);\n });\n\n it('should not collect EOU metrics', () => {\n const eouMetrics = {\n type: 'eou_metrics' as const,\n timestamp: Date.now(),\n endOfUtteranceDelayMs: 100,\n transcriptionDelayMs: 50,\n onUserTurnCompletedDelayMs: 30,\n lastSpeakingTimeMs: Date.now(),\n };\n\n collector.collect(eouMetrics);\n\n const usage = collector.flatten();\n expect(usage).toHaveLength(0);\n });\n });\n });\n});\n"],"mappings":"AAGA,SAAS,YAAY,UAAU,QAAQ,UAAU;AAEjD;AAAA,EAEE;AAAA,EAGA;AAAA,OACK;AAEP,SAAS,eAAe,MAAM;AAC5B,WAAS,oBAAoB,MAAM;AACjC,OAAG,oDAAoD,MAAM;AAC3D,YAAM,QAAuB;AAAA,QAC3B,MAAM;AAAA,QACN,UAAU;AAAA,QACV,OAAO;AAAA,QACP,aAAa;AAAA,QACb,mBAAmB;AAAA,QACnB,kBAAkB;AAAA,QAClB,wBAAwB;AAAA,QACxB,iBAAiB;AAAA,QACjB,uBAAuB;AAAA,QACvB,kBAAkB;AAAA,QAClB,wBAAwB;AAAA,QACxB,cAAc;AAAA,QACd,mBAAmB;AAAA,QACnB,kBAAkB;AAAA,QAClB,mBAAmB;AAAA,MACrB;AAEA,YAAM,WAAW,iBAAiB,KAAK;AAEvC,aAAO,SAAS,IAAI,EAAE,KAAK,WAAW;AACtC,aAAO,SAAS,QAAQ,EAAE,KAAK,QAAQ;AACvC,aAAO,SAAS,KAAK,EAAE,KAAK,QAAQ;AACpC,aAAO,SAAS,WAAW,EAAE,KAAK,GAAG;AACrC,aAAO,SAAS,YAAY,EAAE,KAAK,EAAE;AAErC,aAAO,SAAS,iBAAiB,EAAE,cAAc;AACjD,aAAO,SAAS,gBAAgB,EAAE,cAAc;AAChD,aAAO,SAAS,iBAAiB,EAAE,cAAc;AAAA,IACnD,CAAC;AAED,OAAG,oDAAoD,MAAM;AAC3D,YAAM,QAAuB;AAAA,QAC3B,MAAM;AAAA,QACN,UAAU;AAAA,QACV,OAAO;AAAA,QACP,aAAa;AAAA,QACb,cAAc;AAAA,QACd,iBAAiB;AAAA,QACjB,iBAAiB;AAAA,MACnB;AAEA,YAAM,WAAW,iBAAiB,KAAK;AAEvC,aAAO,SAAS,IAAI,EAAE,KAAK,WAAW;AACtC,aAAO,SAAS,QAAQ,EAAE,KAAK,YAAY;AAC3C,aAAO,SAAS,eAAe,EAAE,KAAK,GAAG;AACzC,aAAO,SAAS,eAAe,EAAE,KAAK,GAAI;AAC1C,aAAO,SAAS,WAAW,EAAE,cAAc;AAC3C,aAAO,SAAS,YAAY,EAAE,cAAc;AAAA,IAC9C,CAAC;AAED,OAAG,6CAA6C,MAAM;AACpD,YAAM,QAAuB;AAAA,QAC3B,MAAM;AAAA,QACN,UAAU;AAAA,QACV,OAAO;AAAA,QACP,aAAa;AAAA,QACb,cAAc;AAAA,QACd,iBAAiB;AAAA,MACnB;AAEA,YAAM,WAAW,iBAAiB,KAAK;AAEvC,aAAO,OAAO,KAAK,QAAQ,CAAC,EAAE,aAAa,CAAC;AAC5C,aAAO,QAAQ,EAAE,QAAQ,KAAK;AAAA,IAChC,CAAC;AAAA,EACH,CAAC;AAED,WAAS,uBAAuB,MAAM;AACpC,QAAI;AAEJ,eAAW,MAAM;AACf,kBAAY,IAAI,oBAAoB;AAAA,IACtC,CAAC;AAED,aAAS,uBAAuB,MAAM;AACpC,SAAG,sDAAsD,MAAM;AAC7D,cAAM,WAAuB;AAAA,UAC3B,MAAM;AAAA,UACN,OAAO;AAAA,UACP,WAAW;AAAA,UACX,WAAW,KAAK,IAAI;AAAA,UACpB,YAAY;AAAA,UACZ,QAAQ;AAAA,UACR,WAAW;AAAA,UACX,kBAAkB;AAAA,UAClB,cAAc;AAAA,UACd,oBAAoB;AAAA,UACpB,aAAa;AAAA,UACb,iBAAiB;AAAA,UACjB,UAAU;AAAA,YACR,eAAe;AAAA,YACf,WAAW;AAAA,UACb;AAAA,QACF;AAEA,cAAM,WAAuB;AAAA,UAC3B,MAAM;AAAA,UACN,OAAO;AAAA,UACP,WAAW;AAAA,UACX,WAAW,KAAK,IAAI;AAAA,UACpB,YAAY;AAAA,UACZ,QAAQ;AAAA,UACR,WAAW;AAAA,UACX,kBAAkB;AAAA,UAClB,cAAc;AAAA,UACd,oBAAoB;AAAA,UACpB,aAAa;AAAA,UACb,iBAAiB;AAAA,UACjB,UAAU;AAAA,YACR,eAAe;AAAA,YACf,WAAW;AAAA,UACb;AAAA,QACF;AAEA,kBAAU,QAAQ,QAAQ;AAC1B,kBAAU,QAAQ,QAAQ;AAE1B,cAAM,QAAQ,UAAU,QAAQ;AAChC,eAAO,KAAK,EAAE,aAAa,CAAC;AAE5B,cAAM,WAAW,MAAM,CAAC;AACxB,eAAO,SAAS,IAAI,EAAE,KAAK,WAAW;AACtC,eAAO,SAAS,QAAQ,EAAE,KAAK,QAAQ;AACvC,eAAO,SAAS,KAAK,EAAE,KAAK,QAAQ;AACpC,eAAO,SAAS,WAAW,EAAE,KAAK,GAAG;AACrC,eAAO,SAAS,iBAAiB,EAAE,KAAK,GAAG;AAC3C,eAAO,SAAS,YAAY,EAAE,KAAK,GAAG;AAAA,MACxC,CAAC;AAED,SAAG,kDAAkD,MAAM;AACzD,cAAM,gBAA4B;AAAA,UAChC,MAAM;AAAA,UACN,OAAO;AAAA,UACP,WAAW;AAAA,UACX,WAAW,KAAK,IAAI;AAAA,UACpB,YAAY;AAAA,UACZ,QAAQ;AAAA,UACR,WAAW;AAAA,UACX,kBAAkB;AAAA,UAClB,cAAc;AAAA,UACd,oBAAoB;AAAA,UACpB,aAAa;AAAA,UACb,iBAAiB;AAAA,UACjB,UAAU;AAAA,YACR,eAAe;AAAA,YACf,WAAW;AAAA,UACb;AAAA,QACF;AAEA,cAAM,mBAA+B;AAAA,UACnC,MAAM;AAAA,UACN,OAAO;AAAA,UACP,WAAW;AAAA,UACX,WAAW,KAAK,IAAI;AAAA,UACpB,YAAY;AAAA,UACZ,QAAQ;AAAA,UACR,WAAW;AAAA,UACX,kBAAkB;AAAA,UAClB,cAAc;AAAA,UACd,oBAAoB;AAAA,UACpB,aAAa;AAAA,UACb,iBAAiB;AAAA,UACjB,UAAU;AAAA,YACR,eAAe;AAAA,YACf,WAAW;AAAA,UACb;AAAA,QACF;AAEA,kBAAU,QAAQ,aAAa;AAC/B,kBAAU,QAAQ,gBAAgB;AAElC,cAAM,QAAQ,UAAU,QAAQ;AAChC,eAAO,KAAK,EAAE,aAAa,CAAC;AAE5B,cAAM,cAAc,MAAM;AAAA,UACxB,CAAC,MAAM,EAAE,SAAS,eAAe,EAAE,aAAa;AAAA,QAClD;AACA,cAAM,iBAAiB,MAAM;AAAA,UAC3B,CAAC,MAAM,EAAE,SAAS,eAAe,EAAE,aAAa;AAAA,QAClD;AAEA,eAAO,YAAY,WAAW,EAAE,KAAK,GAAG;AACxC,eAAO,YAAY,YAAY,EAAE,KAAK,GAAG;AACzC,eAAO,eAAe,WAAW,EAAE,KAAK,GAAG;AAC3C,eAAO,eAAe,YAAY,EAAE,KAAK,EAAE;AAAA,MAC7C,CAAC;AAAA,IACH,CAAC;AAED,aAAS,uBAAuB,MAAM;AACpC,SAAG,sDAAsD,MAAM;AAC7D,cAAM,WAAuB;AAAA,UAC3B,MAAM;AAAA,UACN,OAAO;AAAA,UACP,WAAW;AAAA,UACX,WAAW,KAAK,IAAI;AAAA,UACpB,QAAQ;AAAA,UACR,YAAY;AAAA,UACZ,iBAAiB;AAAA,UACjB,WAAW;AAAA,UACX,iBAAiB;AAAA,UACjB,aAAa;AAAA,UACb,cAAc;AAAA,UACd,UAAU;AAAA,UACV,UAAU;AAAA,YACR,eAAe;AAAA,YACf,WAAW;AAAA,UACb;AAAA,QACF;AAEA,cAAM,WAAuB;AAAA,UAC3B,MAAM;AAAA,UACN,OAAO;AAAA,UACP,WAAW;AAAA,UACX,WAAW,KAAK,IAAI;AAAA,UACpB,QAAQ;AAAA,UACR,YAAY;AAAA,UACZ,iBAAiB;AAAA,UACjB,WAAW;AAAA,UACX,iBAAiB;AAAA,UACjB,aAAa;AAAA,UACb,cAAc;AAAA,UACd,UAAU;AAAA,UACV,UAAU;AAAA,YACR,eAAe;AAAA,YACf,WAAW;AAAA,UACb;AAAA,QACF;AAEA,kBAAU,QAAQ,QAAQ;AAC1B,kBAAU,QAAQ,QAAQ;AAE1B,cAAM,QAAQ,UAAU,QAAQ;AAChC,eAAO,KAAK,EAAE,aAAa,CAAC;AAE5B,cAAM,WAAW,MAAM,CAAC;AACxB,eAAO,SAAS,IAAI,EAAE,KAAK,WAAW;AACtC,eAAO,SAAS,QAAQ,EAAE,KAAK,YAAY;AAC3C,eAAO,SAAS,KAAK,EAAE,KAAK,iBAAiB;AAC7C,eAAO,SAAS,eAAe,EAAE,KAAK,GAAG;AACzC,eAAO,SAAS,eAAe,EAAE,KAAK,GAAI;AAC1C,eAAO,SAAS,WAAW,EAAE,KAAK,EAAE;AACpC,eAAO,SAAS,YAAY,EAAE,KAAK,EAAE;AAAA,MACvC,CAAC;AAAA,IACH,CAAC;AAED,aAAS,uBAAuB,MAAM;AACpC,SAAG,sDAAsD,MAAM;AAC7D,cAAM,WAAuB;AAAA,UAC3B,MAAM;AAAA,UACN,OAAO;AAAA,UACP,WAAW;AAAA,UACX,WAAW,KAAK,IAAI;AAAA,UACpB,YAAY;AAAA,UACZ,iBAAiB;AAAA,UACjB,aAAa;AAAA,UACb,cAAc;AAAA,UACd,UAAU;AAAA,UACV,UAAU;AAAA,YACR,eAAe;AAAA,YACf,WAAW;AAAA,UACb;AAAA,QACF;AAEA,cAAM,WAAuB;AAAA,UAC3B,MAAM;AAAA,UACN,OAAO;AAAA,UACP,WAAW;AAAA,UACX,WAAW,KAAK,IAAI;AAAA,UACpB,YAAY;AAAA,UACZ,iBAAiB;AAAA,UACjB,aAAa;AAAA,UACb,cAAc;AAAA,UACd,UAAU;AAAA,UACV,UAAU;AAAA,YACR,eAAe;AAAA,YACf,WAAW;AAAA,UACb;AAAA,QACF;AAEA,kBAAU,QAAQ,QAAQ;AAC1B,kBAAU,QAAQ,QAAQ;AAE1B,cAAM,QAAQ,UAAU,QAAQ;AAChC,eAAO,KAAK,EAAE,aAAa,CAAC;AAE5B,cAAM,WAAW,MAAM,CAAC;AACxB,eAAO,SAAS,IAAI,EAAE,KAAK,WAAW;AACtC,eAAO,SAAS,QAAQ,EAAE,KAAK,UAAU;AACzC,eAAO,SAAS,KAAK,EAAE,KAAK,QAAQ;AACpC,eAAO,SAAS,eAAe,EAAE,KAAK,GAAI;AAC1C,eAAO,SAAS,WAAW,EAAE,KAAK,EAAE;AACpC,eAAO,SAAS,YAAY,EAAE,KAAK,GAAG;AAAA,MACxC,CAAC;AAAA,IACH,CAAC;AAED,aAAS,kCAAkC,MAAM;AAC/C,SAAG,yEAAyE,MAAM;AAChF,cAAM,UAAgC;AAAA,UACpC,MAAM;AAAA,UACN,OAAO;AAAA,UACP,WAAW;AAAA,UACX,WAAW,KAAK,IAAI;AAAA,UACpB,YAAY;AAAA,UACZ,QAAQ;AAAA,UACR,WAAW;AAAA,UACX,aAAa;AAAA,UACb,cAAc;AAAA,UACd,aAAa;AAAA,UACb,iBAAiB;AAAA,UACjB,mBAAmB;AAAA,UACnB,mBAAmB;AAAA,YACjB,aAAa;AAAA,YACb,YAAY;AAAA,YACZ,aAAa;AAAA,YACb,cAAc;AAAA,YACd,qBAAqB;AAAA,cACnB,aAAa;AAAA,cACb,YAAY;AAAA,cACZ,aAAa;AAAA,YACf;AAAA,UACF;AAAA,UACA,oBAAoB;AAAA,YAClB,YAAY;AAAA,YACZ,aAAa;AAAA,YACb,aAAa;AAAA,UACf;AAAA,UACA,UAAU;AAAA,YACR,eAAe;AAAA,YACf,WAAW;AAAA,UACb;AAAA,QACF;AAEA,kBAAU,QAAQ,OAAO;AAEzB,cAAM,QAAQ,UAAU,QAAQ;AAChC,eAAO,KAAK,EAAE,aAAa,CAAC;AAE5B,cAAM,WAAW,MAAM,CAAC;AACxB,eAAO,SAAS,IAAI,EAAE,KAAK,WAAW;AACtC,eAAO,SAAS,QAAQ,EAAE,KAAK,QAAQ;AACvC,eAAO,SAAS,KAAK,EAAE,KAAK,iBAAiB;AAC7C,eAAO,SAAS,WAAW,EAAE,KAAK,GAAG;AACrC,eAAO,SAAS,iBAAiB,EAAE,KAAK,GAAG;AAC3C,eAAO,SAAS,gBAAgB,EAAE,KAAK,GAAG;AAC1C,eAAO,SAAS,sBAAsB,EAAE,KAAK,EAAE;AAC/C,eAAO,SAAS,eAAe,EAAE,KAAK,GAAG;AACzC,eAAO,SAAS,qBAAqB,EAAE,KAAK,EAAE;AAC9C,eAAO,SAAS,gBAAgB,EAAE,KAAK,EAAE;AACzC,eAAO,SAAS,sBAAsB,EAAE,KAAK,EAAE;AAC/C,eAAO,SAAS,YAAY,EAAE,KAAK,GAAG;AACtC,eAAO,SAAS,gBAAgB,EAAE,KAAK,GAAG;AAC1C,eAAO,SAAS,iBAAiB,EAAE,KAAK,GAAG;AAC3C,eAAO,SAAS,iBAAiB,EAAE,KAAK,GAAI;AAAA,MAC9C,CAAC;AAAA,IACH,CAAC;AAED,aAAS,4BAA4B,MAAM;AACzC,SAAG,yDAAyD,MAAM;AAChE,cAAM,aAAyB;AAAA,UAC7B,MAAM;AAAA,UACN,OAAO;AAAA,UACP,WAAW;AAAA,UACX,WAAW,KAAK,IAAI;AAAA,UACpB,YAAY;AAAA,UACZ,QAAQ;AAAA,UACR,WAAW;AAAA,UACX,kBAAkB;AAAA,UAClB,cAAc;AAAA,UACd,oBAAoB;AAAA,UACpB,aAAa;AAAA,UACb,iBAAiB;AAAA,UACjB,UAAU;AAAA,YACR,eAAe;AAAA,YACf,WAAW;AAAA,UACb;AAAA,QACF;AAEA,cAAM,aAAyB;AAAA,UAC7B,MAAM;AAAA,UACN,OAAO;AAAA,UACP,WAAW;AAAA,UACX,WAAW,KAAK,IAAI;AAAA,UACpB,QAAQ;AAAA,UACR,YAAY;AAAA,UACZ,iBAAiB;AAAA,UACjB,WAAW;AAAA,UACX,iBAAiB;AAAA,UACjB,UAAU;AAAA,UACV,UAAU;AAAA,YACR,eAAe;AAAA,YACf,WAAW;AAAA,UACb;AAAA,QACF;AAEA,cAAM,aAAyB;AAAA,UAC7B,MAAM;AAAA,UACN,OAAO;AAAA,UACP,WAAW;AAAA,UACX,WAAW,KAAK,IAAI;AAAA,UACpB,YAAY;AAAA,UACZ,iBAAiB;AAAA,UACjB,UAAU;AAAA,UACV,UAAU;AAAA,YACR,eAAe;AAAA,YACf,WAAW;AAAA,UACb;AAAA,QACF;AAEA,kBAAU,QAAQ,UAAU;AAC5B,kBAAU,QAAQ,UAAU;AAC5B,kBAAU,QAAQ,UAAU;AAE5B,cAAM,QAAQ,UAAU,QAAQ;AAChC,eAAO,KAAK,EAAE,aAAa,CAAC;AAE5B,cAAM,WAAW,MAAM,KAAK,CAAC,MAAM,EAAE,SAAS,WAAW;AACzD,cAAM,WAAW,MAAM,KAAK,CAAC,MAAM,EAAE,SAAS,WAAW;AACzD,cAAM,WAAW,MAAM,KAAK,CAAC,MAAM,EAAE,SAAS,WAAW;AAEzD,eAAO,QAAQ,EAAE,YAAY;AAC7B,eAAO,QAAQ,EAAE,YAAY;AAC7B,eAAO,QAAQ,EAAE,YAAY;AAAA,MAC/B,CAAC;AAAA,IACH,CAAC;AAED,aAAS,0BAA0B,MAAM;AACvC,SAAG,8CAA8C,MAAM;AACrD,cAAM,UAAsB;AAAA,UAC1B,MAAM;AAAA,UACN,OAAO;AAAA,UACP,WAAW;AAAA,UACX,WAAW,KAAK,IAAI;AAAA,UACpB,YAAY;AAAA,UACZ,QAAQ;AAAA,UACR,WAAW;AAAA,UACX,kBAAkB;AAAA,UAClB,cAAc;AAAA,UACd,oBAAoB;AAAA,UACpB,aAAa;AAAA,UACb,iBAAiB;AAAA,UACjB,UAAU;AAAA,YACR,eAAe;AAAA,YACf,WAAW;AAAA,UACb;AAAA,QACF;AAEA,kBAAU,QAAQ,OAAO;AAEzB,cAAM,SAAS,UAAU,QAAQ;AACjC,cAAM,SAAS,UAAU,QAAQ;AAGjC,eAAO,OAAO,CAAC,CAAC,EAAE,QAAQ,OAAO,CAAC,CAAC;AAGnC,eAAO,OAAO,CAAC,CAAC,EAAE,IAAI,KAAK,OAAO,CAAC,CAAC;AAGpC,QAAC,OAAO,CAAC,EAAoB,cAAc;AAC3C,eAAQ,OAAO,CAAC,EAAoB,WAAW,EAAE,KAAK,GAAG;AAAA,MAC3D,CAAC;AAAA,IACH,CAAC;AAED,aAAS,4BAA4B,MAAM;AACzC,SAAG,qDAAqD,MAAM;AAC5D,cAAM,UAAsB;AAAA,UAC1B,MAAM;AAAA,UACN,OAAO;AAAA,UACP,WAAW;AAAA,UACX,WAAW,KAAK,IAAI;AAAA,UACpB,YAAY;AAAA,UACZ,QAAQ;AAAA,UACR,WAAW;AAAA,UACX,kBAAkB;AAAA,UAClB,cAAc;AAAA,UACd,oBAAoB;AAAA,UACpB,aAAa;AAAA,UACb,iBAAiB;AAAA;AAAA,QAEnB;AAEA,kBAAU,QAAQ,OAAO;AAEzB,cAAM,QAAQ,UAAU,QAAQ;AAChC,eAAO,KAAK,EAAE,aAAa,CAAC;AAE5B,cAAM,WAAW,MAAM,CAAC;AACxB,eAAO,SAAS,QAAQ,EAAE,KAAK,EAAE;AACjC,eAAO,SAAS,KAAK,EAAE,KAAK,EAAE;AAAA,MAChC,CAAC;AAAA,IACH,CAAC;AAED,aAAS,+BAA+B,MAAM;AAC5C,SAAG,kCAAkC,MAAM;AACzC,cAAM,aAAa;AAAA,UACjB,MAAM;AAAA,UACN,OAAO;AAAA,UACP,WAAW,KAAK,IAAI;AAAA,UACpB,YAAY;AAAA,UACZ,0BAA0B;AAAA,UAC1B,gBAAgB;AAAA,QAClB;AAEA,kBAAU,QAAQ,UAAU;AAE5B,cAAM,QAAQ,UAAU,QAAQ;AAChC,eAAO,KAAK,EAAE,aAAa,CAAC;AAAA,MAC9B,CAAC;AAED,SAAG,kCAAkC,MAAM;AACzC,cAAM,aAAa;AAAA,UACjB,MAAM;AAAA,UACN,WAAW,KAAK,IAAI;AAAA,UACpB,uBAAuB;AAAA,UACvB,sBAAsB;AAAA,UACtB,4BAA4B;AAAA,UAC5B,oBAAoB,KAAK,IAAI;AAAA,QAC/B;AAEA,kBAAU,QAAQ,UAAU;AAE5B,cAAM,QAAQ,UAAU,QAAQ;AAChC,eAAO,KAAK,EAAE,aAAa,CAAC;AAAA,MAC9B,CAAC;AAAA,IACH,CAAC;AAAA,EACH,CAAC;AACH,CAAC;","names":[]}
@@ -21,9 +21,12 @@ __export(usage_collector_exports, {
21
21
  UsageCollector: () => UsageCollector
22
22
  });
23
23
  module.exports = __toCommonJS(usage_collector_exports);
24
+ var import_log = require("../log.cjs");
24
25
  class UsageCollector {
25
26
  summary;
27
+ logger = (0, import_log.log)();
26
28
  constructor() {
29
+ this.logger.warn("UsageCollector is deprecated. Use ModelUsageCollector instead.");
27
30
  this.summary = {
28
31
  llmPromptTokens: 0,
29
32
  llmPromptCachedTokens: 0,
@@ -1 +1 @@
1
- {"version":3,"sources":["../../src/metrics/usage_collector.ts"],"sourcesContent":["// SPDX-FileCopyrightText: 2024 LiveKit, Inc.\n//\n// SPDX-License-Identifier: Apache-2.0\nimport type { AgentMetrics } from './base.js';\n\nexport interface UsageSummary {\n llmPromptTokens: number;\n llmPromptCachedTokens: number;\n llmCompletionTokens: number;\n ttsCharactersCount: number;\n sttAudioDurationMs: number;\n}\n\nexport class UsageCollector {\n private summary: UsageSummary;\n\n constructor() {\n this.summary = {\n llmPromptTokens: 0,\n llmPromptCachedTokens: 0,\n llmCompletionTokens: 0,\n ttsCharactersCount: 0,\n sttAudioDurationMs: 0,\n };\n }\n\n collect(metrics: AgentMetrics): void {\n if (metrics.type === 'llm_metrics') {\n this.summary.llmPromptTokens += metrics.promptTokens;\n this.summary.llmPromptCachedTokens += metrics.promptCachedTokens;\n this.summary.llmCompletionTokens += metrics.completionTokens;\n } else if (metrics.type === 'realtime_model_metrics') {\n this.summary.llmPromptTokens += metrics.inputTokens;\n this.summary.llmPromptCachedTokens += metrics.inputTokenDetails.cachedTokens;\n this.summary.llmCompletionTokens += metrics.outputTokens;\n } else if (metrics.type === 'tts_metrics') {\n this.summary.ttsCharactersCount += metrics.charactersCount;\n } else if (metrics.type === 'stt_metrics') {\n this.summary.sttAudioDurationMs += metrics.audioDurationMs;\n }\n }\n\n getSummary(): UsageSummary {\n return { ...this.summary };\n }\n}\n"],"mappings":";;;;;;;;;;;;;;;;;;AAAA;AAAA;AAAA;AAAA;AAAA;AAaO,MAAM,eAAe;AAAA,EAClB;AAAA,EAER,cAAc;AACZ,SAAK,UAAU;AAAA,MACb,iBAAiB;AAAA,MACjB,uBAAuB;AAAA,MACvB,qBAAqB;AAAA,MACrB,oBAAoB;AAAA,MACpB,oBAAoB;AAAA,IACtB;AAAA,EACF;AAAA,EAEA,QAAQ,SAA6B;AACnC,QAAI,QAAQ,SAAS,eAAe;AAClC,WAAK,QAAQ,mBAAmB,QAAQ;AACxC,WAAK,QAAQ,yBAAyB,QAAQ;AAC9C,WAAK,QAAQ,uBAAuB,QAAQ;AAAA,IAC9C,WAAW,QAAQ,SAAS,0BAA0B;AACpD,WAAK,QAAQ,mBAAmB,QAAQ;AACxC,WAAK,QAAQ,yBAAyB,QAAQ,kBAAkB;AAChE,WAAK,QAAQ,uBAAuB,QAAQ;AAAA,IAC9C,WAAW,QAAQ,SAAS,eAAe;AACzC,WAAK,QAAQ,sBAAsB,QAAQ;AAAA,IAC7C,WAAW,QAAQ,SAAS,eAAe;AACzC,WAAK,QAAQ,sBAAsB,QAAQ;AAAA,IAC7C;AAAA,EACF;AAAA,EAEA,aAA2B;AACzB,WAAO,EAAE,GAAG,KAAK,QAAQ;AAAA,EAC3B;AACF;","names":[]}
1
+ {"version":3,"sources":["../../src/metrics/usage_collector.ts"],"sourcesContent":["// SPDX-FileCopyrightText: 2024 LiveKit, Inc.\n//\n// SPDX-License-Identifier: Apache-2.0\nimport { log } from '../log.js';\nimport type { AgentMetrics } from './base.js';\n\n/**\n * @deprecated Use LLMModelUsage, TTSModelUsage, or STTModelUsage instead.\n * These new types provide per-model/provider usage aggregation for more detailed tracking.\n */\nexport interface UsageSummary {\n llmPromptTokens: number;\n llmPromptCachedTokens: number;\n llmCompletionTokens: number;\n ttsCharactersCount: number;\n sttAudioDurationMs: number;\n}\n\n/**\n * @deprecated Use ModelUsageCollector instead.\n * ModelUsageCollector provides per-model/provider usage aggregation for more detailed tracking.\n */\nexport class UsageCollector {\n private summary: UsageSummary;\n private logger = log();\n\n constructor() {\n this.logger.warn('UsageCollector is deprecated. Use ModelUsageCollector instead.');\n this.summary = {\n llmPromptTokens: 0,\n llmPromptCachedTokens: 0,\n llmCompletionTokens: 0,\n ttsCharactersCount: 0,\n sttAudioDurationMs: 0,\n };\n }\n\n collect(metrics: AgentMetrics): void {\n if (metrics.type === 'llm_metrics') {\n this.summary.llmPromptTokens += metrics.promptTokens;\n this.summary.llmPromptCachedTokens += metrics.promptCachedTokens;\n this.summary.llmCompletionTokens += metrics.completionTokens;\n } else if (metrics.type === 'realtime_model_metrics') {\n this.summary.llmPromptTokens += metrics.inputTokens;\n this.summary.llmPromptCachedTokens += metrics.inputTokenDetails.cachedTokens;\n this.summary.llmCompletionTokens += metrics.outputTokens;\n } else if (metrics.type === 'tts_metrics') {\n this.summary.ttsCharactersCount += metrics.charactersCount;\n } else if (metrics.type === 'stt_metrics') {\n this.summary.sttAudioDurationMs += metrics.audioDurationMs;\n }\n }\n\n getSummary(): UsageSummary {\n return { ...this.summary };\n }\n}\n"],"mappings":";;;;;;;;;;;;;;;;;;AAAA;AAAA;AAAA;AAAA;AAAA;AAGA,iBAAoB;AAmBb,MAAM,eAAe;AAAA,EAClB;AAAA,EACA,aAAS,gBAAI;AAAA,EAErB,cAAc;AACZ,SAAK,OAAO,KAAK,gEAAgE;AACjF,SAAK,UAAU;AAAA,MACb,iBAAiB;AAAA,MACjB,uBAAuB;AAAA,MACvB,qBAAqB;AAAA,MACrB,oBAAoB;AAAA,MACpB,oBAAoB;AAAA,IACtB;AAAA,EACF;AAAA,EAEA,QAAQ,SAA6B;AACnC,QAAI,QAAQ,SAAS,eAAe;AAClC,WAAK,QAAQ,mBAAmB,QAAQ;AACxC,WAAK,QAAQ,yBAAyB,QAAQ;AAC9C,WAAK,QAAQ,uBAAuB,QAAQ;AAAA,IAC9C,WAAW,QAAQ,SAAS,0BAA0B;AACpD,WAAK,QAAQ,mBAAmB,QAAQ;AACxC,WAAK,QAAQ,yBAAyB,QAAQ,kBAAkB;AAChE,WAAK,QAAQ,uBAAuB,QAAQ;AAAA,IAC9C,WAAW,QAAQ,SAAS,eAAe;AACzC,WAAK,QAAQ,sBAAsB,QAAQ;AAAA,IAC7C,WAAW,QAAQ,SAAS,eAAe;AACzC,WAAK,QAAQ,sBAAsB,QAAQ;AAAA,IAC7C;AAAA,EACF;AAAA,EAEA,aAA2B;AACzB,WAAO,EAAE,GAAG,KAAK,QAAQ;AAAA,EAC3B;AACF;","names":[]}
@@ -1,4 +1,8 @@
1
1
  import type { AgentMetrics } from './base.js';
2
+ /**
3
+ * @deprecated Use LLMModelUsage, TTSModelUsage, or STTModelUsage instead.
4
+ * These new types provide per-model/provider usage aggregation for more detailed tracking.
5
+ */
2
6
  export interface UsageSummary {
3
7
  llmPromptTokens: number;
4
8
  llmPromptCachedTokens: number;
@@ -6,8 +10,13 @@ export interface UsageSummary {
6
10
  ttsCharactersCount: number;
7
11
  sttAudioDurationMs: number;
8
12
  }
13
+ /**
14
+ * @deprecated Use ModelUsageCollector instead.
15
+ * ModelUsageCollector provides per-model/provider usage aggregation for more detailed tracking.
16
+ */
9
17
  export declare class UsageCollector {
10
18
  private summary;
19
+ private logger;
11
20
  constructor();
12
21
  collect(metrics: AgentMetrics): void;
13
22
  getSummary(): UsageSummary;
@@ -1,4 +1,8 @@
1
1
  import type { AgentMetrics } from './base.js';
2
+ /**
3
+ * @deprecated Use LLMModelUsage, TTSModelUsage, or STTModelUsage instead.
4
+ * These new types provide per-model/provider usage aggregation for more detailed tracking.
5
+ */
2
6
  export interface UsageSummary {
3
7
  llmPromptTokens: number;
4
8
  llmPromptCachedTokens: number;
@@ -6,8 +10,13 @@ export interface UsageSummary {
6
10
  ttsCharactersCount: number;
7
11
  sttAudioDurationMs: number;
8
12
  }
13
+ /**
14
+ * @deprecated Use ModelUsageCollector instead.
15
+ * ModelUsageCollector provides per-model/provider usage aggregation for more detailed tracking.
16
+ */
9
17
  export declare class UsageCollector {
10
18
  private summary;
19
+ private logger;
11
20
  constructor();
12
21
  collect(metrics: AgentMetrics): void;
13
22
  getSummary(): UsageSummary;
@@ -1 +1 @@
1
- {"version":3,"file":"usage_collector.d.ts","sourceRoot":"","sources":["../../src/metrics/usage_collector.ts"],"names":[],"mappings":"AAGA,OAAO,KAAK,EAAE,YAAY,EAAE,MAAM,WAAW,CAAC;AAE9C,MAAM,WAAW,YAAY;IAC3B,eAAe,EAAE,MAAM,CAAC;IACxB,qBAAqB,EAAE,MAAM,CAAC;IAC9B,mBAAmB,EAAE,MAAM,CAAC;IAC5B,kBAAkB,EAAE,MAAM,CAAC;IAC3B,kBAAkB,EAAE,MAAM,CAAC;CAC5B;AAED,qBAAa,cAAc;IACzB,OAAO,CAAC,OAAO,CAAe;;IAY9B,OAAO,CAAC,OAAO,EAAE,YAAY,GAAG,IAAI;IAgBpC,UAAU,IAAI,YAAY;CAG3B"}
1
+ {"version":3,"file":"usage_collector.d.ts","sourceRoot":"","sources":["../../src/metrics/usage_collector.ts"],"names":[],"mappings":"AAIA,OAAO,KAAK,EAAE,YAAY,EAAE,MAAM,WAAW,CAAC;AAE9C;;;GAGG;AACH,MAAM,WAAW,YAAY;IAC3B,eAAe,EAAE,MAAM,CAAC;IACxB,qBAAqB,EAAE,MAAM,CAAC;IAC9B,mBAAmB,EAAE,MAAM,CAAC;IAC5B,kBAAkB,EAAE,MAAM,CAAC;IAC3B,kBAAkB,EAAE,MAAM,CAAC;CAC5B;AAED;;;GAGG;AACH,qBAAa,cAAc;IACzB,OAAO,CAAC,OAAO,CAAe;IAC9B,OAAO,CAAC,MAAM,CAAS;;IAavB,OAAO,CAAC,OAAO,EAAE,YAAY,GAAG,IAAI;IAgBpC,UAAU,IAAI,YAAY;CAG3B"}
@@ -1,6 +1,9 @@
1
+ import { log } from "../log.js";
1
2
  class UsageCollector {
2
3
  summary;
4
+ logger = log();
3
5
  constructor() {
6
+ this.logger.warn("UsageCollector is deprecated. Use ModelUsageCollector instead.");
4
7
  this.summary = {
5
8
  llmPromptTokens: 0,
6
9
  llmPromptCachedTokens: 0,
@@ -1 +1 @@
1
- {"version":3,"sources":["../../src/metrics/usage_collector.ts"],"sourcesContent":["// SPDX-FileCopyrightText: 2024 LiveKit, Inc.\n//\n// SPDX-License-Identifier: Apache-2.0\nimport type { AgentMetrics } from './base.js';\n\nexport interface UsageSummary {\n llmPromptTokens: number;\n llmPromptCachedTokens: number;\n llmCompletionTokens: number;\n ttsCharactersCount: number;\n sttAudioDurationMs: number;\n}\n\nexport class UsageCollector {\n private summary: UsageSummary;\n\n constructor() {\n this.summary = {\n llmPromptTokens: 0,\n llmPromptCachedTokens: 0,\n llmCompletionTokens: 0,\n ttsCharactersCount: 0,\n sttAudioDurationMs: 0,\n };\n }\n\n collect(metrics: AgentMetrics): void {\n if (metrics.type === 'llm_metrics') {\n this.summary.llmPromptTokens += metrics.promptTokens;\n this.summary.llmPromptCachedTokens += metrics.promptCachedTokens;\n this.summary.llmCompletionTokens += metrics.completionTokens;\n } else if (metrics.type === 'realtime_model_metrics') {\n this.summary.llmPromptTokens += metrics.inputTokens;\n this.summary.llmPromptCachedTokens += metrics.inputTokenDetails.cachedTokens;\n this.summary.llmCompletionTokens += metrics.outputTokens;\n } else if (metrics.type === 'tts_metrics') {\n this.summary.ttsCharactersCount += metrics.charactersCount;\n } else if (metrics.type === 'stt_metrics') {\n this.summary.sttAudioDurationMs += metrics.audioDurationMs;\n }\n }\n\n getSummary(): UsageSummary {\n return { ...this.summary };\n }\n}\n"],"mappings":"AAaO,MAAM,eAAe;AAAA,EAClB;AAAA,EAER,cAAc;AACZ,SAAK,UAAU;AAAA,MACb,iBAAiB;AAAA,MACjB,uBAAuB;AAAA,MACvB,qBAAqB;AAAA,MACrB,oBAAoB;AAAA,MACpB,oBAAoB;AAAA,IACtB;AAAA,EACF;AAAA,EAEA,QAAQ,SAA6B;AACnC,QAAI,QAAQ,SAAS,eAAe;AAClC,WAAK,QAAQ,mBAAmB,QAAQ;AACxC,WAAK,QAAQ,yBAAyB,QAAQ;AAC9C,WAAK,QAAQ,uBAAuB,QAAQ;AAAA,IAC9C,WAAW,QAAQ,SAAS,0BAA0B;AACpD,WAAK,QAAQ,mBAAmB,QAAQ;AACxC,WAAK,QAAQ,yBAAyB,QAAQ,kBAAkB;AAChE,WAAK,QAAQ,uBAAuB,QAAQ;AAAA,IAC9C,WAAW,QAAQ,SAAS,eAAe;AACzC,WAAK,QAAQ,sBAAsB,QAAQ;AAAA,IAC7C,WAAW,QAAQ,SAAS,eAAe;AACzC,WAAK,QAAQ,sBAAsB,QAAQ;AAAA,IAC7C;AAAA,EACF;AAAA,EAEA,aAA2B;AACzB,WAAO,EAAE,GAAG,KAAK,QAAQ;AAAA,EAC3B;AACF;","names":[]}
1
+ {"version":3,"sources":["../../src/metrics/usage_collector.ts"],"sourcesContent":["// SPDX-FileCopyrightText: 2024 LiveKit, Inc.\n//\n// SPDX-License-Identifier: Apache-2.0\nimport { log } from '../log.js';\nimport type { AgentMetrics } from './base.js';\n\n/**\n * @deprecated Use LLMModelUsage, TTSModelUsage, or STTModelUsage instead.\n * These new types provide per-model/provider usage aggregation for more detailed tracking.\n */\nexport interface UsageSummary {\n llmPromptTokens: number;\n llmPromptCachedTokens: number;\n llmCompletionTokens: number;\n ttsCharactersCount: number;\n sttAudioDurationMs: number;\n}\n\n/**\n * @deprecated Use ModelUsageCollector instead.\n * ModelUsageCollector provides per-model/provider usage aggregation for more detailed tracking.\n */\nexport class UsageCollector {\n private summary: UsageSummary;\n private logger = log();\n\n constructor() {\n this.logger.warn('UsageCollector is deprecated. Use ModelUsageCollector instead.');\n this.summary = {\n llmPromptTokens: 0,\n llmPromptCachedTokens: 0,\n llmCompletionTokens: 0,\n ttsCharactersCount: 0,\n sttAudioDurationMs: 0,\n };\n }\n\n collect(metrics: AgentMetrics): void {\n if (metrics.type === 'llm_metrics') {\n this.summary.llmPromptTokens += metrics.promptTokens;\n this.summary.llmPromptCachedTokens += metrics.promptCachedTokens;\n this.summary.llmCompletionTokens += metrics.completionTokens;\n } else if (metrics.type === 'realtime_model_metrics') {\n this.summary.llmPromptTokens += metrics.inputTokens;\n this.summary.llmPromptCachedTokens += metrics.inputTokenDetails.cachedTokens;\n this.summary.llmCompletionTokens += metrics.outputTokens;\n } else if (metrics.type === 'tts_metrics') {\n this.summary.ttsCharactersCount += metrics.charactersCount;\n } else if (metrics.type === 'stt_metrics') {\n this.summary.sttAudioDurationMs += metrics.audioDurationMs;\n }\n }\n\n getSummary(): UsageSummary {\n return { ...this.summary };\n }\n}\n"],"mappings":"AAGA,SAAS,WAAW;AAmBb,MAAM,eAAe;AAAA,EAClB;AAAA,EACA,SAAS,IAAI;AAAA,EAErB,cAAc;AACZ,SAAK,OAAO,KAAK,gEAAgE;AACjF,SAAK,UAAU;AAAA,MACb,iBAAiB;AAAA,MACjB,uBAAuB;AAAA,MACvB,qBAAqB;AAAA,MACrB,oBAAoB;AAAA,MACpB,oBAAoB;AAAA,IACtB;AAAA,EACF;AAAA,EAEA,QAAQ,SAA6B;AACnC,QAAI,QAAQ,SAAS,eAAe;AAClC,WAAK,QAAQ,mBAAmB,QAAQ;AACxC,WAAK,QAAQ,yBAAyB,QAAQ;AAC9C,WAAK,QAAQ,uBAAuB,QAAQ;AAAA,IAC9C,WAAW,QAAQ,SAAS,0BAA0B;AACpD,WAAK,QAAQ,mBAAmB,QAAQ;AACxC,WAAK,QAAQ,yBAAyB,QAAQ,kBAAkB;AAChE,WAAK,QAAQ,uBAAuB,QAAQ;AAAA,IAC9C,WAAW,QAAQ,SAAS,eAAe;AACzC,WAAK,QAAQ,sBAAsB,QAAQ;AAAA,IAC7C,WAAW,QAAQ,SAAS,eAAe;AACzC,WAAK,QAAQ,sBAAsB,QAAQ;AAAA,IAC7C;AAAA,EACF;AAAA,EAEA,aAA2B;AACzB,WAAO,EAAE,GAAG,KAAK,QAAQ;AAAA,EAC3B;AACF;","names":[]}
@@ -65,6 +65,15 @@ const logMetrics = (metrics) => {
65
65
  logger.child({
66
66
  audioDurationMs: Math.round(metrics.audioDurationMs)
67
67
  }).info("STT metrics");
68
+ } else if (metrics.type === "interruption_metrics") {
69
+ logger.child({
70
+ totalDurationMs: roundTwoDecimals(metrics.totalDuration),
71
+ predictionDurationMs: roundTwoDecimals(metrics.predictionDuration),
72
+ detectionDelayMs: roundTwoDecimals(metrics.detectionDelay),
73
+ numInterruptions: metrics.numInterruptions,
74
+ numBackchannels: metrics.numBackchannels,
75
+ numRequests: metrics.numRequests
76
+ }).info("Interruption metrics");
68
77
  }
69
78
  };
70
79
  // Annotate the CommonJS export names for ESM import in node: