@livekit/agents 1.0.47 → 1.1.0-dev.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/beta/index.cjs +29 -0
- package/dist/beta/index.cjs.map +1 -0
- package/dist/beta/index.d.cts +2 -0
- package/dist/beta/index.d.ts +2 -0
- package/dist/beta/index.d.ts.map +1 -0
- package/dist/beta/index.js +7 -0
- package/dist/beta/index.js.map +1 -0
- package/dist/beta/workflows/index.cjs +29 -0
- package/dist/beta/workflows/index.cjs.map +1 -0
- package/dist/beta/workflows/index.d.cts +2 -0
- package/dist/beta/workflows/index.d.ts +2 -0
- package/dist/beta/workflows/index.d.ts.map +1 -0
- package/dist/beta/workflows/index.js +7 -0
- package/dist/beta/workflows/index.js.map +1 -0
- package/dist/beta/workflows/task_group.cjs +162 -0
- package/dist/beta/workflows/task_group.cjs.map +1 -0
- package/dist/beta/workflows/task_group.d.cts +32 -0
- package/dist/beta/workflows/task_group.d.ts +32 -0
- package/dist/beta/workflows/task_group.d.ts.map +1 -0
- package/dist/beta/workflows/task_group.js +138 -0
- package/dist/beta/workflows/task_group.js.map +1 -0
- package/dist/constants.cjs +27 -0
- package/dist/constants.cjs.map +1 -1
- package/dist/constants.d.cts +9 -0
- package/dist/constants.d.ts +9 -0
- package/dist/constants.d.ts.map +1 -1
- package/dist/constants.js +18 -0
- package/dist/constants.js.map +1 -1
- package/dist/index.cjs +3 -0
- package/dist/index.cjs.map +1 -1
- package/dist/index.d.cts +2 -1
- package/dist/index.d.ts +2 -1
- package/dist/index.d.ts.map +1 -1
- package/dist/index.js +2 -0
- package/dist/index.js.map +1 -1
- package/dist/inference/api_protos.d.cts +12 -12
- package/dist/inference/api_protos.d.ts +12 -12
- package/dist/inference/interruption/defaults.cjs +81 -0
- package/dist/inference/interruption/defaults.cjs.map +1 -0
- package/dist/inference/interruption/defaults.d.cts +19 -0
- package/dist/inference/interruption/defaults.d.ts +19 -0
- package/dist/inference/interruption/defaults.d.ts.map +1 -0
- package/dist/inference/interruption/defaults.js +46 -0
- package/dist/inference/interruption/defaults.js.map +1 -0
- package/dist/inference/interruption/errors.cjs +44 -0
- package/dist/inference/interruption/errors.cjs.map +1 -0
- package/dist/inference/interruption/errors.d.cts +12 -0
- package/dist/inference/interruption/errors.d.ts +12 -0
- package/dist/inference/interruption/errors.d.ts.map +1 -0
- package/dist/inference/interruption/errors.js +20 -0
- package/dist/inference/interruption/errors.js.map +1 -0
- package/dist/inference/interruption/http_transport.cjs +147 -0
- package/dist/inference/interruption/http_transport.cjs.map +1 -0
- package/dist/inference/interruption/http_transport.d.cts +63 -0
- package/dist/inference/interruption/http_transport.d.ts +63 -0
- package/dist/inference/interruption/http_transport.d.ts.map +1 -0
- package/dist/inference/interruption/http_transport.js +121 -0
- package/dist/inference/interruption/http_transport.js.map +1 -0
- package/dist/inference/interruption/interruption_cache_entry.cjs +58 -0
- package/dist/inference/interruption/interruption_cache_entry.cjs.map +1 -0
- package/dist/inference/interruption/interruption_cache_entry.d.cts +30 -0
- package/dist/inference/interruption/interruption_cache_entry.d.ts +30 -0
- package/dist/inference/interruption/interruption_cache_entry.d.ts.map +1 -0
- package/dist/inference/interruption/interruption_cache_entry.js +34 -0
- package/dist/inference/interruption/interruption_cache_entry.js.map +1 -0
- package/dist/inference/interruption/interruption_detector.cjs +181 -0
- package/dist/inference/interruption/interruption_detector.cjs.map +1 -0
- package/dist/inference/interruption/interruption_detector.d.cts +59 -0
- package/dist/inference/interruption/interruption_detector.d.ts +59 -0
- package/dist/inference/interruption/interruption_detector.d.ts.map +1 -0
- package/dist/inference/interruption/interruption_detector.js +147 -0
- package/dist/inference/interruption/interruption_detector.js.map +1 -0
- package/dist/inference/interruption/interruption_stream.cjs +368 -0
- package/dist/inference/interruption/interruption_stream.cjs.map +1 -0
- package/dist/inference/interruption/interruption_stream.d.cts +46 -0
- package/dist/inference/interruption/interruption_stream.d.ts +46 -0
- package/dist/inference/interruption/interruption_stream.d.ts.map +1 -0
- package/dist/inference/interruption/interruption_stream.js +344 -0
- package/dist/inference/interruption/interruption_stream.js.map +1 -0
- package/dist/inference/interruption/types.cjs +17 -0
- package/dist/inference/interruption/types.cjs.map +1 -0
- package/dist/inference/interruption/types.d.cts +66 -0
- package/dist/inference/interruption/types.d.ts +66 -0
- package/dist/inference/interruption/types.d.ts.map +1 -0
- package/dist/inference/interruption/types.js +1 -0
- package/dist/inference/interruption/types.js.map +1 -0
- package/dist/inference/interruption/utils.cjs +130 -0
- package/dist/inference/interruption/utils.cjs.map +1 -0
- package/dist/inference/interruption/utils.d.cts +41 -0
- package/dist/inference/interruption/utils.d.ts +41 -0
- package/dist/inference/interruption/utils.d.ts.map +1 -0
- package/dist/inference/interruption/utils.js +105 -0
- package/dist/inference/interruption/utils.js.map +1 -0
- package/dist/inference/interruption/utils.test.cjs +105 -0
- package/dist/inference/interruption/utils.test.cjs.map +1 -0
- package/dist/inference/interruption/utils.test.js +104 -0
- package/dist/inference/interruption/utils.test.js.map +1 -0
- package/dist/inference/interruption/ws_transport.cjs +329 -0
- package/dist/inference/interruption/ws_transport.cjs.map +1 -0
- package/dist/inference/interruption/ws_transport.d.cts +33 -0
- package/dist/inference/interruption/ws_transport.d.ts +33 -0
- package/dist/inference/interruption/ws_transport.d.ts.map +1 -0
- package/dist/inference/interruption/ws_transport.js +295 -0
- package/dist/inference/interruption/ws_transport.js.map +1 -0
- package/dist/inference/llm.cjs +14 -10
- package/dist/inference/llm.cjs.map +1 -1
- package/dist/inference/llm.d.cts +2 -1
- package/dist/inference/llm.d.ts +2 -1
- package/dist/inference/llm.d.ts.map +1 -1
- package/dist/inference/llm.js +8 -10
- package/dist/inference/llm.js.map +1 -1
- package/dist/inference/stt.cjs +7 -2
- package/dist/inference/stt.cjs.map +1 -1
- package/dist/inference/stt.d.cts +2 -0
- package/dist/inference/stt.d.ts +2 -0
- package/dist/inference/stt.d.ts.map +1 -1
- package/dist/inference/stt.js +8 -3
- package/dist/inference/stt.js.map +1 -1
- package/dist/inference/tts.cjs +7 -2
- package/dist/inference/tts.cjs.map +1 -1
- package/dist/inference/tts.d.cts +2 -0
- package/dist/inference/tts.d.ts +2 -0
- package/dist/inference/tts.d.ts.map +1 -1
- package/dist/inference/tts.js +8 -3
- package/dist/inference/tts.js.map +1 -1
- package/dist/inference/utils.cjs +26 -7
- package/dist/inference/utils.cjs.map +1 -1
- package/dist/inference/utils.d.cts +13 -0
- package/dist/inference/utils.d.ts +13 -0
- package/dist/inference/utils.d.ts.map +1 -1
- package/dist/inference/utils.js +18 -2
- package/dist/inference/utils.js.map +1 -1
- package/dist/llm/chat_context.cjs +108 -2
- package/dist/llm/chat_context.cjs.map +1 -1
- package/dist/llm/chat_context.d.cts +28 -1
- package/dist/llm/chat_context.d.ts +28 -1
- package/dist/llm/chat_context.d.ts.map +1 -1
- package/dist/llm/chat_context.js +108 -2
- package/dist/llm/chat_context.js.map +1 -1
- package/dist/llm/chat_context.test.cjs +43 -0
- package/dist/llm/chat_context.test.cjs.map +1 -1
- package/dist/llm/chat_context.test.js +43 -0
- package/dist/llm/chat_context.test.js.map +1 -1
- package/dist/llm/index.cjs +2 -0
- package/dist/llm/index.cjs.map +1 -1
- package/dist/llm/index.d.cts +2 -2
- package/dist/llm/index.d.ts +2 -2
- package/dist/llm/index.d.ts.map +1 -1
- package/dist/llm/index.js +3 -1
- package/dist/llm/index.js.map +1 -1
- package/dist/llm/llm.cjs +16 -1
- package/dist/llm/llm.cjs.map +1 -1
- package/dist/llm/llm.d.cts +9 -0
- package/dist/llm/llm.d.ts +9 -0
- package/dist/llm/llm.d.ts.map +1 -1
- package/dist/llm/llm.js +16 -1
- package/dist/llm/llm.js.map +1 -1
- package/dist/llm/provider_format/index.d.cts +1 -1
- package/dist/llm/provider_format/index.d.ts +1 -1
- package/dist/llm/realtime.cjs +3 -0
- package/dist/llm/realtime.cjs.map +1 -1
- package/dist/llm/realtime.d.cts +1 -0
- package/dist/llm/realtime.d.ts +1 -0
- package/dist/llm/realtime.d.ts.map +1 -1
- package/dist/llm/realtime.js +3 -0
- package/dist/llm/realtime.js.map +1 -1
- package/dist/llm/tool_context.cjs +7 -0
- package/dist/llm/tool_context.cjs.map +1 -1
- package/dist/llm/tool_context.d.cts +10 -2
- package/dist/llm/tool_context.d.ts +10 -2
- package/dist/llm/tool_context.d.ts.map +1 -1
- package/dist/llm/tool_context.js +6 -0
- package/dist/llm/tool_context.js.map +1 -1
- package/dist/metrics/base.cjs.map +1 -1
- package/dist/metrics/base.d.cts +45 -1
- package/dist/metrics/base.d.ts +45 -1
- package/dist/metrics/base.d.ts.map +1 -1
- package/dist/metrics/index.cjs +5 -0
- package/dist/metrics/index.cjs.map +1 -1
- package/dist/metrics/index.d.cts +2 -1
- package/dist/metrics/index.d.ts +2 -1
- package/dist/metrics/index.d.ts.map +1 -1
- package/dist/metrics/index.js +6 -0
- package/dist/metrics/index.js.map +1 -1
- package/dist/metrics/model_usage.cjs +189 -0
- package/dist/metrics/model_usage.cjs.map +1 -0
- package/dist/metrics/model_usage.d.cts +92 -0
- package/dist/metrics/model_usage.d.ts +92 -0
- package/dist/metrics/model_usage.d.ts.map +1 -0
- package/dist/metrics/model_usage.js +164 -0
- package/dist/metrics/model_usage.js.map +1 -0
- package/dist/metrics/model_usage.test.cjs +474 -0
- package/dist/metrics/model_usage.test.cjs.map +1 -0
- package/dist/metrics/model_usage.test.js +476 -0
- package/dist/metrics/model_usage.test.js.map +1 -0
- package/dist/metrics/usage_collector.cjs +3 -0
- package/dist/metrics/usage_collector.cjs.map +1 -1
- package/dist/metrics/usage_collector.d.cts +9 -0
- package/dist/metrics/usage_collector.d.ts +9 -0
- package/dist/metrics/usage_collector.d.ts.map +1 -1
- package/dist/metrics/usage_collector.js +3 -0
- package/dist/metrics/usage_collector.js.map +1 -1
- package/dist/metrics/utils.cjs +9 -0
- package/dist/metrics/utils.cjs.map +1 -1
- package/dist/metrics/utils.d.ts.map +1 -1
- package/dist/metrics/utils.js +9 -0
- package/dist/metrics/utils.js.map +1 -1
- package/dist/stream/multi_input_stream.test.cjs +4 -0
- package/dist/stream/multi_input_stream.test.cjs.map +1 -1
- package/dist/stream/multi_input_stream.test.js +5 -1
- package/dist/stream/multi_input_stream.test.js.map +1 -1
- package/dist/stream/stream_channel.cjs +31 -0
- package/dist/stream/stream_channel.cjs.map +1 -1
- package/dist/stream/stream_channel.d.cts +4 -2
- package/dist/stream/stream_channel.d.ts +4 -2
- package/dist/stream/stream_channel.d.ts.map +1 -1
- package/dist/stream/stream_channel.js +31 -0
- package/dist/stream/stream_channel.js.map +1 -1
- package/dist/stt/stt.cjs +34 -2
- package/dist/stt/stt.cjs.map +1 -1
- package/dist/stt/stt.d.cts +22 -0
- package/dist/stt/stt.d.ts +22 -0
- package/dist/stt/stt.d.ts.map +1 -1
- package/dist/stt/stt.js +34 -2
- package/dist/stt/stt.js.map +1 -1
- package/dist/telemetry/otel_http_exporter.cjs +24 -5
- package/dist/telemetry/otel_http_exporter.cjs.map +1 -1
- package/dist/telemetry/otel_http_exporter.d.cts +1 -0
- package/dist/telemetry/otel_http_exporter.d.ts +1 -0
- package/dist/telemetry/otel_http_exporter.d.ts.map +1 -1
- package/dist/telemetry/otel_http_exporter.js +24 -5
- package/dist/telemetry/otel_http_exporter.js.map +1 -1
- package/dist/telemetry/trace_types.cjs +5 -5
- package/dist/telemetry/trace_types.cjs.map +1 -1
- package/dist/telemetry/trace_types.d.cts +9 -5
- package/dist/telemetry/trace_types.d.ts +9 -5
- package/dist/telemetry/trace_types.d.ts.map +1 -1
- package/dist/telemetry/trace_types.js +5 -5
- package/dist/telemetry/trace_types.js.map +1 -1
- package/dist/telemetry/traces.cjs +47 -8
- package/dist/telemetry/traces.cjs.map +1 -1
- package/dist/telemetry/traces.d.ts.map +1 -1
- package/dist/telemetry/traces.js +47 -8
- package/dist/telemetry/traces.js.map +1 -1
- package/dist/tts/tts.cjs +64 -2
- package/dist/tts/tts.cjs.map +1 -1
- package/dist/tts/tts.d.cts +34 -0
- package/dist/tts/tts.d.ts +34 -0
- package/dist/tts/tts.d.ts.map +1 -1
- package/dist/tts/tts.js +64 -2
- package/dist/tts/tts.js.map +1 -1
- package/dist/utils.cjs +1 -0
- package/dist/utils.cjs.map +1 -1
- package/dist/utils.d.ts.map +1 -1
- package/dist/utils.js +1 -0
- package/dist/utils.js.map +1 -1
- package/dist/version.cjs +1 -1
- package/dist/version.js +1 -1
- package/dist/voice/agent.cjs +34 -4
- package/dist/voice/agent.cjs.map +1 -1
- package/dist/voice/agent.d.cts +11 -2
- package/dist/voice/agent.d.ts +11 -2
- package/dist/voice/agent.d.ts.map +1 -1
- package/dist/voice/agent.js +34 -4
- package/dist/voice/agent.js.map +1 -1
- package/dist/voice/agent_activity.cjs +292 -44
- package/dist/voice/agent_activity.cjs.map +1 -1
- package/dist/voice/agent_activity.d.cts +27 -6
- package/dist/voice/agent_activity.d.ts +27 -6
- package/dist/voice/agent_activity.d.ts.map +1 -1
- package/dist/voice/agent_activity.js +293 -45
- package/dist/voice/agent_activity.js.map +1 -1
- package/dist/voice/agent_session.cjs +105 -48
- package/dist/voice/agent_session.cjs.map +1 -1
- package/dist/voice/agent_session.d.cts +90 -20
- package/dist/voice/agent_session.d.ts +90 -20
- package/dist/voice/agent_session.d.ts.map +1 -1
- package/dist/voice/agent_session.js +105 -46
- package/dist/voice/agent_session.js.map +1 -1
- package/dist/voice/audio_recognition.cjs +287 -6
- package/dist/voice/audio_recognition.cjs.map +1 -1
- package/dist/voice/audio_recognition.d.cts +42 -3
- package/dist/voice/audio_recognition.d.ts +42 -3
- package/dist/voice/audio_recognition.d.ts.map +1 -1
- package/dist/voice/audio_recognition.js +289 -7
- package/dist/voice/audio_recognition.js.map +1 -1
- package/dist/voice/client_events.cjs +554 -0
- package/dist/voice/client_events.cjs.map +1 -0
- package/dist/voice/client_events.d.cts +195 -0
- package/dist/voice/client_events.d.ts +195 -0
- package/dist/voice/client_events.d.ts.map +1 -0
- package/dist/voice/client_events.js +548 -0
- package/dist/voice/client_events.js.map +1 -0
- package/dist/voice/events.cjs +1 -0
- package/dist/voice/events.cjs.map +1 -1
- package/dist/voice/events.d.cts +8 -5
- package/dist/voice/events.d.ts +8 -5
- package/dist/voice/events.d.ts.map +1 -1
- package/dist/voice/events.js +1 -0
- package/dist/voice/events.js.map +1 -1
- package/dist/voice/generation.cjs +43 -8
- package/dist/voice/generation.cjs.map +1 -1
- package/dist/voice/generation.d.cts +3 -3
- package/dist/voice/generation.d.ts +3 -3
- package/dist/voice/generation.d.ts.map +1 -1
- package/dist/voice/generation.js +43 -8
- package/dist/voice/generation.js.map +1 -1
- package/dist/voice/index.cjs +1 -0
- package/dist/voice/index.cjs.map +1 -1
- package/dist/voice/index.d.cts +1 -0
- package/dist/voice/index.d.ts +1 -0
- package/dist/voice/index.d.ts.map +1 -1
- package/dist/voice/index.js +1 -0
- package/dist/voice/index.js.map +1 -1
- package/dist/voice/report.cjs +20 -8
- package/dist/voice/report.cjs.map +1 -1
- package/dist/voice/report.d.cts +5 -0
- package/dist/voice/report.d.ts +5 -0
- package/dist/voice/report.d.ts.map +1 -1
- package/dist/voice/report.js +20 -8
- package/dist/voice/report.js.map +1 -1
- package/dist/voice/report.test.cjs +106 -0
- package/dist/voice/report.test.cjs.map +1 -0
- package/dist/voice/report.test.js +105 -0
- package/dist/voice/report.test.js.map +1 -0
- package/dist/voice/room_io/room_io.cjs +16 -41
- package/dist/voice/room_io/room_io.cjs.map +1 -1
- package/dist/voice/room_io/room_io.d.cts +4 -9
- package/dist/voice/room_io/room_io.d.ts +4 -9
- package/dist/voice/room_io/room_io.d.ts.map +1 -1
- package/dist/voice/room_io/room_io.js +17 -43
- package/dist/voice/room_io/room_io.js.map +1 -1
- package/dist/voice/testing/fake_llm.cjs +127 -0
- package/dist/voice/testing/fake_llm.cjs.map +1 -0
- package/dist/voice/testing/fake_llm.d.cts +30 -0
- package/dist/voice/testing/fake_llm.d.ts +30 -0
- package/dist/voice/testing/fake_llm.d.ts.map +1 -0
- package/dist/voice/testing/fake_llm.js +103 -0
- package/dist/voice/testing/fake_llm.js.map +1 -0
- package/dist/voice/testing/index.cjs +3 -0
- package/dist/voice/testing/index.cjs.map +1 -1
- package/dist/voice/testing/index.d.cts +1 -0
- package/dist/voice/testing/index.d.ts +1 -0
- package/dist/voice/testing/index.d.ts.map +1 -1
- package/dist/voice/testing/index.js +2 -0
- package/dist/voice/testing/index.js.map +1 -1
- package/dist/voice/turn_config/endpointing.cjs +33 -0
- package/dist/voice/turn_config/endpointing.cjs.map +1 -0
- package/dist/voice/turn_config/endpointing.d.cts +30 -0
- package/dist/voice/turn_config/endpointing.d.ts +30 -0
- package/dist/voice/turn_config/endpointing.d.ts.map +1 -0
- package/dist/voice/turn_config/endpointing.js +9 -0
- package/dist/voice/turn_config/endpointing.js.map +1 -0
- package/dist/voice/turn_config/interruption.cjs +37 -0
- package/dist/voice/turn_config/interruption.cjs.map +1 -0
- package/dist/voice/turn_config/interruption.d.cts +53 -0
- package/dist/voice/turn_config/interruption.d.ts +53 -0
- package/dist/voice/turn_config/interruption.d.ts.map +1 -0
- package/dist/voice/turn_config/interruption.js +13 -0
- package/dist/voice/turn_config/interruption.js.map +1 -0
- package/dist/voice/turn_config/turn_handling.cjs +35 -0
- package/dist/voice/turn_config/turn_handling.cjs.map +1 -0
- package/dist/voice/turn_config/turn_handling.d.cts +36 -0
- package/dist/voice/turn_config/turn_handling.d.ts +36 -0
- package/dist/voice/turn_config/turn_handling.d.ts.map +1 -0
- package/dist/voice/turn_config/turn_handling.js +11 -0
- package/dist/voice/turn_config/turn_handling.js.map +1 -0
- package/dist/voice/turn_config/utils.cjs +97 -0
- package/dist/voice/turn_config/utils.cjs.map +1 -0
- package/dist/voice/turn_config/utils.d.cts +25 -0
- package/dist/voice/turn_config/utils.d.ts +25 -0
- package/dist/voice/turn_config/utils.d.ts.map +1 -0
- package/dist/voice/turn_config/utils.js +73 -0
- package/dist/voice/turn_config/utils.js.map +1 -0
- package/dist/voice/turn_config/utils.test.cjs +86 -0
- package/dist/voice/turn_config/utils.test.cjs.map +1 -0
- package/dist/voice/turn_config/utils.test.js +85 -0
- package/dist/voice/turn_config/utils.test.js.map +1 -0
- package/dist/voice/wire_format.cjs +798 -0
- package/dist/voice/wire_format.cjs.map +1 -0
- package/dist/voice/wire_format.d.cts +5503 -0
- package/dist/voice/wire_format.d.ts +5503 -0
- package/dist/voice/wire_format.d.ts.map +1 -0
- package/dist/voice/wire_format.js +728 -0
- package/dist/voice/wire_format.js.map +1 -0
- package/package.json +2 -1
- package/src/beta/index.ts +9 -0
- package/src/beta/workflows/index.ts +9 -0
- package/src/beta/workflows/task_group.ts +194 -0
- package/src/constants.ts +13 -0
- package/src/index.ts +2 -1
- package/src/inference/interruption/defaults.ts +51 -0
- package/src/inference/interruption/errors.ts +25 -0
- package/src/inference/interruption/http_transport.ts +187 -0
- package/src/inference/interruption/interruption_cache_entry.ts +50 -0
- package/src/inference/interruption/interruption_detector.ts +188 -0
- package/src/inference/interruption/interruption_stream.ts +467 -0
- package/src/inference/interruption/types.ts +84 -0
- package/src/inference/interruption/utils.test.ts +132 -0
- package/src/inference/interruption/utils.ts +137 -0
- package/src/inference/interruption/ws_transport.ts +402 -0
- package/src/inference/llm.ts +9 -12
- package/src/inference/stt.ts +10 -3
- package/src/inference/tts.ts +10 -3
- package/src/inference/utils.ts +29 -1
- package/src/llm/chat_context.test.ts +48 -0
- package/src/llm/chat_context.ts +161 -0
- package/src/llm/index.ts +2 -0
- package/src/llm/llm.ts +16 -0
- package/src/llm/realtime.ts +4 -0
- package/src/llm/tool_context.ts +14 -0
- package/src/metrics/base.ts +48 -1
- package/src/metrics/index.ts +11 -0
- package/src/metrics/model_usage.test.ts +545 -0
- package/src/metrics/model_usage.ts +262 -0
- package/src/metrics/usage_collector.ts +11 -0
- package/src/metrics/utils.ts +11 -0
- package/src/stream/multi_input_stream.test.ts +6 -1
- package/src/stream/stream_channel.ts +34 -2
- package/src/stt/stt.ts +38 -0
- package/src/telemetry/otel_http_exporter.ts +28 -5
- package/src/telemetry/trace_types.ts +11 -8
- package/src/telemetry/traces.ts +111 -54
- package/src/tts/tts.ts +69 -1
- package/src/utils.ts +5 -0
- package/src/voice/agent.ts +41 -3
- package/src/voice/agent_activity.ts +371 -34
- package/src/voice/agent_session.ts +207 -59
- package/src/voice/audio_recognition.ts +385 -9
- package/src/voice/client_events.ts +838 -0
- package/src/voice/events.ts +14 -4
- package/src/voice/generation.ts +52 -9
- package/src/voice/index.ts +1 -0
- package/src/voice/report.test.ts +117 -0
- package/src/voice/report.ts +29 -6
- package/src/voice/room_io/room_io.ts +21 -64
- package/src/voice/testing/fake_llm.ts +138 -0
- package/src/voice/testing/index.ts +2 -0
- package/src/voice/turn_config/endpointing.ts +33 -0
- package/src/voice/turn_config/interruption.ts +56 -0
- package/src/voice/turn_config/turn_handling.ts +45 -0
- package/src/voice/turn_config/utils.test.ts +100 -0
- package/src/voice/turn_config/utils.ts +103 -0
- package/src/voice/wire_format.ts +827 -0
|
@@ -0,0 +1,476 @@
|
|
|
1
|
+
import { beforeEach, describe, expect, it } from "vitest";
|
|
2
|
+
import {
|
|
3
|
+
ModelUsageCollector,
|
|
4
|
+
filterZeroValues
|
|
5
|
+
} from "./model_usage.js";
|
|
6
|
+
describe("model_usage", () => {
|
|
7
|
+
describe("filterZeroValues", () => {
|
|
8
|
+
it("should filter out zero values from LLMModelUsage", () => {
|
|
9
|
+
const usage = {
|
|
10
|
+
type: "llm_usage",
|
|
11
|
+
provider: "openai",
|
|
12
|
+
model: "gpt-4o",
|
|
13
|
+
inputTokens: 100,
|
|
14
|
+
inputCachedTokens: 0,
|
|
15
|
+
inputAudioTokens: 0,
|
|
16
|
+
inputCachedAudioTokens: 0,
|
|
17
|
+
inputTextTokens: 0,
|
|
18
|
+
inputCachedTextTokens: 0,
|
|
19
|
+
inputImageTokens: 0,
|
|
20
|
+
inputCachedImageTokens: 0,
|
|
21
|
+
outputTokens: 50,
|
|
22
|
+
outputAudioTokens: 0,
|
|
23
|
+
outputTextTokens: 0,
|
|
24
|
+
sessionDurationMs: 0
|
|
25
|
+
};
|
|
26
|
+
const filtered = filterZeroValues(usage);
|
|
27
|
+
expect(filtered.type).toBe("llm_usage");
|
|
28
|
+
expect(filtered.provider).toBe("openai");
|
|
29
|
+
expect(filtered.model).toBe("gpt-4o");
|
|
30
|
+
expect(filtered.inputTokens).toBe(100);
|
|
31
|
+
expect(filtered.outputTokens).toBe(50);
|
|
32
|
+
expect(filtered.inputCachedTokens).toBeUndefined();
|
|
33
|
+
expect(filtered.inputAudioTokens).toBeUndefined();
|
|
34
|
+
expect(filtered.sessionDurationMs).toBeUndefined();
|
|
35
|
+
});
|
|
36
|
+
it("should filter out zero values from TTSModelUsage", () => {
|
|
37
|
+
const usage = {
|
|
38
|
+
type: "tts_usage",
|
|
39
|
+
provider: "elevenlabs",
|
|
40
|
+
model: "eleven_turbo_v2",
|
|
41
|
+
inputTokens: 0,
|
|
42
|
+
outputTokens: 0,
|
|
43
|
+
charactersCount: 500,
|
|
44
|
+
audioDurationMs: 3e3
|
|
45
|
+
};
|
|
46
|
+
const filtered = filterZeroValues(usage);
|
|
47
|
+
expect(filtered.type).toBe("tts_usage");
|
|
48
|
+
expect(filtered.provider).toBe("elevenlabs");
|
|
49
|
+
expect(filtered.charactersCount).toBe(500);
|
|
50
|
+
expect(filtered.audioDurationMs).toBe(3e3);
|
|
51
|
+
expect(filtered.inputTokens).toBeUndefined();
|
|
52
|
+
expect(filtered.outputTokens).toBeUndefined();
|
|
53
|
+
});
|
|
54
|
+
it("should keep all values when none are zero", () => {
|
|
55
|
+
const usage = {
|
|
56
|
+
type: "stt_usage",
|
|
57
|
+
provider: "deepgram",
|
|
58
|
+
model: "nova-2",
|
|
59
|
+
inputTokens: 10,
|
|
60
|
+
outputTokens: 20,
|
|
61
|
+
audioDurationMs: 5e3
|
|
62
|
+
};
|
|
63
|
+
const filtered = filterZeroValues(usage);
|
|
64
|
+
expect(Object.keys(filtered)).toHaveLength(6);
|
|
65
|
+
expect(filtered).toEqual(usage);
|
|
66
|
+
});
|
|
67
|
+
});
|
|
68
|
+
describe("ModelUsageCollector", () => {
|
|
69
|
+
let collector;
|
|
70
|
+
beforeEach(() => {
|
|
71
|
+
collector = new ModelUsageCollector();
|
|
72
|
+
});
|
|
73
|
+
describe("collect LLM metrics", () => {
|
|
74
|
+
it("should aggregate LLM metrics by provider and model", () => {
|
|
75
|
+
const metrics1 = {
|
|
76
|
+
type: "llm_metrics",
|
|
77
|
+
label: "test",
|
|
78
|
+
requestId: "req1",
|
|
79
|
+
timestamp: Date.now(),
|
|
80
|
+
durationMs: 100,
|
|
81
|
+
ttftMs: 50,
|
|
82
|
+
cancelled: false,
|
|
83
|
+
completionTokens: 100,
|
|
84
|
+
promptTokens: 200,
|
|
85
|
+
promptCachedTokens: 50,
|
|
86
|
+
totalTokens: 300,
|
|
87
|
+
tokensPerSecond: 10,
|
|
88
|
+
metadata: {
|
|
89
|
+
modelProvider: "openai",
|
|
90
|
+
modelName: "gpt-4o"
|
|
91
|
+
}
|
|
92
|
+
};
|
|
93
|
+
const metrics2 = {
|
|
94
|
+
type: "llm_metrics",
|
|
95
|
+
label: "test",
|
|
96
|
+
requestId: "req2",
|
|
97
|
+
timestamp: Date.now(),
|
|
98
|
+
durationMs: 150,
|
|
99
|
+
ttftMs: 60,
|
|
100
|
+
cancelled: false,
|
|
101
|
+
completionTokens: 150,
|
|
102
|
+
promptTokens: 300,
|
|
103
|
+
promptCachedTokens: 75,
|
|
104
|
+
totalTokens: 450,
|
|
105
|
+
tokensPerSecond: 12,
|
|
106
|
+
metadata: {
|
|
107
|
+
modelProvider: "openai",
|
|
108
|
+
modelName: "gpt-4o"
|
|
109
|
+
}
|
|
110
|
+
};
|
|
111
|
+
collector.collect(metrics1);
|
|
112
|
+
collector.collect(metrics2);
|
|
113
|
+
const usage = collector.flatten();
|
|
114
|
+
expect(usage).toHaveLength(1);
|
|
115
|
+
const llmUsage = usage[0];
|
|
116
|
+
expect(llmUsage.type).toBe("llm_usage");
|
|
117
|
+
expect(llmUsage.provider).toBe("openai");
|
|
118
|
+
expect(llmUsage.model).toBe("gpt-4o");
|
|
119
|
+
expect(llmUsage.inputTokens).toBe(500);
|
|
120
|
+
expect(llmUsage.inputCachedTokens).toBe(125);
|
|
121
|
+
expect(llmUsage.outputTokens).toBe(250);
|
|
122
|
+
});
|
|
123
|
+
it("should separate metrics by different providers", () => {
|
|
124
|
+
const openaiMetrics = {
|
|
125
|
+
type: "llm_metrics",
|
|
126
|
+
label: "test",
|
|
127
|
+
requestId: "req1",
|
|
128
|
+
timestamp: Date.now(),
|
|
129
|
+
durationMs: 100,
|
|
130
|
+
ttftMs: 50,
|
|
131
|
+
cancelled: false,
|
|
132
|
+
completionTokens: 100,
|
|
133
|
+
promptTokens: 200,
|
|
134
|
+
promptCachedTokens: 0,
|
|
135
|
+
totalTokens: 300,
|
|
136
|
+
tokensPerSecond: 10,
|
|
137
|
+
metadata: {
|
|
138
|
+
modelProvider: "openai",
|
|
139
|
+
modelName: "gpt-4o"
|
|
140
|
+
}
|
|
141
|
+
};
|
|
142
|
+
const anthropicMetrics = {
|
|
143
|
+
type: "llm_metrics",
|
|
144
|
+
label: "test",
|
|
145
|
+
requestId: "req2",
|
|
146
|
+
timestamp: Date.now(),
|
|
147
|
+
durationMs: 120,
|
|
148
|
+
ttftMs: 55,
|
|
149
|
+
cancelled: false,
|
|
150
|
+
completionTokens: 80,
|
|
151
|
+
promptTokens: 150,
|
|
152
|
+
promptCachedTokens: 0,
|
|
153
|
+
totalTokens: 230,
|
|
154
|
+
tokensPerSecond: 8,
|
|
155
|
+
metadata: {
|
|
156
|
+
modelProvider: "anthropic",
|
|
157
|
+
modelName: "claude-3-5-sonnet"
|
|
158
|
+
}
|
|
159
|
+
};
|
|
160
|
+
collector.collect(openaiMetrics);
|
|
161
|
+
collector.collect(anthropicMetrics);
|
|
162
|
+
const usage = collector.flatten();
|
|
163
|
+
expect(usage).toHaveLength(2);
|
|
164
|
+
const openaiUsage = usage.find(
|
|
165
|
+
(u) => u.type === "llm_usage" && u.provider === "openai"
|
|
166
|
+
);
|
|
167
|
+
const anthropicUsage = usage.find(
|
|
168
|
+
(u) => u.type === "llm_usage" && u.provider === "anthropic"
|
|
169
|
+
);
|
|
170
|
+
expect(openaiUsage.inputTokens).toBe(200);
|
|
171
|
+
expect(openaiUsage.outputTokens).toBe(100);
|
|
172
|
+
expect(anthropicUsage.inputTokens).toBe(150);
|
|
173
|
+
expect(anthropicUsage.outputTokens).toBe(80);
|
|
174
|
+
});
|
|
175
|
+
});
|
|
176
|
+
describe("collect TTS metrics", () => {
|
|
177
|
+
it("should aggregate TTS metrics by provider and model", () => {
|
|
178
|
+
const metrics1 = {
|
|
179
|
+
type: "tts_metrics",
|
|
180
|
+
label: "test",
|
|
181
|
+
requestId: "req1",
|
|
182
|
+
timestamp: Date.now(),
|
|
183
|
+
ttfbMs: 100,
|
|
184
|
+
durationMs: 500,
|
|
185
|
+
audioDurationMs: 3e3,
|
|
186
|
+
cancelled: false,
|
|
187
|
+
charactersCount: 100,
|
|
188
|
+
inputTokens: 10,
|
|
189
|
+
outputTokens: 20,
|
|
190
|
+
streamed: true,
|
|
191
|
+
metadata: {
|
|
192
|
+
modelProvider: "elevenlabs",
|
|
193
|
+
modelName: "eleven_turbo_v2"
|
|
194
|
+
}
|
|
195
|
+
};
|
|
196
|
+
const metrics2 = {
|
|
197
|
+
type: "tts_metrics",
|
|
198
|
+
label: "test",
|
|
199
|
+
requestId: "req2",
|
|
200
|
+
timestamp: Date.now(),
|
|
201
|
+
ttfbMs: 120,
|
|
202
|
+
durationMs: 600,
|
|
203
|
+
audioDurationMs: 4e3,
|
|
204
|
+
cancelled: false,
|
|
205
|
+
charactersCount: 200,
|
|
206
|
+
inputTokens: 15,
|
|
207
|
+
outputTokens: 25,
|
|
208
|
+
streamed: true,
|
|
209
|
+
metadata: {
|
|
210
|
+
modelProvider: "elevenlabs",
|
|
211
|
+
modelName: "eleven_turbo_v2"
|
|
212
|
+
}
|
|
213
|
+
};
|
|
214
|
+
collector.collect(metrics1);
|
|
215
|
+
collector.collect(metrics2);
|
|
216
|
+
const usage = collector.flatten();
|
|
217
|
+
expect(usage).toHaveLength(1);
|
|
218
|
+
const ttsUsage = usage[0];
|
|
219
|
+
expect(ttsUsage.type).toBe("tts_usage");
|
|
220
|
+
expect(ttsUsage.provider).toBe("elevenlabs");
|
|
221
|
+
expect(ttsUsage.model).toBe("eleven_turbo_v2");
|
|
222
|
+
expect(ttsUsage.charactersCount).toBe(300);
|
|
223
|
+
expect(ttsUsage.audioDurationMs).toBe(7e3);
|
|
224
|
+
expect(ttsUsage.inputTokens).toBe(25);
|
|
225
|
+
expect(ttsUsage.outputTokens).toBe(45);
|
|
226
|
+
});
|
|
227
|
+
});
|
|
228
|
+
describe("collect STT metrics", () => {
|
|
229
|
+
it("should aggregate STT metrics by provider and model", () => {
|
|
230
|
+
const metrics1 = {
|
|
231
|
+
type: "stt_metrics",
|
|
232
|
+
label: "test",
|
|
233
|
+
requestId: "req1",
|
|
234
|
+
timestamp: Date.now(),
|
|
235
|
+
durationMs: 0,
|
|
236
|
+
audioDurationMs: 5e3,
|
|
237
|
+
inputTokens: 50,
|
|
238
|
+
outputTokens: 100,
|
|
239
|
+
streamed: true,
|
|
240
|
+
metadata: {
|
|
241
|
+
modelProvider: "deepgram",
|
|
242
|
+
modelName: "nova-2"
|
|
243
|
+
}
|
|
244
|
+
};
|
|
245
|
+
const metrics2 = {
|
|
246
|
+
type: "stt_metrics",
|
|
247
|
+
label: "test",
|
|
248
|
+
requestId: "req2",
|
|
249
|
+
timestamp: Date.now(),
|
|
250
|
+
durationMs: 0,
|
|
251
|
+
audioDurationMs: 3e3,
|
|
252
|
+
inputTokens: 30,
|
|
253
|
+
outputTokens: 60,
|
|
254
|
+
streamed: true,
|
|
255
|
+
metadata: {
|
|
256
|
+
modelProvider: "deepgram",
|
|
257
|
+
modelName: "nova-2"
|
|
258
|
+
}
|
|
259
|
+
};
|
|
260
|
+
collector.collect(metrics1);
|
|
261
|
+
collector.collect(metrics2);
|
|
262
|
+
const usage = collector.flatten();
|
|
263
|
+
expect(usage).toHaveLength(1);
|
|
264
|
+
const sttUsage = usage[0];
|
|
265
|
+
expect(sttUsage.type).toBe("stt_usage");
|
|
266
|
+
expect(sttUsage.provider).toBe("deepgram");
|
|
267
|
+
expect(sttUsage.model).toBe("nova-2");
|
|
268
|
+
expect(sttUsage.audioDurationMs).toBe(8e3);
|
|
269
|
+
expect(sttUsage.inputTokens).toBe(80);
|
|
270
|
+
expect(sttUsage.outputTokens).toBe(160);
|
|
271
|
+
});
|
|
272
|
+
});
|
|
273
|
+
describe("collect realtime model metrics", () => {
|
|
274
|
+
it("should aggregate realtime model metrics with detailed token breakdown", () => {
|
|
275
|
+
const metrics = {
|
|
276
|
+
type: "realtime_model_metrics",
|
|
277
|
+
label: "test",
|
|
278
|
+
requestId: "req1",
|
|
279
|
+
timestamp: Date.now(),
|
|
280
|
+
durationMs: 1e3,
|
|
281
|
+
ttftMs: 100,
|
|
282
|
+
cancelled: false,
|
|
283
|
+
inputTokens: 500,
|
|
284
|
+
outputTokens: 300,
|
|
285
|
+
totalTokens: 800,
|
|
286
|
+
tokensPerSecond: 10,
|
|
287
|
+
sessionDurationMs: 5e3,
|
|
288
|
+
inputTokenDetails: {
|
|
289
|
+
audioTokens: 200,
|
|
290
|
+
textTokens: 250,
|
|
291
|
+
imageTokens: 50,
|
|
292
|
+
cachedTokens: 100,
|
|
293
|
+
cachedTokensDetails: {
|
|
294
|
+
audioTokens: 30,
|
|
295
|
+
textTokens: 50,
|
|
296
|
+
imageTokens: 20
|
|
297
|
+
}
|
|
298
|
+
},
|
|
299
|
+
outputTokenDetails: {
|
|
300
|
+
textTokens: 200,
|
|
301
|
+
audioTokens: 100,
|
|
302
|
+
imageTokens: 0
|
|
303
|
+
},
|
|
304
|
+
metadata: {
|
|
305
|
+
modelProvider: "openai",
|
|
306
|
+
modelName: "gpt-4o-realtime"
|
|
307
|
+
}
|
|
308
|
+
};
|
|
309
|
+
collector.collect(metrics);
|
|
310
|
+
const usage = collector.flatten();
|
|
311
|
+
expect(usage).toHaveLength(1);
|
|
312
|
+
const llmUsage = usage[0];
|
|
313
|
+
expect(llmUsage.type).toBe("llm_usage");
|
|
314
|
+
expect(llmUsage.provider).toBe("openai");
|
|
315
|
+
expect(llmUsage.model).toBe("gpt-4o-realtime");
|
|
316
|
+
expect(llmUsage.inputTokens).toBe(500);
|
|
317
|
+
expect(llmUsage.inputCachedTokens).toBe(100);
|
|
318
|
+
expect(llmUsage.inputAudioTokens).toBe(200);
|
|
319
|
+
expect(llmUsage.inputCachedAudioTokens).toBe(30);
|
|
320
|
+
expect(llmUsage.inputTextTokens).toBe(250);
|
|
321
|
+
expect(llmUsage.inputCachedTextTokens).toBe(50);
|
|
322
|
+
expect(llmUsage.inputImageTokens).toBe(50);
|
|
323
|
+
expect(llmUsage.inputCachedImageTokens).toBe(20);
|
|
324
|
+
expect(llmUsage.outputTokens).toBe(300);
|
|
325
|
+
expect(llmUsage.outputTextTokens).toBe(200);
|
|
326
|
+
expect(llmUsage.outputAudioTokens).toBe(100);
|
|
327
|
+
expect(llmUsage.sessionDurationMs).toBe(5e3);
|
|
328
|
+
});
|
|
329
|
+
});
|
|
330
|
+
describe("mixed metrics collection", () => {
|
|
331
|
+
it("should collect and separate LLM, TTS, and STT metrics", () => {
|
|
332
|
+
const llmMetrics = {
|
|
333
|
+
type: "llm_metrics",
|
|
334
|
+
label: "test",
|
|
335
|
+
requestId: "req1",
|
|
336
|
+
timestamp: Date.now(),
|
|
337
|
+
durationMs: 100,
|
|
338
|
+
ttftMs: 50,
|
|
339
|
+
cancelled: false,
|
|
340
|
+
completionTokens: 100,
|
|
341
|
+
promptTokens: 200,
|
|
342
|
+
promptCachedTokens: 0,
|
|
343
|
+
totalTokens: 300,
|
|
344
|
+
tokensPerSecond: 10,
|
|
345
|
+
metadata: {
|
|
346
|
+
modelProvider: "openai",
|
|
347
|
+
modelName: "gpt-4o"
|
|
348
|
+
}
|
|
349
|
+
};
|
|
350
|
+
const ttsMetrics = {
|
|
351
|
+
type: "tts_metrics",
|
|
352
|
+
label: "test",
|
|
353
|
+
requestId: "req2",
|
|
354
|
+
timestamp: Date.now(),
|
|
355
|
+
ttfbMs: 100,
|
|
356
|
+
durationMs: 500,
|
|
357
|
+
audioDurationMs: 3e3,
|
|
358
|
+
cancelled: false,
|
|
359
|
+
charactersCount: 100,
|
|
360
|
+
streamed: true,
|
|
361
|
+
metadata: {
|
|
362
|
+
modelProvider: "elevenlabs",
|
|
363
|
+
modelName: "eleven_turbo_v2"
|
|
364
|
+
}
|
|
365
|
+
};
|
|
366
|
+
const sttMetrics = {
|
|
367
|
+
type: "stt_metrics",
|
|
368
|
+
label: "test",
|
|
369
|
+
requestId: "req3",
|
|
370
|
+
timestamp: Date.now(),
|
|
371
|
+
durationMs: 0,
|
|
372
|
+
audioDurationMs: 5e3,
|
|
373
|
+
streamed: true,
|
|
374
|
+
metadata: {
|
|
375
|
+
modelProvider: "deepgram",
|
|
376
|
+
modelName: "nova-2"
|
|
377
|
+
}
|
|
378
|
+
};
|
|
379
|
+
collector.collect(llmMetrics);
|
|
380
|
+
collector.collect(ttsMetrics);
|
|
381
|
+
collector.collect(sttMetrics);
|
|
382
|
+
const usage = collector.flatten();
|
|
383
|
+
expect(usage).toHaveLength(3);
|
|
384
|
+
const llmUsage = usage.find((u) => u.type === "llm_usage");
|
|
385
|
+
const ttsUsage = usage.find((u) => u.type === "tts_usage");
|
|
386
|
+
const sttUsage = usage.find((u) => u.type === "stt_usage");
|
|
387
|
+
expect(llmUsage).toBeDefined();
|
|
388
|
+
expect(ttsUsage).toBeDefined();
|
|
389
|
+
expect(sttUsage).toBeDefined();
|
|
390
|
+
});
|
|
391
|
+
});
|
|
392
|
+
describe("flatten returns copies", () => {
|
|
393
|
+
it("should return deep copies of usage objects", () => {
|
|
394
|
+
const metrics = {
|
|
395
|
+
type: "llm_metrics",
|
|
396
|
+
label: "test",
|
|
397
|
+
requestId: "req1",
|
|
398
|
+
timestamp: Date.now(),
|
|
399
|
+
durationMs: 100,
|
|
400
|
+
ttftMs: 50,
|
|
401
|
+
cancelled: false,
|
|
402
|
+
completionTokens: 100,
|
|
403
|
+
promptTokens: 200,
|
|
404
|
+
promptCachedTokens: 0,
|
|
405
|
+
totalTokens: 300,
|
|
406
|
+
tokensPerSecond: 10,
|
|
407
|
+
metadata: {
|
|
408
|
+
modelProvider: "openai",
|
|
409
|
+
modelName: "gpt-4o"
|
|
410
|
+
}
|
|
411
|
+
};
|
|
412
|
+
collector.collect(metrics);
|
|
413
|
+
const usage1 = collector.flatten();
|
|
414
|
+
const usage2 = collector.flatten();
|
|
415
|
+
expect(usage1[0]).toEqual(usage2[0]);
|
|
416
|
+
expect(usage1[0]).not.toBe(usage2[0]);
|
|
417
|
+
usage1[0].inputTokens = 9999;
|
|
418
|
+
expect(usage2[0].inputTokens).toBe(200);
|
|
419
|
+
});
|
|
420
|
+
});
|
|
421
|
+
describe("handles missing metadata", () => {
|
|
422
|
+
it("should use empty strings when metadata is missing", () => {
|
|
423
|
+
const metrics = {
|
|
424
|
+
type: "llm_metrics",
|
|
425
|
+
label: "test",
|
|
426
|
+
requestId: "req1",
|
|
427
|
+
timestamp: Date.now(),
|
|
428
|
+
durationMs: 100,
|
|
429
|
+
ttftMs: 50,
|
|
430
|
+
cancelled: false,
|
|
431
|
+
completionTokens: 100,
|
|
432
|
+
promptTokens: 200,
|
|
433
|
+
promptCachedTokens: 0,
|
|
434
|
+
totalTokens: 300,
|
|
435
|
+
tokensPerSecond: 10
|
|
436
|
+
// No metadata
|
|
437
|
+
};
|
|
438
|
+
collector.collect(metrics);
|
|
439
|
+
const usage = collector.flatten();
|
|
440
|
+
expect(usage).toHaveLength(1);
|
|
441
|
+
const llmUsage = usage[0];
|
|
442
|
+
expect(llmUsage.provider).toBe("");
|
|
443
|
+
expect(llmUsage.model).toBe("");
|
|
444
|
+
});
|
|
445
|
+
});
|
|
446
|
+
describe("ignores VAD and EOU metrics", () => {
|
|
447
|
+
it("should not collect VAD metrics", () => {
|
|
448
|
+
const vadMetrics = {
|
|
449
|
+
type: "vad_metrics",
|
|
450
|
+
label: "test",
|
|
451
|
+
timestamp: Date.now(),
|
|
452
|
+
idleTimeMs: 100,
|
|
453
|
+
inferenceDurationTotalMs: 50,
|
|
454
|
+
inferenceCount: 10
|
|
455
|
+
};
|
|
456
|
+
collector.collect(vadMetrics);
|
|
457
|
+
const usage = collector.flatten();
|
|
458
|
+
expect(usage).toHaveLength(0);
|
|
459
|
+
});
|
|
460
|
+
it("should not collect EOU metrics", () => {
|
|
461
|
+
const eouMetrics = {
|
|
462
|
+
type: "eou_metrics",
|
|
463
|
+
timestamp: Date.now(),
|
|
464
|
+
endOfUtteranceDelayMs: 100,
|
|
465
|
+
transcriptionDelayMs: 50,
|
|
466
|
+
onUserTurnCompletedDelayMs: 30,
|
|
467
|
+
lastSpeakingTimeMs: Date.now()
|
|
468
|
+
};
|
|
469
|
+
collector.collect(eouMetrics);
|
|
470
|
+
const usage = collector.flatten();
|
|
471
|
+
expect(usage).toHaveLength(0);
|
|
472
|
+
});
|
|
473
|
+
});
|
|
474
|
+
});
|
|
475
|
+
});
|
|
476
|
+
//# sourceMappingURL=model_usage.test.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"sources":["../../src/metrics/model_usage.test.ts"],"sourcesContent":["// SPDX-FileCopyrightText: 2024 LiveKit, Inc.\n//\n// SPDX-License-Identifier: Apache-2.0\nimport { beforeEach, describe, expect, it } from 'vitest';\nimport type { LLMMetrics, RealtimeModelMetrics, STTMetrics, TTSMetrics } from './base.js';\nimport {\n type LLMModelUsage,\n ModelUsageCollector,\n type STTModelUsage,\n type TTSModelUsage,\n filterZeroValues,\n} from './model_usage.js';\n\ndescribe('model_usage', () => {\n describe('filterZeroValues', () => {\n it('should filter out zero values from LLMModelUsage', () => {\n const usage: LLMModelUsage = {\n type: 'llm_usage',\n provider: 'openai',\n model: 'gpt-4o',\n inputTokens: 100,\n inputCachedTokens: 0,\n inputAudioTokens: 0,\n inputCachedAudioTokens: 0,\n inputTextTokens: 0,\n inputCachedTextTokens: 0,\n inputImageTokens: 0,\n inputCachedImageTokens: 0,\n outputTokens: 50,\n outputAudioTokens: 0,\n outputTextTokens: 0,\n sessionDurationMs: 0,\n };\n\n const filtered = filterZeroValues(usage);\n\n expect(filtered.type).toBe('llm_usage');\n expect(filtered.provider).toBe('openai');\n expect(filtered.model).toBe('gpt-4o');\n expect(filtered.inputTokens).toBe(100);\n expect(filtered.outputTokens).toBe(50);\n // Zero values should be filtered out\n expect(filtered.inputCachedTokens).toBeUndefined();\n expect(filtered.inputAudioTokens).toBeUndefined();\n expect(filtered.sessionDurationMs).toBeUndefined();\n });\n\n it('should filter out zero values from TTSModelUsage', () => {\n const usage: TTSModelUsage = {\n type: 'tts_usage',\n provider: 'elevenlabs',\n model: 'eleven_turbo_v2',\n inputTokens: 0,\n outputTokens: 0,\n charactersCount: 500,\n audioDurationMs: 3000,\n };\n\n const filtered = filterZeroValues(usage);\n\n expect(filtered.type).toBe('tts_usage');\n expect(filtered.provider).toBe('elevenlabs');\n expect(filtered.charactersCount).toBe(500);\n expect(filtered.audioDurationMs).toBe(3000);\n expect(filtered.inputTokens).toBeUndefined();\n expect(filtered.outputTokens).toBeUndefined();\n });\n\n it('should keep all values when none are zero', () => {\n const usage: STTModelUsage = {\n type: 'stt_usage',\n provider: 'deepgram',\n model: 'nova-2',\n inputTokens: 10,\n outputTokens: 20,\n audioDurationMs: 5000,\n };\n\n const filtered = filterZeroValues(usage);\n\n expect(Object.keys(filtered)).toHaveLength(6);\n expect(filtered).toEqual(usage);\n });\n });\n\n describe('ModelUsageCollector', () => {\n let collector: ModelUsageCollector;\n\n beforeEach(() => {\n collector = new ModelUsageCollector();\n });\n\n describe('collect LLM metrics', () => {\n it('should aggregate LLM metrics by provider and model', () => {\n const metrics1: LLMMetrics = {\n type: 'llm_metrics',\n label: 'test',\n requestId: 'req1',\n timestamp: Date.now(),\n durationMs: 100,\n ttftMs: 50,\n cancelled: false,\n completionTokens: 100,\n promptTokens: 200,\n promptCachedTokens: 50,\n totalTokens: 300,\n tokensPerSecond: 10,\n metadata: {\n modelProvider: 'openai',\n modelName: 'gpt-4o',\n },\n };\n\n const metrics2: LLMMetrics = {\n type: 'llm_metrics',\n label: 'test',\n requestId: 'req2',\n timestamp: Date.now(),\n durationMs: 150,\n ttftMs: 60,\n cancelled: false,\n completionTokens: 150,\n promptTokens: 300,\n promptCachedTokens: 75,\n totalTokens: 450,\n tokensPerSecond: 12,\n metadata: {\n modelProvider: 'openai',\n modelName: 'gpt-4o',\n },\n };\n\n collector.collect(metrics1);\n collector.collect(metrics2);\n\n const usage = collector.flatten();\n expect(usage).toHaveLength(1);\n\n const llmUsage = usage[0] as LLMModelUsage;\n expect(llmUsage.type).toBe('llm_usage');\n expect(llmUsage.provider).toBe('openai');\n expect(llmUsage.model).toBe('gpt-4o');\n expect(llmUsage.inputTokens).toBe(500); // 200 + 300\n expect(llmUsage.inputCachedTokens).toBe(125); // 50 + 75\n expect(llmUsage.outputTokens).toBe(250); // 100 + 150\n });\n\n it('should separate metrics by different providers', () => {\n const openaiMetrics: LLMMetrics = {\n type: 'llm_metrics',\n label: 'test',\n requestId: 'req1',\n timestamp: Date.now(),\n durationMs: 100,\n ttftMs: 50,\n cancelled: false,\n completionTokens: 100,\n promptTokens: 200,\n promptCachedTokens: 0,\n totalTokens: 300,\n tokensPerSecond: 10,\n metadata: {\n modelProvider: 'openai',\n modelName: 'gpt-4o',\n },\n };\n\n const anthropicMetrics: LLMMetrics = {\n type: 'llm_metrics',\n label: 'test',\n requestId: 'req2',\n timestamp: Date.now(),\n durationMs: 120,\n ttftMs: 55,\n cancelled: false,\n completionTokens: 80,\n promptTokens: 150,\n promptCachedTokens: 0,\n totalTokens: 230,\n tokensPerSecond: 8,\n metadata: {\n modelProvider: 'anthropic',\n modelName: 'claude-3-5-sonnet',\n },\n };\n\n collector.collect(openaiMetrics);\n collector.collect(anthropicMetrics);\n\n const usage = collector.flatten();\n expect(usage).toHaveLength(2);\n\n const openaiUsage = usage.find(\n (u) => u.type === 'llm_usage' && u.provider === 'openai',\n ) as LLMModelUsage;\n const anthropicUsage = usage.find(\n (u) => u.type === 'llm_usage' && u.provider === 'anthropic',\n ) as LLMModelUsage;\n\n expect(openaiUsage.inputTokens).toBe(200);\n expect(openaiUsage.outputTokens).toBe(100);\n expect(anthropicUsage.inputTokens).toBe(150);\n expect(anthropicUsage.outputTokens).toBe(80);\n });\n });\n\n describe('collect TTS metrics', () => {\n it('should aggregate TTS metrics by provider and model', () => {\n const metrics1: TTSMetrics = {\n type: 'tts_metrics',\n label: 'test',\n requestId: 'req1',\n timestamp: Date.now(),\n ttfbMs: 100,\n durationMs: 500,\n audioDurationMs: 3000,\n cancelled: false,\n charactersCount: 100,\n inputTokens: 10,\n outputTokens: 20,\n streamed: true,\n metadata: {\n modelProvider: 'elevenlabs',\n modelName: 'eleven_turbo_v2',\n },\n };\n\n const metrics2: TTSMetrics = {\n type: 'tts_metrics',\n label: 'test',\n requestId: 'req2',\n timestamp: Date.now(),\n ttfbMs: 120,\n durationMs: 600,\n audioDurationMs: 4000,\n cancelled: false,\n charactersCount: 200,\n inputTokens: 15,\n outputTokens: 25,\n streamed: true,\n metadata: {\n modelProvider: 'elevenlabs',\n modelName: 'eleven_turbo_v2',\n },\n };\n\n collector.collect(metrics1);\n collector.collect(metrics2);\n\n const usage = collector.flatten();\n expect(usage).toHaveLength(1);\n\n const ttsUsage = usage[0] as TTSModelUsage;\n expect(ttsUsage.type).toBe('tts_usage');\n expect(ttsUsage.provider).toBe('elevenlabs');\n expect(ttsUsage.model).toBe('eleven_turbo_v2');\n expect(ttsUsage.charactersCount).toBe(300); // 100 + 200\n expect(ttsUsage.audioDurationMs).toBe(7000); // 3000 + 4000\n expect(ttsUsage.inputTokens).toBe(25); // 10 + 15\n expect(ttsUsage.outputTokens).toBe(45); // 20 + 25\n });\n });\n\n describe('collect STT metrics', () => {\n it('should aggregate STT metrics by provider and model', () => {\n const metrics1: STTMetrics = {\n type: 'stt_metrics',\n label: 'test',\n requestId: 'req1',\n timestamp: Date.now(),\n durationMs: 0,\n audioDurationMs: 5000,\n inputTokens: 50,\n outputTokens: 100,\n streamed: true,\n metadata: {\n modelProvider: 'deepgram',\n modelName: 'nova-2',\n },\n };\n\n const metrics2: STTMetrics = {\n type: 'stt_metrics',\n label: 'test',\n requestId: 'req2',\n timestamp: Date.now(),\n durationMs: 0,\n audioDurationMs: 3000,\n inputTokens: 30,\n outputTokens: 60,\n streamed: true,\n metadata: {\n modelProvider: 'deepgram',\n modelName: 'nova-2',\n },\n };\n\n collector.collect(metrics1);\n collector.collect(metrics2);\n\n const usage = collector.flatten();\n expect(usage).toHaveLength(1);\n\n const sttUsage = usage[0] as STTModelUsage;\n expect(sttUsage.type).toBe('stt_usage');\n expect(sttUsage.provider).toBe('deepgram');\n expect(sttUsage.model).toBe('nova-2');\n expect(sttUsage.audioDurationMs).toBe(8000); // 5000 + 3000\n expect(sttUsage.inputTokens).toBe(80); // 50 + 30\n expect(sttUsage.outputTokens).toBe(160); // 100 + 60\n });\n });\n\n describe('collect realtime model metrics', () => {\n it('should aggregate realtime model metrics with detailed token breakdown', () => {\n const metrics: RealtimeModelMetrics = {\n type: 'realtime_model_metrics',\n label: 'test',\n requestId: 'req1',\n timestamp: Date.now(),\n durationMs: 1000,\n ttftMs: 100,\n cancelled: false,\n inputTokens: 500,\n outputTokens: 300,\n totalTokens: 800,\n tokensPerSecond: 10,\n sessionDurationMs: 5000,\n inputTokenDetails: {\n audioTokens: 200,\n textTokens: 250,\n imageTokens: 50,\n cachedTokens: 100,\n cachedTokensDetails: {\n audioTokens: 30,\n textTokens: 50,\n imageTokens: 20,\n },\n },\n outputTokenDetails: {\n textTokens: 200,\n audioTokens: 100,\n imageTokens: 0,\n },\n metadata: {\n modelProvider: 'openai',\n modelName: 'gpt-4o-realtime',\n },\n };\n\n collector.collect(metrics);\n\n const usage = collector.flatten();\n expect(usage).toHaveLength(1);\n\n const llmUsage = usage[0] as LLMModelUsage;\n expect(llmUsage.type).toBe('llm_usage');\n expect(llmUsage.provider).toBe('openai');\n expect(llmUsage.model).toBe('gpt-4o-realtime');\n expect(llmUsage.inputTokens).toBe(500);\n expect(llmUsage.inputCachedTokens).toBe(100);\n expect(llmUsage.inputAudioTokens).toBe(200);\n expect(llmUsage.inputCachedAudioTokens).toBe(30);\n expect(llmUsage.inputTextTokens).toBe(250);\n expect(llmUsage.inputCachedTextTokens).toBe(50);\n expect(llmUsage.inputImageTokens).toBe(50);\n expect(llmUsage.inputCachedImageTokens).toBe(20);\n expect(llmUsage.outputTokens).toBe(300);\n expect(llmUsage.outputTextTokens).toBe(200);\n expect(llmUsage.outputAudioTokens).toBe(100);\n expect(llmUsage.sessionDurationMs).toBe(5000);\n });\n });\n\n describe('mixed metrics collection', () => {\n it('should collect and separate LLM, TTS, and STT metrics', () => {\n const llmMetrics: LLMMetrics = {\n type: 'llm_metrics',\n label: 'test',\n requestId: 'req1',\n timestamp: Date.now(),\n durationMs: 100,\n ttftMs: 50,\n cancelled: false,\n completionTokens: 100,\n promptTokens: 200,\n promptCachedTokens: 0,\n totalTokens: 300,\n tokensPerSecond: 10,\n metadata: {\n modelProvider: 'openai',\n modelName: 'gpt-4o',\n },\n };\n\n const ttsMetrics: TTSMetrics = {\n type: 'tts_metrics',\n label: 'test',\n requestId: 'req2',\n timestamp: Date.now(),\n ttfbMs: 100,\n durationMs: 500,\n audioDurationMs: 3000,\n cancelled: false,\n charactersCount: 100,\n streamed: true,\n metadata: {\n modelProvider: 'elevenlabs',\n modelName: 'eleven_turbo_v2',\n },\n };\n\n const sttMetrics: STTMetrics = {\n type: 'stt_metrics',\n label: 'test',\n requestId: 'req3',\n timestamp: Date.now(),\n durationMs: 0,\n audioDurationMs: 5000,\n streamed: true,\n metadata: {\n modelProvider: 'deepgram',\n modelName: 'nova-2',\n },\n };\n\n collector.collect(llmMetrics);\n collector.collect(ttsMetrics);\n collector.collect(sttMetrics);\n\n const usage = collector.flatten();\n expect(usage).toHaveLength(3);\n\n const llmUsage = usage.find((u) => u.type === 'llm_usage');\n const ttsUsage = usage.find((u) => u.type === 'tts_usage');\n const sttUsage = usage.find((u) => u.type === 'stt_usage');\n\n expect(llmUsage).toBeDefined();\n expect(ttsUsage).toBeDefined();\n expect(sttUsage).toBeDefined();\n });\n });\n\n describe('flatten returns copies', () => {\n it('should return deep copies of usage objects', () => {\n const metrics: LLMMetrics = {\n type: 'llm_metrics',\n label: 'test',\n requestId: 'req1',\n timestamp: Date.now(),\n durationMs: 100,\n ttftMs: 50,\n cancelled: false,\n completionTokens: 100,\n promptTokens: 200,\n promptCachedTokens: 0,\n totalTokens: 300,\n tokensPerSecond: 10,\n metadata: {\n modelProvider: 'openai',\n modelName: 'gpt-4o',\n },\n };\n\n collector.collect(metrics);\n\n const usage1 = collector.flatten();\n const usage2 = collector.flatten();\n\n // Should be equal values\n expect(usage1[0]).toEqual(usage2[0]);\n\n // But not the same object reference\n expect(usage1[0]).not.toBe(usage2[0]);\n\n // Modifying one shouldn't affect the other\n (usage1[0] as LLMModelUsage).inputTokens = 9999;\n expect((usage2[0] as LLMModelUsage).inputTokens).toBe(200);\n });\n });\n\n describe('handles missing metadata', () => {\n it('should use empty strings when metadata is missing', () => {\n const metrics: LLMMetrics = {\n type: 'llm_metrics',\n label: 'test',\n requestId: 'req1',\n timestamp: Date.now(),\n durationMs: 100,\n ttftMs: 50,\n cancelled: false,\n completionTokens: 100,\n promptTokens: 200,\n promptCachedTokens: 0,\n totalTokens: 300,\n tokensPerSecond: 10,\n // No metadata\n };\n\n collector.collect(metrics);\n\n const usage = collector.flatten();\n expect(usage).toHaveLength(1);\n\n const llmUsage = usage[0] as LLMModelUsage;\n expect(llmUsage.provider).toBe('');\n expect(llmUsage.model).toBe('');\n });\n });\n\n describe('ignores VAD and EOU metrics', () => {\n it('should not collect VAD metrics', () => {\n const vadMetrics = {\n type: 'vad_metrics' as const,\n label: 'test',\n timestamp: Date.now(),\n idleTimeMs: 100,\n inferenceDurationTotalMs: 50,\n inferenceCount: 10,\n };\n\n collector.collect(vadMetrics);\n\n const usage = collector.flatten();\n expect(usage).toHaveLength(0);\n });\n\n it('should not collect EOU metrics', () => {\n const eouMetrics = {\n type: 'eou_metrics' as const,\n timestamp: Date.now(),\n endOfUtteranceDelayMs: 100,\n transcriptionDelayMs: 50,\n onUserTurnCompletedDelayMs: 30,\n lastSpeakingTimeMs: Date.now(),\n };\n\n collector.collect(eouMetrics);\n\n const usage = collector.flatten();\n expect(usage).toHaveLength(0);\n });\n });\n });\n});\n"],"mappings":"AAGA,SAAS,YAAY,UAAU,QAAQ,UAAU;AAEjD;AAAA,EAEE;AAAA,EAGA;AAAA,OACK;AAEP,SAAS,eAAe,MAAM;AAC5B,WAAS,oBAAoB,MAAM;AACjC,OAAG,oDAAoD,MAAM;AAC3D,YAAM,QAAuB;AAAA,QAC3B,MAAM;AAAA,QACN,UAAU;AAAA,QACV,OAAO;AAAA,QACP,aAAa;AAAA,QACb,mBAAmB;AAAA,QACnB,kBAAkB;AAAA,QAClB,wBAAwB;AAAA,QACxB,iBAAiB;AAAA,QACjB,uBAAuB;AAAA,QACvB,kBAAkB;AAAA,QAClB,wBAAwB;AAAA,QACxB,cAAc;AAAA,QACd,mBAAmB;AAAA,QACnB,kBAAkB;AAAA,QAClB,mBAAmB;AAAA,MACrB;AAEA,YAAM,WAAW,iBAAiB,KAAK;AAEvC,aAAO,SAAS,IAAI,EAAE,KAAK,WAAW;AACtC,aAAO,SAAS,QAAQ,EAAE,KAAK,QAAQ;AACvC,aAAO,SAAS,KAAK,EAAE,KAAK,QAAQ;AACpC,aAAO,SAAS,WAAW,EAAE,KAAK,GAAG;AACrC,aAAO,SAAS,YAAY,EAAE,KAAK,EAAE;AAErC,aAAO,SAAS,iBAAiB,EAAE,cAAc;AACjD,aAAO,SAAS,gBAAgB,EAAE,cAAc;AAChD,aAAO,SAAS,iBAAiB,EAAE,cAAc;AAAA,IACnD,CAAC;AAED,OAAG,oDAAoD,MAAM;AAC3D,YAAM,QAAuB;AAAA,QAC3B,MAAM;AAAA,QACN,UAAU;AAAA,QACV,OAAO;AAAA,QACP,aAAa;AAAA,QACb,cAAc;AAAA,QACd,iBAAiB;AAAA,QACjB,iBAAiB;AAAA,MACnB;AAEA,YAAM,WAAW,iBAAiB,KAAK;AAEvC,aAAO,SAAS,IAAI,EAAE,KAAK,WAAW;AACtC,aAAO,SAAS,QAAQ,EAAE,KAAK,YAAY;AAC3C,aAAO,SAAS,eAAe,EAAE,KAAK,GAAG;AACzC,aAAO,SAAS,eAAe,EAAE,KAAK,GAAI;AAC1C,aAAO,SAAS,WAAW,EAAE,cAAc;AAC3C,aAAO,SAAS,YAAY,EAAE,cAAc;AAAA,IAC9C,CAAC;AAED,OAAG,6CAA6C,MAAM;AACpD,YAAM,QAAuB;AAAA,QAC3B,MAAM;AAAA,QACN,UAAU;AAAA,QACV,OAAO;AAAA,QACP,aAAa;AAAA,QACb,cAAc;AAAA,QACd,iBAAiB;AAAA,MACnB;AAEA,YAAM,WAAW,iBAAiB,KAAK;AAEvC,aAAO,OAAO,KAAK,QAAQ,CAAC,EAAE,aAAa,CAAC;AAC5C,aAAO,QAAQ,EAAE,QAAQ,KAAK;AAAA,IAChC,CAAC;AAAA,EACH,CAAC;AAED,WAAS,uBAAuB,MAAM;AACpC,QAAI;AAEJ,eAAW,MAAM;AACf,kBAAY,IAAI,oBAAoB;AAAA,IACtC,CAAC;AAED,aAAS,uBAAuB,MAAM;AACpC,SAAG,sDAAsD,MAAM;AAC7D,cAAM,WAAuB;AAAA,UAC3B,MAAM;AAAA,UACN,OAAO;AAAA,UACP,WAAW;AAAA,UACX,WAAW,KAAK,IAAI;AAAA,UACpB,YAAY;AAAA,UACZ,QAAQ;AAAA,UACR,WAAW;AAAA,UACX,kBAAkB;AAAA,UAClB,cAAc;AAAA,UACd,oBAAoB;AAAA,UACpB,aAAa;AAAA,UACb,iBAAiB;AAAA,UACjB,UAAU;AAAA,YACR,eAAe;AAAA,YACf,WAAW;AAAA,UACb;AAAA,QACF;AAEA,cAAM,WAAuB;AAAA,UAC3B,MAAM;AAAA,UACN,OAAO;AAAA,UACP,WAAW;AAAA,UACX,WAAW,KAAK,IAAI;AAAA,UACpB,YAAY;AAAA,UACZ,QAAQ;AAAA,UACR,WAAW;AAAA,UACX,kBAAkB;AAAA,UAClB,cAAc;AAAA,UACd,oBAAoB;AAAA,UACpB,aAAa;AAAA,UACb,iBAAiB;AAAA,UACjB,UAAU;AAAA,YACR,eAAe;AAAA,YACf,WAAW;AAAA,UACb;AAAA,QACF;AAEA,kBAAU,QAAQ,QAAQ;AAC1B,kBAAU,QAAQ,QAAQ;AAE1B,cAAM,QAAQ,UAAU,QAAQ;AAChC,eAAO,KAAK,EAAE,aAAa,CAAC;AAE5B,cAAM,WAAW,MAAM,CAAC;AACxB,eAAO,SAAS,IAAI,EAAE,KAAK,WAAW;AACtC,eAAO,SAAS,QAAQ,EAAE,KAAK,QAAQ;AACvC,eAAO,SAAS,KAAK,EAAE,KAAK,QAAQ;AACpC,eAAO,SAAS,WAAW,EAAE,KAAK,GAAG;AACrC,eAAO,SAAS,iBAAiB,EAAE,KAAK,GAAG;AAC3C,eAAO,SAAS,YAAY,EAAE,KAAK,GAAG;AAAA,MACxC,CAAC;AAED,SAAG,kDAAkD,MAAM;AACzD,cAAM,gBAA4B;AAAA,UAChC,MAAM;AAAA,UACN,OAAO;AAAA,UACP,WAAW;AAAA,UACX,WAAW,KAAK,IAAI;AAAA,UACpB,YAAY;AAAA,UACZ,QAAQ;AAAA,UACR,WAAW;AAAA,UACX,kBAAkB;AAAA,UAClB,cAAc;AAAA,UACd,oBAAoB;AAAA,UACpB,aAAa;AAAA,UACb,iBAAiB;AAAA,UACjB,UAAU;AAAA,YACR,eAAe;AAAA,YACf,WAAW;AAAA,UACb;AAAA,QACF;AAEA,cAAM,mBAA+B;AAAA,UACnC,MAAM;AAAA,UACN,OAAO;AAAA,UACP,WAAW;AAAA,UACX,WAAW,KAAK,IAAI;AAAA,UACpB,YAAY;AAAA,UACZ,QAAQ;AAAA,UACR,WAAW;AAAA,UACX,kBAAkB;AAAA,UAClB,cAAc;AAAA,UACd,oBAAoB;AAAA,UACpB,aAAa;AAAA,UACb,iBAAiB;AAAA,UACjB,UAAU;AAAA,YACR,eAAe;AAAA,YACf,WAAW;AAAA,UACb;AAAA,QACF;AAEA,kBAAU,QAAQ,aAAa;AAC/B,kBAAU,QAAQ,gBAAgB;AAElC,cAAM,QAAQ,UAAU,QAAQ;AAChC,eAAO,KAAK,EAAE,aAAa,CAAC;AAE5B,cAAM,cAAc,MAAM;AAAA,UACxB,CAAC,MAAM,EAAE,SAAS,eAAe,EAAE,aAAa;AAAA,QAClD;AACA,cAAM,iBAAiB,MAAM;AAAA,UAC3B,CAAC,MAAM,EAAE,SAAS,eAAe,EAAE,aAAa;AAAA,QAClD;AAEA,eAAO,YAAY,WAAW,EAAE,KAAK,GAAG;AACxC,eAAO,YAAY,YAAY,EAAE,KAAK,GAAG;AACzC,eAAO,eAAe,WAAW,EAAE,KAAK,GAAG;AAC3C,eAAO,eAAe,YAAY,EAAE,KAAK,EAAE;AAAA,MAC7C,CAAC;AAAA,IACH,CAAC;AAED,aAAS,uBAAuB,MAAM;AACpC,SAAG,sDAAsD,MAAM;AAC7D,cAAM,WAAuB;AAAA,UAC3B,MAAM;AAAA,UACN,OAAO;AAAA,UACP,WAAW;AAAA,UACX,WAAW,KAAK,IAAI;AAAA,UACpB,QAAQ;AAAA,UACR,YAAY;AAAA,UACZ,iBAAiB;AAAA,UACjB,WAAW;AAAA,UACX,iBAAiB;AAAA,UACjB,aAAa;AAAA,UACb,cAAc;AAAA,UACd,UAAU;AAAA,UACV,UAAU;AAAA,YACR,eAAe;AAAA,YACf,WAAW;AAAA,UACb;AAAA,QACF;AAEA,cAAM,WAAuB;AAAA,UAC3B,MAAM;AAAA,UACN,OAAO;AAAA,UACP,WAAW;AAAA,UACX,WAAW,KAAK,IAAI;AAAA,UACpB,QAAQ;AAAA,UACR,YAAY;AAAA,UACZ,iBAAiB;AAAA,UACjB,WAAW;AAAA,UACX,iBAAiB;AAAA,UACjB,aAAa;AAAA,UACb,cAAc;AAAA,UACd,UAAU;AAAA,UACV,UAAU;AAAA,YACR,eAAe;AAAA,YACf,WAAW;AAAA,UACb;AAAA,QACF;AAEA,kBAAU,QAAQ,QAAQ;AAC1B,kBAAU,QAAQ,QAAQ;AAE1B,cAAM,QAAQ,UAAU,QAAQ;AAChC,eAAO,KAAK,EAAE,aAAa,CAAC;AAE5B,cAAM,WAAW,MAAM,CAAC;AACxB,eAAO,SAAS,IAAI,EAAE,KAAK,WAAW;AACtC,eAAO,SAAS,QAAQ,EAAE,KAAK,YAAY;AAC3C,eAAO,SAAS,KAAK,EAAE,KAAK,iBAAiB;AAC7C,eAAO,SAAS,eAAe,EAAE,KAAK,GAAG;AACzC,eAAO,SAAS,eAAe,EAAE,KAAK,GAAI;AAC1C,eAAO,SAAS,WAAW,EAAE,KAAK,EAAE;AACpC,eAAO,SAAS,YAAY,EAAE,KAAK,EAAE;AAAA,MACvC,CAAC;AAAA,IACH,CAAC;AAED,aAAS,uBAAuB,MAAM;AACpC,SAAG,sDAAsD,MAAM;AAC7D,cAAM,WAAuB;AAAA,UAC3B,MAAM;AAAA,UACN,OAAO;AAAA,UACP,WAAW;AAAA,UACX,WAAW,KAAK,IAAI;AAAA,UACpB,YAAY;AAAA,UACZ,iBAAiB;AAAA,UACjB,aAAa;AAAA,UACb,cAAc;AAAA,UACd,UAAU;AAAA,UACV,UAAU;AAAA,YACR,eAAe;AAAA,YACf,WAAW;AAAA,UACb;AAAA,QACF;AAEA,cAAM,WAAuB;AAAA,UAC3B,MAAM;AAAA,UACN,OAAO;AAAA,UACP,WAAW;AAAA,UACX,WAAW,KAAK,IAAI;AAAA,UACpB,YAAY;AAAA,UACZ,iBAAiB;AAAA,UACjB,aAAa;AAAA,UACb,cAAc;AAAA,UACd,UAAU;AAAA,UACV,UAAU;AAAA,YACR,eAAe;AAAA,YACf,WAAW;AAAA,UACb;AAAA,QACF;AAEA,kBAAU,QAAQ,QAAQ;AAC1B,kBAAU,QAAQ,QAAQ;AAE1B,cAAM,QAAQ,UAAU,QAAQ;AAChC,eAAO,KAAK,EAAE,aAAa,CAAC;AAE5B,cAAM,WAAW,MAAM,CAAC;AACxB,eAAO,SAAS,IAAI,EAAE,KAAK,WAAW;AACtC,eAAO,SAAS,QAAQ,EAAE,KAAK,UAAU;AACzC,eAAO,SAAS,KAAK,EAAE,KAAK,QAAQ;AACpC,eAAO,SAAS,eAAe,EAAE,KAAK,GAAI;AAC1C,eAAO,SAAS,WAAW,EAAE,KAAK,EAAE;AACpC,eAAO,SAAS,YAAY,EAAE,KAAK,GAAG;AAAA,MACxC,CAAC;AAAA,IACH,CAAC;AAED,aAAS,kCAAkC,MAAM;AAC/C,SAAG,yEAAyE,MAAM;AAChF,cAAM,UAAgC;AAAA,UACpC,MAAM;AAAA,UACN,OAAO;AAAA,UACP,WAAW;AAAA,UACX,WAAW,KAAK,IAAI;AAAA,UACpB,YAAY;AAAA,UACZ,QAAQ;AAAA,UACR,WAAW;AAAA,UACX,aAAa;AAAA,UACb,cAAc;AAAA,UACd,aAAa;AAAA,UACb,iBAAiB;AAAA,UACjB,mBAAmB;AAAA,UACnB,mBAAmB;AAAA,YACjB,aAAa;AAAA,YACb,YAAY;AAAA,YACZ,aAAa;AAAA,YACb,cAAc;AAAA,YACd,qBAAqB;AAAA,cACnB,aAAa;AAAA,cACb,YAAY;AAAA,cACZ,aAAa;AAAA,YACf;AAAA,UACF;AAAA,UACA,oBAAoB;AAAA,YAClB,YAAY;AAAA,YACZ,aAAa;AAAA,YACb,aAAa;AAAA,UACf;AAAA,UACA,UAAU;AAAA,YACR,eAAe;AAAA,YACf,WAAW;AAAA,UACb;AAAA,QACF;AAEA,kBAAU,QAAQ,OAAO;AAEzB,cAAM,QAAQ,UAAU,QAAQ;AAChC,eAAO,KAAK,EAAE,aAAa,CAAC;AAE5B,cAAM,WAAW,MAAM,CAAC;AACxB,eAAO,SAAS,IAAI,EAAE,KAAK,WAAW;AACtC,eAAO,SAAS,QAAQ,EAAE,KAAK,QAAQ;AACvC,eAAO,SAAS,KAAK,EAAE,KAAK,iBAAiB;AAC7C,eAAO,SAAS,WAAW,EAAE,KAAK,GAAG;AACrC,eAAO,SAAS,iBAAiB,EAAE,KAAK,GAAG;AAC3C,eAAO,SAAS,gBAAgB,EAAE,KAAK,GAAG;AAC1C,eAAO,SAAS,sBAAsB,EAAE,KAAK,EAAE;AAC/C,eAAO,SAAS,eAAe,EAAE,KAAK,GAAG;AACzC,eAAO,SAAS,qBAAqB,EAAE,KAAK,EAAE;AAC9C,eAAO,SAAS,gBAAgB,EAAE,KAAK,EAAE;AACzC,eAAO,SAAS,sBAAsB,EAAE,KAAK,EAAE;AAC/C,eAAO,SAAS,YAAY,EAAE,KAAK,GAAG;AACtC,eAAO,SAAS,gBAAgB,EAAE,KAAK,GAAG;AAC1C,eAAO,SAAS,iBAAiB,EAAE,KAAK,GAAG;AAC3C,eAAO,SAAS,iBAAiB,EAAE,KAAK,GAAI;AAAA,MAC9C,CAAC;AAAA,IACH,CAAC;AAED,aAAS,4BAA4B,MAAM;AACzC,SAAG,yDAAyD,MAAM;AAChE,cAAM,aAAyB;AAAA,UAC7B,MAAM;AAAA,UACN,OAAO;AAAA,UACP,WAAW;AAAA,UACX,WAAW,KAAK,IAAI;AAAA,UACpB,YAAY;AAAA,UACZ,QAAQ;AAAA,UACR,WAAW;AAAA,UACX,kBAAkB;AAAA,UAClB,cAAc;AAAA,UACd,oBAAoB;AAAA,UACpB,aAAa;AAAA,UACb,iBAAiB;AAAA,UACjB,UAAU;AAAA,YACR,eAAe;AAAA,YACf,WAAW;AAAA,UACb;AAAA,QACF;AAEA,cAAM,aAAyB;AAAA,UAC7B,MAAM;AAAA,UACN,OAAO;AAAA,UACP,WAAW;AAAA,UACX,WAAW,KAAK,IAAI;AAAA,UACpB,QAAQ;AAAA,UACR,YAAY;AAAA,UACZ,iBAAiB;AAAA,UACjB,WAAW;AAAA,UACX,iBAAiB;AAAA,UACjB,UAAU;AAAA,UACV,UAAU;AAAA,YACR,eAAe;AAAA,YACf,WAAW;AAAA,UACb;AAAA,QACF;AAEA,cAAM,aAAyB;AAAA,UAC7B,MAAM;AAAA,UACN,OAAO;AAAA,UACP,WAAW;AAAA,UACX,WAAW,KAAK,IAAI;AAAA,UACpB,YAAY;AAAA,UACZ,iBAAiB;AAAA,UACjB,UAAU;AAAA,UACV,UAAU;AAAA,YACR,eAAe;AAAA,YACf,WAAW;AAAA,UACb;AAAA,QACF;AAEA,kBAAU,QAAQ,UAAU;AAC5B,kBAAU,QAAQ,UAAU;AAC5B,kBAAU,QAAQ,UAAU;AAE5B,cAAM,QAAQ,UAAU,QAAQ;AAChC,eAAO,KAAK,EAAE,aAAa,CAAC;AAE5B,cAAM,WAAW,MAAM,KAAK,CAAC,MAAM,EAAE,SAAS,WAAW;AACzD,cAAM,WAAW,MAAM,KAAK,CAAC,MAAM,EAAE,SAAS,WAAW;AACzD,cAAM,WAAW,MAAM,KAAK,CAAC,MAAM,EAAE,SAAS,WAAW;AAEzD,eAAO,QAAQ,EAAE,YAAY;AAC7B,eAAO,QAAQ,EAAE,YAAY;AAC7B,eAAO,QAAQ,EAAE,YAAY;AAAA,MAC/B,CAAC;AAAA,IACH,CAAC;AAED,aAAS,0BAA0B,MAAM;AACvC,SAAG,8CAA8C,MAAM;AACrD,cAAM,UAAsB;AAAA,UAC1B,MAAM;AAAA,UACN,OAAO;AAAA,UACP,WAAW;AAAA,UACX,WAAW,KAAK,IAAI;AAAA,UACpB,YAAY;AAAA,UACZ,QAAQ;AAAA,UACR,WAAW;AAAA,UACX,kBAAkB;AAAA,UAClB,cAAc;AAAA,UACd,oBAAoB;AAAA,UACpB,aAAa;AAAA,UACb,iBAAiB;AAAA,UACjB,UAAU;AAAA,YACR,eAAe;AAAA,YACf,WAAW;AAAA,UACb;AAAA,QACF;AAEA,kBAAU,QAAQ,OAAO;AAEzB,cAAM,SAAS,UAAU,QAAQ;AACjC,cAAM,SAAS,UAAU,QAAQ;AAGjC,eAAO,OAAO,CAAC,CAAC,EAAE,QAAQ,OAAO,CAAC,CAAC;AAGnC,eAAO,OAAO,CAAC,CAAC,EAAE,IAAI,KAAK,OAAO,CAAC,CAAC;AAGpC,QAAC,OAAO,CAAC,EAAoB,cAAc;AAC3C,eAAQ,OAAO,CAAC,EAAoB,WAAW,EAAE,KAAK,GAAG;AAAA,MAC3D,CAAC;AAAA,IACH,CAAC;AAED,aAAS,4BAA4B,MAAM;AACzC,SAAG,qDAAqD,MAAM;AAC5D,cAAM,UAAsB;AAAA,UAC1B,MAAM;AAAA,UACN,OAAO;AAAA,UACP,WAAW;AAAA,UACX,WAAW,KAAK,IAAI;AAAA,UACpB,YAAY;AAAA,UACZ,QAAQ;AAAA,UACR,WAAW;AAAA,UACX,kBAAkB;AAAA,UAClB,cAAc;AAAA,UACd,oBAAoB;AAAA,UACpB,aAAa;AAAA,UACb,iBAAiB;AAAA;AAAA,QAEnB;AAEA,kBAAU,QAAQ,OAAO;AAEzB,cAAM,QAAQ,UAAU,QAAQ;AAChC,eAAO,KAAK,EAAE,aAAa,CAAC;AAE5B,cAAM,WAAW,MAAM,CAAC;AACxB,eAAO,SAAS,QAAQ,EAAE,KAAK,EAAE;AACjC,eAAO,SAAS,KAAK,EAAE,KAAK,EAAE;AAAA,MAChC,CAAC;AAAA,IACH,CAAC;AAED,aAAS,+BAA+B,MAAM;AAC5C,SAAG,kCAAkC,MAAM;AACzC,cAAM,aAAa;AAAA,UACjB,MAAM;AAAA,UACN,OAAO;AAAA,UACP,WAAW,KAAK,IAAI;AAAA,UACpB,YAAY;AAAA,UACZ,0BAA0B;AAAA,UAC1B,gBAAgB;AAAA,QAClB;AAEA,kBAAU,QAAQ,UAAU;AAE5B,cAAM,QAAQ,UAAU,QAAQ;AAChC,eAAO,KAAK,EAAE,aAAa,CAAC;AAAA,MAC9B,CAAC;AAED,SAAG,kCAAkC,MAAM;AACzC,cAAM,aAAa;AAAA,UACjB,MAAM;AAAA,UACN,WAAW,KAAK,IAAI;AAAA,UACpB,uBAAuB;AAAA,UACvB,sBAAsB;AAAA,UACtB,4BAA4B;AAAA,UAC5B,oBAAoB,KAAK,IAAI;AAAA,QAC/B;AAEA,kBAAU,QAAQ,UAAU;AAE5B,cAAM,QAAQ,UAAU,QAAQ;AAChC,eAAO,KAAK,EAAE,aAAa,CAAC;AAAA,MAC9B,CAAC;AAAA,IACH,CAAC;AAAA,EACH,CAAC;AACH,CAAC;","names":[]}
|
|
@@ -21,9 +21,12 @@ __export(usage_collector_exports, {
|
|
|
21
21
|
UsageCollector: () => UsageCollector
|
|
22
22
|
});
|
|
23
23
|
module.exports = __toCommonJS(usage_collector_exports);
|
|
24
|
+
var import_log = require("../log.cjs");
|
|
24
25
|
class UsageCollector {
|
|
25
26
|
summary;
|
|
27
|
+
logger = (0, import_log.log)();
|
|
26
28
|
constructor() {
|
|
29
|
+
this.logger.warn("UsageCollector is deprecated. Use ModelUsageCollector instead.");
|
|
27
30
|
this.summary = {
|
|
28
31
|
llmPromptTokens: 0,
|
|
29
32
|
llmPromptCachedTokens: 0,
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"sources":["../../src/metrics/usage_collector.ts"],"sourcesContent":["// SPDX-FileCopyrightText: 2024 LiveKit, Inc.\n//\n// SPDX-License-Identifier: Apache-2.0\nimport type { AgentMetrics } from './base.js';\n\nexport interface UsageSummary {\n llmPromptTokens: number;\n llmPromptCachedTokens: number;\n llmCompletionTokens: number;\n ttsCharactersCount: number;\n sttAudioDurationMs: number;\n}\n\nexport class UsageCollector {\n private summary: UsageSummary;\n\n constructor() {\n this.summary = {\n llmPromptTokens: 0,\n llmPromptCachedTokens: 0,\n llmCompletionTokens: 0,\n ttsCharactersCount: 0,\n sttAudioDurationMs: 0,\n };\n }\n\n collect(metrics: AgentMetrics): void {\n if (metrics.type === 'llm_metrics') {\n this.summary.llmPromptTokens += metrics.promptTokens;\n this.summary.llmPromptCachedTokens += metrics.promptCachedTokens;\n this.summary.llmCompletionTokens += metrics.completionTokens;\n } else if (metrics.type === 'realtime_model_metrics') {\n this.summary.llmPromptTokens += metrics.inputTokens;\n this.summary.llmPromptCachedTokens += metrics.inputTokenDetails.cachedTokens;\n this.summary.llmCompletionTokens += metrics.outputTokens;\n } else if (metrics.type === 'tts_metrics') {\n this.summary.ttsCharactersCount += metrics.charactersCount;\n } else if (metrics.type === 'stt_metrics') {\n this.summary.sttAudioDurationMs += metrics.audioDurationMs;\n }\n }\n\n getSummary(): UsageSummary {\n return { ...this.summary };\n }\n}\n"],"mappings":";;;;;;;;;;;;;;;;;;AAAA;AAAA;AAAA;AAAA;AAAA;
|
|
1
|
+
{"version":3,"sources":["../../src/metrics/usage_collector.ts"],"sourcesContent":["// SPDX-FileCopyrightText: 2024 LiveKit, Inc.\n//\n// SPDX-License-Identifier: Apache-2.0\nimport { log } from '../log.js';\nimport type { AgentMetrics } from './base.js';\n\n/**\n * @deprecated Use LLMModelUsage, TTSModelUsage, or STTModelUsage instead.\n * These new types provide per-model/provider usage aggregation for more detailed tracking.\n */\nexport interface UsageSummary {\n llmPromptTokens: number;\n llmPromptCachedTokens: number;\n llmCompletionTokens: number;\n ttsCharactersCount: number;\n sttAudioDurationMs: number;\n}\n\n/**\n * @deprecated Use ModelUsageCollector instead.\n * ModelUsageCollector provides per-model/provider usage aggregation for more detailed tracking.\n */\nexport class UsageCollector {\n private summary: UsageSummary;\n private logger = log();\n\n constructor() {\n this.logger.warn('UsageCollector is deprecated. Use ModelUsageCollector instead.');\n this.summary = {\n llmPromptTokens: 0,\n llmPromptCachedTokens: 0,\n llmCompletionTokens: 0,\n ttsCharactersCount: 0,\n sttAudioDurationMs: 0,\n };\n }\n\n collect(metrics: AgentMetrics): void {\n if (metrics.type === 'llm_metrics') {\n this.summary.llmPromptTokens += metrics.promptTokens;\n this.summary.llmPromptCachedTokens += metrics.promptCachedTokens;\n this.summary.llmCompletionTokens += metrics.completionTokens;\n } else if (metrics.type === 'realtime_model_metrics') {\n this.summary.llmPromptTokens += metrics.inputTokens;\n this.summary.llmPromptCachedTokens += metrics.inputTokenDetails.cachedTokens;\n this.summary.llmCompletionTokens += metrics.outputTokens;\n } else if (metrics.type === 'tts_metrics') {\n this.summary.ttsCharactersCount += metrics.charactersCount;\n } else if (metrics.type === 'stt_metrics') {\n this.summary.sttAudioDurationMs += metrics.audioDurationMs;\n }\n }\n\n getSummary(): UsageSummary {\n return { ...this.summary };\n }\n}\n"],"mappings":";;;;;;;;;;;;;;;;;;AAAA;AAAA;AAAA;AAAA;AAAA;AAGA,iBAAoB;AAmBb,MAAM,eAAe;AAAA,EAClB;AAAA,EACA,aAAS,gBAAI;AAAA,EAErB,cAAc;AACZ,SAAK,OAAO,KAAK,gEAAgE;AACjF,SAAK,UAAU;AAAA,MACb,iBAAiB;AAAA,MACjB,uBAAuB;AAAA,MACvB,qBAAqB;AAAA,MACrB,oBAAoB;AAAA,MACpB,oBAAoB;AAAA,IACtB;AAAA,EACF;AAAA,EAEA,QAAQ,SAA6B;AACnC,QAAI,QAAQ,SAAS,eAAe;AAClC,WAAK,QAAQ,mBAAmB,QAAQ;AACxC,WAAK,QAAQ,yBAAyB,QAAQ;AAC9C,WAAK,QAAQ,uBAAuB,QAAQ;AAAA,IAC9C,WAAW,QAAQ,SAAS,0BAA0B;AACpD,WAAK,QAAQ,mBAAmB,QAAQ;AACxC,WAAK,QAAQ,yBAAyB,QAAQ,kBAAkB;AAChE,WAAK,QAAQ,uBAAuB,QAAQ;AAAA,IAC9C,WAAW,QAAQ,SAAS,eAAe;AACzC,WAAK,QAAQ,sBAAsB,QAAQ;AAAA,IAC7C,WAAW,QAAQ,SAAS,eAAe;AACzC,WAAK,QAAQ,sBAAsB,QAAQ;AAAA,IAC7C;AAAA,EACF;AAAA,EAEA,aAA2B;AACzB,WAAO,EAAE,GAAG,KAAK,QAAQ;AAAA,EAC3B;AACF;","names":[]}
|
|
@@ -1,4 +1,8 @@
|
|
|
1
1
|
import type { AgentMetrics } from './base.js';
|
|
2
|
+
/**
|
|
3
|
+
* @deprecated Use LLMModelUsage, TTSModelUsage, or STTModelUsage instead.
|
|
4
|
+
* These new types provide per-model/provider usage aggregation for more detailed tracking.
|
|
5
|
+
*/
|
|
2
6
|
export interface UsageSummary {
|
|
3
7
|
llmPromptTokens: number;
|
|
4
8
|
llmPromptCachedTokens: number;
|
|
@@ -6,8 +10,13 @@ export interface UsageSummary {
|
|
|
6
10
|
ttsCharactersCount: number;
|
|
7
11
|
sttAudioDurationMs: number;
|
|
8
12
|
}
|
|
13
|
+
/**
|
|
14
|
+
* @deprecated Use ModelUsageCollector instead.
|
|
15
|
+
* ModelUsageCollector provides per-model/provider usage aggregation for more detailed tracking.
|
|
16
|
+
*/
|
|
9
17
|
export declare class UsageCollector {
|
|
10
18
|
private summary;
|
|
19
|
+
private logger;
|
|
11
20
|
constructor();
|
|
12
21
|
collect(metrics: AgentMetrics): void;
|
|
13
22
|
getSummary(): UsageSummary;
|
|
@@ -1,4 +1,8 @@
|
|
|
1
1
|
import type { AgentMetrics } from './base.js';
|
|
2
|
+
/**
|
|
3
|
+
* @deprecated Use LLMModelUsage, TTSModelUsage, or STTModelUsage instead.
|
|
4
|
+
* These new types provide per-model/provider usage aggregation for more detailed tracking.
|
|
5
|
+
*/
|
|
2
6
|
export interface UsageSummary {
|
|
3
7
|
llmPromptTokens: number;
|
|
4
8
|
llmPromptCachedTokens: number;
|
|
@@ -6,8 +10,13 @@ export interface UsageSummary {
|
|
|
6
10
|
ttsCharactersCount: number;
|
|
7
11
|
sttAudioDurationMs: number;
|
|
8
12
|
}
|
|
13
|
+
/**
|
|
14
|
+
* @deprecated Use ModelUsageCollector instead.
|
|
15
|
+
* ModelUsageCollector provides per-model/provider usage aggregation for more detailed tracking.
|
|
16
|
+
*/
|
|
9
17
|
export declare class UsageCollector {
|
|
10
18
|
private summary;
|
|
19
|
+
private logger;
|
|
11
20
|
constructor();
|
|
12
21
|
collect(metrics: AgentMetrics): void;
|
|
13
22
|
getSummary(): UsageSummary;
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"usage_collector.d.ts","sourceRoot":"","sources":["../../src/metrics/usage_collector.ts"],"names":[],"mappings":"
|
|
1
|
+
{"version":3,"file":"usage_collector.d.ts","sourceRoot":"","sources":["../../src/metrics/usage_collector.ts"],"names":[],"mappings":"AAIA,OAAO,KAAK,EAAE,YAAY,EAAE,MAAM,WAAW,CAAC;AAE9C;;;GAGG;AACH,MAAM,WAAW,YAAY;IAC3B,eAAe,EAAE,MAAM,CAAC;IACxB,qBAAqB,EAAE,MAAM,CAAC;IAC9B,mBAAmB,EAAE,MAAM,CAAC;IAC5B,kBAAkB,EAAE,MAAM,CAAC;IAC3B,kBAAkB,EAAE,MAAM,CAAC;CAC5B;AAED;;;GAGG;AACH,qBAAa,cAAc;IACzB,OAAO,CAAC,OAAO,CAAe;IAC9B,OAAO,CAAC,MAAM,CAAS;;IAavB,OAAO,CAAC,OAAO,EAAE,YAAY,GAAG,IAAI;IAgBpC,UAAU,IAAI,YAAY;CAG3B"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"sources":["../../src/metrics/usage_collector.ts"],"sourcesContent":["// SPDX-FileCopyrightText: 2024 LiveKit, Inc.\n//\n// SPDX-License-Identifier: Apache-2.0\nimport type { AgentMetrics } from './base.js';\n\nexport interface UsageSummary {\n llmPromptTokens: number;\n llmPromptCachedTokens: number;\n llmCompletionTokens: number;\n ttsCharactersCount: number;\n sttAudioDurationMs: number;\n}\n\nexport class UsageCollector {\n private summary: UsageSummary;\n\n constructor() {\n this.summary = {\n llmPromptTokens: 0,\n llmPromptCachedTokens: 0,\n llmCompletionTokens: 0,\n ttsCharactersCount: 0,\n sttAudioDurationMs: 0,\n };\n }\n\n collect(metrics: AgentMetrics): void {\n if (metrics.type === 'llm_metrics') {\n this.summary.llmPromptTokens += metrics.promptTokens;\n this.summary.llmPromptCachedTokens += metrics.promptCachedTokens;\n this.summary.llmCompletionTokens += metrics.completionTokens;\n } else if (metrics.type === 'realtime_model_metrics') {\n this.summary.llmPromptTokens += metrics.inputTokens;\n this.summary.llmPromptCachedTokens += metrics.inputTokenDetails.cachedTokens;\n this.summary.llmCompletionTokens += metrics.outputTokens;\n } else if (metrics.type === 'tts_metrics') {\n this.summary.ttsCharactersCount += metrics.charactersCount;\n } else if (metrics.type === 'stt_metrics') {\n this.summary.sttAudioDurationMs += metrics.audioDurationMs;\n }\n }\n\n getSummary(): UsageSummary {\n return { ...this.summary };\n }\n}\n"],"mappings":"
|
|
1
|
+
{"version":3,"sources":["../../src/metrics/usage_collector.ts"],"sourcesContent":["// SPDX-FileCopyrightText: 2024 LiveKit, Inc.\n//\n// SPDX-License-Identifier: Apache-2.0\nimport { log } from '../log.js';\nimport type { AgentMetrics } from './base.js';\n\n/**\n * @deprecated Use LLMModelUsage, TTSModelUsage, or STTModelUsage instead.\n * These new types provide per-model/provider usage aggregation for more detailed tracking.\n */\nexport interface UsageSummary {\n llmPromptTokens: number;\n llmPromptCachedTokens: number;\n llmCompletionTokens: number;\n ttsCharactersCount: number;\n sttAudioDurationMs: number;\n}\n\n/**\n * @deprecated Use ModelUsageCollector instead.\n * ModelUsageCollector provides per-model/provider usage aggregation for more detailed tracking.\n */\nexport class UsageCollector {\n private summary: UsageSummary;\n private logger = log();\n\n constructor() {\n this.logger.warn('UsageCollector is deprecated. Use ModelUsageCollector instead.');\n this.summary = {\n llmPromptTokens: 0,\n llmPromptCachedTokens: 0,\n llmCompletionTokens: 0,\n ttsCharactersCount: 0,\n sttAudioDurationMs: 0,\n };\n }\n\n collect(metrics: AgentMetrics): void {\n if (metrics.type === 'llm_metrics') {\n this.summary.llmPromptTokens += metrics.promptTokens;\n this.summary.llmPromptCachedTokens += metrics.promptCachedTokens;\n this.summary.llmCompletionTokens += metrics.completionTokens;\n } else if (metrics.type === 'realtime_model_metrics') {\n this.summary.llmPromptTokens += metrics.inputTokens;\n this.summary.llmPromptCachedTokens += metrics.inputTokenDetails.cachedTokens;\n this.summary.llmCompletionTokens += metrics.outputTokens;\n } else if (metrics.type === 'tts_metrics') {\n this.summary.ttsCharactersCount += metrics.charactersCount;\n } else if (metrics.type === 'stt_metrics') {\n this.summary.sttAudioDurationMs += metrics.audioDurationMs;\n }\n }\n\n getSummary(): UsageSummary {\n return { ...this.summary };\n }\n}\n"],"mappings":"AAGA,SAAS,WAAW;AAmBb,MAAM,eAAe;AAAA,EAClB;AAAA,EACA,SAAS,IAAI;AAAA,EAErB,cAAc;AACZ,SAAK,OAAO,KAAK,gEAAgE;AACjF,SAAK,UAAU;AAAA,MACb,iBAAiB;AAAA,MACjB,uBAAuB;AAAA,MACvB,qBAAqB;AAAA,MACrB,oBAAoB;AAAA,MACpB,oBAAoB;AAAA,IACtB;AAAA,EACF;AAAA,EAEA,QAAQ,SAA6B;AACnC,QAAI,QAAQ,SAAS,eAAe;AAClC,WAAK,QAAQ,mBAAmB,QAAQ;AACxC,WAAK,QAAQ,yBAAyB,QAAQ;AAC9C,WAAK,QAAQ,uBAAuB,QAAQ;AAAA,IAC9C,WAAW,QAAQ,SAAS,0BAA0B;AACpD,WAAK,QAAQ,mBAAmB,QAAQ;AACxC,WAAK,QAAQ,yBAAyB,QAAQ,kBAAkB;AAChE,WAAK,QAAQ,uBAAuB,QAAQ;AAAA,IAC9C,WAAW,QAAQ,SAAS,eAAe;AACzC,WAAK,QAAQ,sBAAsB,QAAQ;AAAA,IAC7C,WAAW,QAAQ,SAAS,eAAe;AACzC,WAAK,QAAQ,sBAAsB,QAAQ;AAAA,IAC7C;AAAA,EACF;AAAA,EAEA,aAA2B;AACzB,WAAO,EAAE,GAAG,KAAK,QAAQ;AAAA,EAC3B;AACF;","names":[]}
|
package/dist/metrics/utils.cjs
CHANGED
|
@@ -65,6 +65,15 @@ const logMetrics = (metrics) => {
|
|
|
65
65
|
logger.child({
|
|
66
66
|
audioDurationMs: Math.round(metrics.audioDurationMs)
|
|
67
67
|
}).info("STT metrics");
|
|
68
|
+
} else if (metrics.type === "interruption_metrics") {
|
|
69
|
+
logger.child({
|
|
70
|
+
totalDurationMs: roundTwoDecimals(metrics.totalDuration),
|
|
71
|
+
predictionDurationMs: roundTwoDecimals(metrics.predictionDuration),
|
|
72
|
+
detectionDelayMs: roundTwoDecimals(metrics.detectionDelay),
|
|
73
|
+
numInterruptions: metrics.numInterruptions,
|
|
74
|
+
numBackchannels: metrics.numBackchannels,
|
|
75
|
+
numRequests: metrics.numRequests
|
|
76
|
+
}).info("Interruption metrics");
|
|
68
77
|
}
|
|
69
78
|
};
|
|
70
79
|
// Annotate the CommonJS export names for ESM import in node:
|