@omnimedia/omnitool 1.1.0-5 → 1.1.0-51
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +143 -80
- package/package.json +18 -16
- package/s/context.ts +0 -7
- package/s/demo/demo.bundle.ts +102 -43
- package/s/demo/demo.css +246 -11
- package/s/demo/routines/export-test.ts +16 -0
- package/s/demo/routines/filmstrip-test.ts +22 -18
- package/s/demo/routines/load-video.ts +2 -3
- package/s/demo/routines/playback-test.ts +76 -0
- package/s/demo/routines/timeline-setup.ts +24 -0
- package/s/demo/routines/transcode-test.ts +21 -5
- package/s/demo/routines/transitions-test.ts +2 -2
- package/s/demo/routines/waveform-test.ts +8 -6
- package/s/driver/driver-worker.ts +9 -0
- package/s/driver/driver.test.ts +1 -1
- package/s/driver/driver.ts +32 -44
- package/s/driver/fns/schematic.ts +52 -25
- package/s/driver/fns/work.ts +71 -200
- package/s/driver/parts/compositor.ts +178 -0
- package/s/driver/parts/machina.ts +19 -20
- package/s/index.html.ts +103 -19
- package/s/index.ts +1 -0
- package/s/tests.bundle.ts +11 -0
- package/s/tests.html.ts +28 -0
- package/s/timeline/index.ts +3 -0
- package/s/timeline/parts/basics.ts +1 -1
- package/s/timeline/parts/filmstrip.ts +43 -16
- package/s/timeline/parts/item.ts +59 -6
- package/s/timeline/parts/media.ts +32 -1
- package/s/timeline/parts/resource-pool.ts +8 -5
- package/s/timeline/parts/resource.ts +2 -0
- package/s/timeline/parts/waveform.ts +3 -4
- package/s/timeline/renderers/export/parts/audio-gain.ts +17 -0
- package/s/timeline/renderers/export/parts/audio-mix.ts +133 -0
- package/s/timeline/renderers/export/parts/cursor.ts +109 -0
- package/s/timeline/renderers/export/parts/produce-audio.ts +64 -0
- package/s/timeline/renderers/export/parts/produce-video.ts +50 -0
- package/s/timeline/renderers/export/parts/resamplers.ts +48 -0
- package/s/timeline/renderers/export/produce.ts +28 -0
- package/s/timeline/renderers/parts/handy.ts +360 -0
- package/s/timeline/renderers/parts/samplers/audio/parts/find.ts +19 -0
- package/s/timeline/renderers/parts/samplers/audio/parts/init.ts +60 -0
- package/s/timeline/renderers/parts/samplers/audio/parts/sink.ts +38 -0
- package/s/timeline/renderers/parts/samplers/audio/parts/types.ts +16 -0
- package/s/timeline/renderers/parts/samplers/audio/sampler.ts +35 -0
- package/s/timeline/renderers/parts/samplers/visual/parts/defaults.ts +16 -0
- package/s/timeline/renderers/parts/samplers/visual/parts/sample.ts +59 -0
- package/s/timeline/renderers/parts/samplers/visual/parts/sequence.ts +111 -0
- package/s/timeline/renderers/parts/samplers/visual/parts/sink.ts +38 -0
- package/s/timeline/renderers/parts/samplers/visual/parts/transition.ts +28 -0
- package/s/timeline/renderers/parts/samplers/visual/parts/types.ts +10 -0
- package/s/timeline/renderers/parts/samplers/visual/sampler.ts +28 -0
- package/s/timeline/renderers/parts/schedulers.ts +96 -0
- package/s/timeline/renderers/player/parts/playback.ts +125 -0
- package/s/timeline/renderers/player/player.ts +71 -0
- package/s/timeline/renderers/renderers.test.ts +387 -0
- package/s/timeline/sugar/helpers.ts +85 -0
- package/s/timeline/sugar/o.ts +178 -38
- package/s/timeline/sugar/omni.test.ts +210 -0
- package/s/timeline/sugar/omni.ts +38 -12
- package/s/timeline/types.ts +29 -0
- package/s/timeline/utils/checksum.ts +3 -1
- package/s/timeline/utils/datafile.ts +15 -4
- package/s/timeline/utils/dummy-data.ts +3 -3
- package/s/timeline/utils/matrix.ts +33 -0
- package/s/units/fps.ts +8 -0
- package/s/units/ms.ts +8 -0
- package/s/units/seconds.ts +8 -0
- package/x/WebGLRenderer-7X274AYV.js +2 -0
- package/x/WebGLRenderer-7X274AYV.js.map +7 -0
- package/x/WebGPURenderer-XMCMEXAO.js +2 -0
- package/x/WebGPURenderer-XMCMEXAO.js.map +7 -0
- package/x/browserAll-6TVTCHHE.js +2 -0
- package/x/browserAll-6TVTCHHE.js.map +7 -0
- package/x/chunk-4ONWQOPX.js +157 -0
- package/x/chunk-4ONWQOPX.js.map +7 -0
- package/x/chunk-63NSCXPX.js +2 -0
- package/x/chunk-63NSCXPX.js.map +7 -0
- package/x/chunk-A45M2HJC.js +2 -0
- package/x/chunk-A45M2HJC.js.map +7 -0
- package/x/chunk-OTQK6FAJ.js +15 -0
- package/x/chunk-OTQK6FAJ.js.map +7 -0
- package/x/chunk-Q7JBQNE4.js +42 -0
- package/x/chunk-Q7JBQNE4.js.map +7 -0
- package/x/chunk-W33LM336.js +393 -0
- package/x/chunk-W33LM336.js.map +7 -0
- package/x/chunk-W5CN46AR.js +327 -0
- package/x/chunk-W5CN46AR.js.map +7 -0
- package/x/chunk-WFT3KTZG.js +269 -0
- package/x/chunk-WFT3KTZG.js.map +7 -0
- package/x/context.d.ts +1 -4
- package/x/context.js +1 -5
- package/x/context.js.map +1 -1
- package/x/demo/WebGLRenderer-NLGJGAXK.js +2 -0
- package/x/demo/WebGLRenderer-NLGJGAXK.js.map +7 -0
- package/x/demo/WebGPURenderer-RBOFXPL5.js +2 -0
- package/x/demo/WebGPURenderer-RBOFXPL5.js.map +7 -0
- package/x/demo/browserAll-5AZHDDG6.js +2 -0
- package/x/demo/browserAll-5AZHDDG6.js.map +7 -0
- package/x/demo/chunk-5ZZYIILO.js +393 -0
- package/x/demo/chunk-5ZZYIILO.js.map +7 -0
- package/x/demo/chunk-P3PTHHFE.js +42 -0
- package/x/demo/chunk-P3PTHHFE.js.map +7 -0
- package/x/demo/chunk-PYG4RZZ2.js +269 -0
- package/x/demo/chunk-PYG4RZZ2.js.map +7 -0
- package/x/demo/chunk-Q4MWBZHL.js +157 -0
- package/x/demo/chunk-Q4MWBZHL.js.map +7 -0
- package/x/demo/chunk-T3METYEY.js +15 -0
- package/x/demo/chunk-T3METYEY.js.map +7 -0
- package/x/demo/chunk-USLRKDKD.js +2 -0
- package/x/demo/chunk-USLRKDKD.js.map +7 -0
- package/x/demo/chunk-YISSXWBT.js +327 -0
- package/x/demo/chunk-YISSXWBT.js.map +7 -0
- package/x/demo/chunk-YJQWVIHX.js +2 -0
- package/x/demo/chunk-YJQWVIHX.js.map +7 -0
- package/x/demo/demo.bundle.js +92 -40
- package/x/demo/demo.bundle.js.map +1 -1
- package/x/demo/demo.bundle.min.js +2450 -35
- package/x/demo/demo.bundle.min.js.map +4 -4
- package/x/demo/demo.css +246 -11
- package/x/demo/routines/export-test.d.ts +2 -0
- package/x/demo/routines/export-test.js +10 -0
- package/x/demo/routines/export-test.js.map +1 -0
- package/x/demo/routines/filmstrip-test.d.ts +1 -1
- package/x/demo/routines/filmstrip-test.js +20 -17
- package/x/demo/routines/filmstrip-test.js.map +1 -1
- package/x/demo/routines/load-video.d.ts +1 -1
- package/x/demo/routines/load-video.js +1 -2
- package/x/demo/routines/load-video.js.map +1 -1
- package/x/demo/routines/playback-test.d.ts +2 -0
- package/x/demo/routines/playback-test.js +65 -0
- package/x/demo/routines/playback-test.js.map +1 -0
- package/x/demo/routines/timeline-setup.d.ts +6 -0
- package/x/demo/routines/timeline-setup.js +13 -0
- package/x/demo/routines/timeline-setup.js.map +1 -0
- package/x/demo/routines/transcode-test.js +17 -5
- package/x/demo/routines/transcode-test.js.map +1 -1
- package/x/demo/routines/transitions-test.js +2 -2
- package/x/demo/routines/transitions-test.js.map +1 -1
- package/x/demo/routines/waveform-test.d.ts +2 -1
- package/x/demo/routines/waveform-test.js +7 -6
- package/x/demo/routines/waveform-test.js.map +1 -1
- package/x/demo/webworkerAll-QKIC5O27.js +2 -0
- package/x/demo/webworkerAll-QKIC5O27.js.map +7 -0
- package/x/driver/WebGLRenderer-OMRWYQIV.js +2 -0
- package/x/driver/WebGLRenderer-OMRWYQIV.js.map +7 -0
- package/x/driver/WebGPURenderer-KQJB2OJJ.js +2 -0
- package/x/driver/WebGPURenderer-KQJB2OJJ.js.map +7 -0
- package/x/driver/browserAll-YBZEJCN3.js +2 -0
- package/x/driver/browserAll-YBZEJCN3.js.map +7 -0
- package/x/driver/chunk-3L3MB5NY.js +393 -0
- package/x/driver/chunk-3L3MB5NY.js.map +7 -0
- package/x/driver/chunk-42BQ4XKE.js +269 -0
- package/x/driver/chunk-42BQ4XKE.js.map +7 -0
- package/x/driver/chunk-4HAYG3N5.js +327 -0
- package/x/driver/chunk-4HAYG3N5.js.map +7 -0
- package/x/driver/chunk-BFBY7VYB.js +42 -0
- package/x/driver/chunk-BFBY7VYB.js.map +7 -0
- package/x/driver/chunk-KM6O72WE.js +157 -0
- package/x/driver/chunk-KM6O72WE.js.map +7 -0
- package/x/driver/chunk-N6HD4WYJ.js +2 -0
- package/x/driver/chunk-N6HD4WYJ.js.map +7 -0
- package/x/driver/chunk-WCZ2O3UN.js +15 -0
- package/x/driver/chunk-WCZ2O3UN.js.map +7 -0
- package/x/driver/chunk-XWNSF3WJ.js +2 -0
- package/x/driver/chunk-XWNSF3WJ.js.map +7 -0
- package/x/driver/driver-worker.d.ts +1 -0
- package/x/driver/driver-worker.js +6 -0
- package/x/driver/driver-worker.js.map +1 -0
- package/x/driver/driver.d.ts +12 -9
- package/x/driver/driver.js +31 -41
- package/x/driver/driver.js.map +1 -1
- package/x/driver/driver.test.js +1 -1
- package/x/driver/driver.test.js.map +1 -1
- package/x/driver/driver.worker.bundle.min.js +139 -1135
- package/x/driver/driver.worker.bundle.min.js.map +4 -4
- package/x/driver/fns/host.d.ts +10 -4
- package/x/driver/fns/schematic.d.ts +47 -24
- package/x/driver/fns/work.d.ts +12 -7
- package/x/driver/fns/work.js +68 -170
- package/x/driver/fns/work.js.map +1 -1
- package/x/driver/parts/compositor.d.ts +15 -0
- package/x/driver/parts/compositor.js +152 -0
- package/x/driver/parts/compositor.js.map +1 -0
- package/x/driver/parts/machina.d.ts +0 -20
- package/x/driver/parts/machina.js +6 -10
- package/x/driver/parts/machina.js.map +1 -1
- package/x/driver/webworkerAll-BKJQW6P7.js +2 -0
- package/x/driver/webworkerAll-BKJQW6P7.js.map +7 -0
- package/x/features/speech/transcribe/parts/prep-audio.d.ts +1 -1
- package/x/features/speech/transcribe/worker.bundle.min.js +703 -703
- package/x/features/speech/transcribe/worker.bundle.min.js.map +4 -4
- package/x/index.d.ts +1 -0
- package/x/index.html +349 -30
- package/x/index.html.js +103 -19
- package/x/index.html.js.map +1 -1
- package/x/index.js +1 -0
- package/x/index.js.map +1 -1
- package/x/tests.bundle.js +8 -0
- package/x/tests.bundle.js.map +1 -0
- package/x/tests.bundle.min.js +2464 -0
- package/x/tests.bundle.min.js.map +7 -0
- package/x/tests.html +316 -0
- package/x/tests.html.d.ts +2 -0
- package/x/tests.html.js +22 -0
- package/x/tests.html.js.map +1 -0
- package/x/timeline/index.d.ts +3 -0
- package/x/timeline/index.js +3 -0
- package/x/timeline/index.js.map +1 -1
- package/x/timeline/parts/basics.d.ts +1 -1
- package/x/timeline/parts/filmstrip.d.ts +4 -3
- package/x/timeline/parts/filmstrip.js +29 -10
- package/x/timeline/parts/filmstrip.js.map +1 -1
- package/x/timeline/parts/item.d.ts +51 -8
- package/x/timeline/parts/item.js +7 -3
- package/x/timeline/parts/item.js.map +1 -1
- package/x/timeline/parts/media.d.ts +5 -0
- package/x/timeline/parts/media.js +27 -1
- package/x/timeline/parts/media.js.map +1 -1
- package/x/timeline/parts/resource-pool.d.ts +3 -0
- package/x/timeline/parts/resource-pool.js +7 -4
- package/x/timeline/parts/resource-pool.js.map +1 -1
- package/x/timeline/parts/resource.d.ts +2 -0
- package/x/timeline/parts/waveform.d.ts +2 -1
- package/x/timeline/parts/waveform.js +2 -4
- package/x/timeline/parts/waveform.js.map +1 -1
- package/x/timeline/renderers/export/parts/audio-gain.d.ts +1 -0
- package/x/timeline/renderers/export/parts/audio-gain.js +13 -0
- package/x/timeline/renderers/export/parts/audio-gain.js.map +1 -0
- package/x/timeline/renderers/export/parts/audio-mix.d.ts +21 -0
- package/x/timeline/renderers/export/parts/audio-mix.js +89 -0
- package/x/timeline/renderers/export/parts/audio-mix.js.map +1 -0
- package/x/timeline/renderers/export/parts/cursor.d.ts +19 -0
- package/x/timeline/renderers/export/parts/cursor.js +80 -0
- package/x/timeline/renderers/export/parts/cursor.js.map +1 -0
- package/x/timeline/renderers/export/parts/produce-audio.d.ts +6 -0
- package/x/timeline/renderers/export/parts/produce-audio.js +41 -0
- package/x/timeline/renderers/export/parts/produce-audio.js.map +1 -0
- package/x/timeline/renderers/export/parts/produce-video.d.ts +10 -0
- package/x/timeline/renderers/export/parts/produce-video.js +27 -0
- package/x/timeline/renderers/export/parts/produce-video.js.map +1 -0
- package/x/timeline/renderers/export/parts/resamplers.d.ts +12 -0
- package/x/timeline/renderers/export/parts/resamplers.js +29 -0
- package/x/timeline/renderers/export/parts/resamplers.js.map +1 -0
- package/x/timeline/renderers/export/produce.d.ts +13 -0
- package/x/timeline/renderers/export/produce.js +15 -0
- package/x/timeline/renderers/export/produce.js.map +1 -0
- package/x/timeline/renderers/parts/handy.d.ts +30 -0
- package/x/timeline/renderers/parts/handy.js +219 -0
- package/x/timeline/renderers/parts/handy.js.map +1 -0
- package/x/timeline/renderers/parts/samplers/audio/parts/find.d.ts +6 -0
- package/x/timeline/renderers/parts/samplers/audio/parts/find.js +15 -0
- package/x/timeline/renderers/parts/samplers/audio/parts/find.js.map +1 -0
- package/x/timeline/renderers/parts/samplers/audio/parts/init.d.ts +5 -0
- package/x/timeline/renderers/parts/samplers/audio/parts/init.js +40 -0
- package/x/timeline/renderers/parts/samplers/audio/parts/init.js.map +1 -0
- package/x/timeline/renderers/parts/samplers/audio/parts/sink.d.ts +8 -0
- package/x/timeline/renderers/parts/samplers/audio/parts/sink.js +24 -0
- package/x/timeline/renderers/parts/samplers/audio/parts/sink.js.map +1 -0
- package/x/timeline/renderers/parts/samplers/audio/parts/types.d.ts +14 -0
- package/x/timeline/renderers/parts/samplers/audio/parts/types.js.map +1 -0
- package/x/timeline/renderers/parts/samplers/audio/sampler.d.ts +11 -0
- package/x/timeline/renderers/parts/samplers/audio/sampler.js +22 -0
- package/x/timeline/renderers/parts/samplers/audio/sampler.js.map +1 -0
- package/x/timeline/renderers/parts/samplers/visual/parts/defaults.d.ts +5 -0
- package/x/timeline/renderers/parts/samplers/visual/parts/defaults.js +10 -0
- package/x/timeline/renderers/parts/samplers/visual/parts/defaults.js.map +1 -0
- package/x/timeline/renderers/parts/samplers/visual/parts/sample.d.ts +5 -0
- package/x/timeline/renderers/parts/samplers/visual/parts/sample.js +38 -0
- package/x/timeline/renderers/parts/samplers/visual/parts/sample.js.map +1 -0
- package/x/timeline/renderers/parts/samplers/visual/parts/sequence.d.ts +5 -0
- package/x/timeline/renderers/parts/samplers/visual/parts/sequence.js +75 -0
- package/x/timeline/renderers/parts/samplers/visual/parts/sequence.js.map +1 -0
- package/x/timeline/renderers/parts/samplers/visual/parts/sink.d.ts +8 -0
- package/x/timeline/renderers/parts/samplers/visual/parts/sink.js +24 -0
- package/x/timeline/renderers/parts/samplers/visual/parts/sink.js.map +1 -0
- package/x/timeline/renderers/parts/samplers/visual/parts/transition.d.ts +3 -0
- package/x/timeline/renderers/parts/samplers/visual/parts/transition.js +18 -0
- package/x/timeline/renderers/parts/samplers/visual/parts/transition.js.map +1 -0
- package/x/timeline/renderers/parts/samplers/visual/parts/types.d.ts +8 -0
- package/x/timeline/renderers/parts/samplers/visual/parts/types.js +2 -0
- package/x/timeline/renderers/parts/samplers/visual/parts/types.js.map +1 -0
- package/x/timeline/renderers/parts/samplers/visual/sampler.d.ts +7 -0
- package/x/timeline/renderers/parts/samplers/visual/sampler.js +17 -0
- package/x/timeline/renderers/parts/samplers/visual/sampler.js.map +1 -0
- package/x/timeline/renderers/parts/schedulers.d.ts +17 -0
- package/x/timeline/renderers/parts/schedulers.js +64 -0
- package/x/timeline/renderers/parts/schedulers.js.map +1 -0
- package/x/timeline/renderers/player/parts/playback.d.ts +31 -0
- package/x/timeline/renderers/player/parts/playback.js +91 -0
- package/x/timeline/renderers/player/parts/playback.js.map +1 -0
- package/x/timeline/renderers/player/player.d.ts +23 -0
- package/x/timeline/renderers/player/player.js +50 -0
- package/x/timeline/renderers/player/player.js.map +1 -0
- package/x/timeline/renderers/renderers.test.d.ts +32 -0
- package/x/timeline/renderers/renderers.test.js +306 -0
- package/x/timeline/renderers/renderers.test.js.map +1 -0
- package/x/timeline/sugar/helpers.d.ts +30 -0
- package/x/timeline/sugar/helpers.js +46 -0
- package/x/timeline/sugar/helpers.js.map +1 -0
- package/x/timeline/sugar/o.d.ts +31 -4
- package/x/timeline/sugar/o.js +144 -38
- package/x/timeline/sugar/o.js.map +1 -1
- package/x/timeline/sugar/omni.d.ts +11 -2
- package/x/timeline/sugar/omni.js +30 -9
- package/x/timeline/sugar/omni.js.map +1 -1
- package/x/timeline/sugar/omni.test.d.ts +27 -0
- package/x/timeline/sugar/omni.test.js +128 -0
- package/x/timeline/sugar/omni.test.js.map +1 -0
- package/x/timeline/types.d.ts +24 -0
- package/x/timeline/types.js +2 -0
- package/x/timeline/types.js.map +1 -0
- package/x/timeline/utils/checksum.d.ts +3 -2
- package/x/timeline/utils/checksum.js.map +1 -1
- package/x/timeline/utils/datafile.d.ts +5 -3
- package/x/timeline/utils/datafile.js +18 -5
- package/x/timeline/utils/datafile.js.map +1 -1
- package/x/timeline/utils/dummy-data.d.ts +1 -2
- package/x/timeline/utils/dummy-data.js +4 -2
- package/x/timeline/utils/dummy-data.js.map +1 -1
- package/x/timeline/utils/matrix.d.ts +8 -0
- package/x/timeline/utils/matrix.js +26 -0
- package/x/timeline/utils/matrix.js.map +1 -0
- package/x/units/fps.d.ts +6 -0
- package/x/units/fps.js +2 -0
- package/x/units/fps.js.map +1 -0
- package/x/units/ms.d.ts +6 -0
- package/x/units/ms.js +2 -0
- package/x/units/ms.js.map +1 -0
- package/x/units/seconds.d.ts +6 -0
- package/x/units/seconds.js +2 -0
- package/x/units/seconds.js.map +1 -0
- package/x/webworkerAll-VVIU3M54.js +2 -0
- package/x/webworkerAll-VVIU3M54.js.map +7 -0
- package/s/tests.test.ts +0 -8
- package/s/timeline/sugar/omni-test.ts +0 -38
- package/s/tools/common/loader.ts +0 -26
- package/s/tools/common/transformer-pipeline.ts +0 -26
- package/s/tools/speech-recognition/common/model.ts +0 -26
- package/s/tools/speech-recognition/whisper/fns/host.ts +0 -25
- package/s/tools/speech-recognition/whisper/fns/schematic.ts +0 -23
- package/s/tools/speech-recognition/whisper/fns/work.ts +0 -91
- package/s/tools/speech-recognition/whisper/parts/types.ts +0 -38
- package/s/tools/speech-recognition/whisper/parts/worker.bundle.ts +0 -7
- package/s/tools/speech-recognition/whisper/tool.ts +0 -70
- package/x/tests.test.js +0 -6
- package/x/tests.test.js.map +0 -1
- package/x/timeline/sugar/omni-test.d.ts +0 -1
- package/x/timeline/sugar/omni-test.js +0 -22
- package/x/timeline/sugar/omni-test.js.map +0 -1
- package/x/tools/common/loader.d.ts +0 -19
- package/x/tools/common/loader.js +0 -18
- package/x/tools/common/loader.js.map +0 -1
- package/x/tools/common/transformer-pipeline.d.ts +0 -8
- package/x/tools/common/transformer-pipeline.js +0 -24
- package/x/tools/common/transformer-pipeline.js.map +0 -1
- package/x/tools/speech-recognition/common/model.d.ts +0 -14
- package/x/tools/speech-recognition/common/model.js +0 -16
- package/x/tools/speech-recognition/common/model.js.map +0 -1
- package/x/tools/speech-recognition/whisper/fns/host.d.ts +0 -13
- package/x/tools/speech-recognition/whisper/fns/host.js +0 -19
- package/x/tools/speech-recognition/whisper/fns/host.js.map +0 -1
- package/x/tools/speech-recognition/whisper/fns/schematic.d.ts +0 -19
- package/x/tools/speech-recognition/whisper/fns/schematic.js +0 -2
- package/x/tools/speech-recognition/whisper/fns/schematic.js.map +0 -1
- package/x/tools/speech-recognition/whisper/fns/work.d.ts +0 -12
- package/x/tools/speech-recognition/whisper/fns/work.js +0 -74
- package/x/tools/speech-recognition/whisper/fns/work.js.map +0 -1
- package/x/tools/speech-recognition/whisper/parts/types.d.ts +0 -31
- package/x/tools/speech-recognition/whisper/parts/types.js.map +0 -1
- package/x/tools/speech-recognition/whisper/parts/worker.bundle.d.ts +0 -1
- package/x/tools/speech-recognition/whisper/parts/worker.bundle.js +0 -4
- package/x/tools/speech-recognition/whisper/parts/worker.bundle.js.map +0 -1
- package/x/tools/speech-recognition/whisper/parts/worker.bundle.min.js +0 -8
- package/x/tools/speech-recognition/whisper/parts/worker.bundle.min.js.map +0 -7
- package/x/tools/speech-recognition/whisper/tool.d.ts +0 -12
- package/x/tools/speech-recognition/whisper/tool.js +0 -63
- package/x/tools/speech-recognition/whisper/tool.js.map +0 -1
- /package/x/{tests.test.d.ts → tests.bundle.d.ts} +0 -0
- /package/x/{tools/speech-recognition/whisper → timeline/renderers/parts/samplers/audio}/parts/types.js +0 -0
|
@@ -1,26 +0,0 @@
|
|
|
1
|
-
import {pub} from "@e280/stz"
|
|
2
|
-
|
|
3
|
-
import {Loader} from "../../common/loader.js"
|
|
4
|
-
import {DecoderSource} from "../../../driver/fns/schematic.js"
|
|
5
|
-
import {SpeechRecognizerModels, Word, WordGroup} from "../whisper/parts/types.js"
|
|
6
|
-
|
|
7
|
-
export abstract class SpeechRecognizer extends Loader {
|
|
8
|
-
multilingual = true
|
|
9
|
-
|
|
10
|
-
static speechRecognizerEvents = {
|
|
11
|
-
onTranscriptionChunk: pub<Word[]>(),
|
|
12
|
-
onTranscribeProgress: pub<[number]>()
|
|
13
|
-
}
|
|
14
|
-
|
|
15
|
-
abstract transcribe(input: DecoderSource): Promise<WordGroup>
|
|
16
|
-
|
|
17
|
-
setMultilingual(value: boolean) {
|
|
18
|
-
this.multilingual = value
|
|
19
|
-
}
|
|
20
|
-
|
|
21
|
-
detectLanguage?(input: Blob | AudioBuffer): Promise<string>
|
|
22
|
-
|
|
23
|
-
setModel(value: SpeechRecognizerModels) {
|
|
24
|
-
this.model = value
|
|
25
|
-
}
|
|
26
|
-
}
|
|
@@ -1,25 +0,0 @@
|
|
|
1
|
-
|
|
2
|
-
import {Comrade} from "@e280/comrade"
|
|
3
|
-
import {ProgressItem} from "../parts/types.js"
|
|
4
|
-
import {SpeechRecognizerHostEvents, WhisperSchematic} from "./schematic.js"
|
|
5
|
-
|
|
6
|
-
export const setupWhisperHost = (events: SpeechRecognizerHostEvents) => (
|
|
7
|
-
Comrade.host<WhisperSchematic>(_shell => ({
|
|
8
|
-
async updateModelLoadProgress(item) {
|
|
9
|
-
events.onModelLoadProgress.pub(item)
|
|
10
|
-
},
|
|
11
|
-
async deliverTranscriptionChunk(chunk) {
|
|
12
|
-
events.onTranscriptionChunk.pub({
|
|
13
|
-
text: chunk.text,
|
|
14
|
-
timestamp: chunk.timestamp
|
|
15
|
-
})
|
|
16
|
-
},
|
|
17
|
-
async updateTps(value) {
|
|
18
|
-
events.onTpsUpdate.pub(value)
|
|
19
|
-
},
|
|
20
|
-
async updateTranscribeProgress(value) {
|
|
21
|
-
events.onTranscribeProgress(value)
|
|
22
|
-
}
|
|
23
|
-
}))
|
|
24
|
-
)
|
|
25
|
-
|
|
@@ -1,23 +0,0 @@
|
|
|
1
|
-
import {Pub} from "@e280/stz"
|
|
2
|
-
import {AsSchematic} from "@e280/comrade"
|
|
3
|
-
|
|
4
|
-
import {LoaderEvents} from "../../../common/loader.js"
|
|
5
|
-
import {ProgressItem, TranscriptionChunk, TranscriptionMessage, TranscriptionResult, Word} from "../parts/types.js"
|
|
6
|
-
|
|
7
|
-
export type WhisperSchematic = AsSchematic<{
|
|
8
|
-
work: {
|
|
9
|
-
transcribe(input: TranscriptionMessage): Promise<TranscriptionResult | null>
|
|
10
|
-
},
|
|
11
|
-
|
|
12
|
-
host: {
|
|
13
|
-
updateModelLoadProgress(item: ProgressItem): Promise<void>
|
|
14
|
-
deliverTranscriptionChunk(chunk: TranscriptionChunk): Promise<void>
|
|
15
|
-
updateTps(value: number): Promise<void>
|
|
16
|
-
updateTranscribeProgress(value: number): Promise<void>
|
|
17
|
-
}
|
|
18
|
-
}>
|
|
19
|
-
|
|
20
|
-
export interface SpeechRecognizerHostEvents extends LoaderEvents {
|
|
21
|
-
onTranscriptionChunk: Pub<Word[]>
|
|
22
|
-
onTranscribeProgress: Pub<[number]>
|
|
23
|
-
}
|
|
@@ -1,91 +0,0 @@
|
|
|
1
|
-
import {Comrade} from "@e280/comrade"
|
|
2
|
-
//@ts-ignore
|
|
3
|
-
import {pipeline, WhisperTextStreamer} from "https://cdn.jsdelivr.net/npm/@huggingface/transformers@3.7.0/dist/transformers.min.js"
|
|
4
|
-
|
|
5
|
-
import {WhisperSchematic} from "./schematic.js"
|
|
6
|
-
import {TranscriptionChunk} from "../parts/types.js"
|
|
7
|
-
import {PipelineFactory} from "../../../common/transformer-pipeline.js"
|
|
8
|
-
|
|
9
|
-
// TODO suspicious globals, probably bad
|
|
10
|
-
const pipeline = new PipelineFactory("automatic-speech-recognition")
|
|
11
|
-
let transcriber: any
|
|
12
|
-
|
|
13
|
-
export const setupWhisperWork = Comrade.work<WhisperSchematic>(shell => ({
|
|
14
|
-
async transcribe({audio, model, language, duration}) {
|
|
15
|
-
const isDistil = model.startsWith("distil-whisper/")
|
|
16
|
-
|
|
17
|
-
if(!pipeline.model || pipeline.model !== model) {
|
|
18
|
-
pipeline.instance?.dispose()?.()
|
|
19
|
-
pipeline.instance = null
|
|
20
|
-
transcriber = await pipeline.createInstance(
|
|
21
|
-
model,
|
|
22
|
-
(data) => {
|
|
23
|
-
if(data.progress)
|
|
24
|
-
shell.host.updateModelLoadProgress({
|
|
25
|
-
id: data.file,
|
|
26
|
-
progress: data.progress
|
|
27
|
-
})
|
|
28
|
-
}
|
|
29
|
-
)
|
|
30
|
-
}
|
|
31
|
-
|
|
32
|
-
const timePrecision =
|
|
33
|
-
transcriber.processor.feature_extractor.config.chunk_length /
|
|
34
|
-
transcriber.model.config.max_source_positions
|
|
35
|
-
|
|
36
|
-
const chunkLength = isDistil ? 20 : 30
|
|
37
|
-
const strideLength = isDistil ? 3 : 5
|
|
38
|
-
|
|
39
|
-
let chunkCount = 0
|
|
40
|
-
let startTime: number | null = null
|
|
41
|
-
let tokenCount = 0
|
|
42
|
-
let tps = 0
|
|
43
|
-
|
|
44
|
-
const chunkDuration = chunkLength - strideLength
|
|
45
|
-
|
|
46
|
-
const estimateProgress = () => {
|
|
47
|
-
const audioProgressSeconds = chunkCount * chunkDuration
|
|
48
|
-
return Math.min(audioProgressSeconds / duration, 1)
|
|
49
|
-
}
|
|
50
|
-
|
|
51
|
-
const streamer = new WhisperTextStreamer(transcriber.tokenizer, {
|
|
52
|
-
time_precision: timePrecision,
|
|
53
|
-
token_callback_function: () => {
|
|
54
|
-
startTime ??= performance.now()
|
|
55
|
-
if (++tokenCount > 1) {
|
|
56
|
-
tps = (tokenCount / (performance.now() - startTime)) * 1000
|
|
57
|
-
shell.host.updateTps(tps)
|
|
58
|
-
}
|
|
59
|
-
},
|
|
60
|
-
callback_function: (textChunk: any) => {
|
|
61
|
-
shell.host.deliverTranscriptionChunk(textChunk)
|
|
62
|
-
},
|
|
63
|
-
on_finalize: () => {
|
|
64
|
-
startTime = null
|
|
65
|
-
tokenCount = 0
|
|
66
|
-
chunkCount++
|
|
67
|
-
const progress = estimateProgress()
|
|
68
|
-
shell.host.updateTranscribeProgress(progress)
|
|
69
|
-
},
|
|
70
|
-
})
|
|
71
|
-
|
|
72
|
-
const output = await transcriber(audio, {
|
|
73
|
-
top_k: 0,
|
|
74
|
-
do_sample: false,
|
|
75
|
-
chunk_length_s: chunkLength,
|
|
76
|
-
stride_length_s: strideLength,
|
|
77
|
-
language,
|
|
78
|
-
task: "transcribe",
|
|
79
|
-
return_timestamps: "word", // if using "word" the on_chunk_start & end is not called thus we cant retrieve timestamps, only after whole thing finishes
|
|
80
|
-
force_full_sequences: false,
|
|
81
|
-
streamer,
|
|
82
|
-
})
|
|
83
|
-
|
|
84
|
-
if (!output) return null
|
|
85
|
-
|
|
86
|
-
return {
|
|
87
|
-
tps,
|
|
88
|
-
...output,
|
|
89
|
-
}
|
|
90
|
-
}
|
|
91
|
-
}))
|
|
@@ -1,38 +0,0 @@
|
|
|
1
|
-
export interface ProgressItem {
|
|
2
|
-
id: string
|
|
3
|
-
progress: number
|
|
4
|
-
}
|
|
5
|
-
|
|
6
|
-
export type Word = {
|
|
7
|
-
text: string
|
|
8
|
-
timestamp: [start: number, end: number]
|
|
9
|
-
}
|
|
10
|
-
|
|
11
|
-
export type WordGroup = Word[]
|
|
12
|
-
export type Transcript = WordGroup[]
|
|
13
|
-
|
|
14
|
-
export interface TranscriptionChunk {
|
|
15
|
-
text: string
|
|
16
|
-
offset: number
|
|
17
|
-
timestamp: [number, number]
|
|
18
|
-
finalised: boolean
|
|
19
|
-
}
|
|
20
|
-
|
|
21
|
-
export interface TranscriptionMessage {
|
|
22
|
-
audio: Float32Array
|
|
23
|
-
model: string
|
|
24
|
-
subtask: string | null
|
|
25
|
-
language: string | null
|
|
26
|
-
duration: number
|
|
27
|
-
}
|
|
28
|
-
|
|
29
|
-
export interface TranscriptionResult {
|
|
30
|
-
text: string
|
|
31
|
-
chunks: TranscriptionChunk[]
|
|
32
|
-
tps: number
|
|
33
|
-
}
|
|
34
|
-
|
|
35
|
-
export type ProgressCallback = (data: any) => void
|
|
36
|
-
|
|
37
|
-
export type SpeechRecognizerModels = "onnx-community/whisper-tiny_timestamped"
|
|
38
|
-
export type SpeechRecognizerSubtasks = "transcribe"
|
|
@@ -1,70 +0,0 @@
|
|
|
1
|
-
import {Comrade, Thread} from "@e280/comrade"
|
|
2
|
-
|
|
3
|
-
import {WordGroup} from "./parts/types.js"
|
|
4
|
-
import {context} from "../../../context.js"
|
|
5
|
-
import {setupWhisperHost} from "./fns/host.js"
|
|
6
|
-
import {SpeechRecognizer} from "../common/model.js"
|
|
7
|
-
import {WhisperSchematic} from "./fns/schematic.js"
|
|
8
|
-
|
|
9
|
-
export class Whisper extends SpeechRecognizer {
|
|
10
|
-
constructor(public thread: Thread<WhisperSchematic>) {
|
|
11
|
-
super('whisper', "onnx-community/whisper-tiny_timestamped")
|
|
12
|
-
}
|
|
13
|
-
|
|
14
|
-
static async setup() {
|
|
15
|
-
const thread = await Comrade.thread<WhisperSchematic>({
|
|
16
|
-
label: "OmnitoolDriver",
|
|
17
|
-
workerUrl: new URL("/tools/speech-recognition/whisper/parts/worker.bundle.min.js", import.meta.url),
|
|
18
|
-
setupHost: setupWhisperHost({
|
|
19
|
-
...this.loaderEvents,
|
|
20
|
-
...this.speechRecognizerEvents
|
|
21
|
-
})
|
|
22
|
-
})
|
|
23
|
-
return new this(thread)
|
|
24
|
-
}
|
|
25
|
-
|
|
26
|
-
async init() {
|
|
27
|
-
// there should be called loading of the model in worker instead when transcribe is called ..
|
|
28
|
-
}
|
|
29
|
-
|
|
30
|
-
async #transcribe(source: Blob, options?: {multilingual?: boolean, language?: string}) {
|
|
31
|
-
const arrayBuffer = await source.arrayBuffer()
|
|
32
|
-
const audioCTX = new AudioContext({sampleRate: 16000})
|
|
33
|
-
const audioData = await audioCTX.decodeAudioData(arrayBuffer)
|
|
34
|
-
let audio
|
|
35
|
-
if (audioData.numberOfChannels === 2) {
|
|
36
|
-
const SCALING_FACTOR = Math.sqrt(2)
|
|
37
|
-
const left = audioData.getChannelData(0)
|
|
38
|
-
const right = audioData.getChannelData(1)
|
|
39
|
-
audio = new Float32Array(left.length)
|
|
40
|
-
for (let i = 0; i < audioData.length; ++i) {
|
|
41
|
-
audio[i] = (SCALING_FACTOR * (left[i] + right[i])) / 2
|
|
42
|
-
}
|
|
43
|
-
} else {
|
|
44
|
-
audio = audioData.getChannelData(0)
|
|
45
|
-
}
|
|
46
|
-
const driver = await context.driver
|
|
47
|
-
const duration = await driver.getAudioDuration(source)
|
|
48
|
-
return await this.thread.work.transcribe({
|
|
49
|
-
audio,
|
|
50
|
-
duration,
|
|
51
|
-
model: this.model,
|
|
52
|
-
subtask: this.multilingual ? "transcribe" : null,
|
|
53
|
-
language:
|
|
54
|
-
this.multilingual && options?.language !== "auto"
|
|
55
|
-
? options?.language ?? "english"
|
|
56
|
-
: null
|
|
57
|
-
})
|
|
58
|
-
}
|
|
59
|
-
|
|
60
|
-
async transcribe(source: Blob): Promise<WordGroup> {
|
|
61
|
-
const result = await this.#transcribe(source)
|
|
62
|
-
|
|
63
|
-
const words = result?.chunks.map((chunk: any) => ({
|
|
64
|
-
text: chunk.text.trim(),
|
|
65
|
-
timestamp: chunk.timestamp,
|
|
66
|
-
})) as WordGroup
|
|
67
|
-
|
|
68
|
-
return words
|
|
69
|
-
}
|
|
70
|
-
}
|
package/x/tests.test.js
DELETED
package/x/tests.test.js.map
DELETED
|
@@ -1 +0,0 @@
|
|
|
1
|
-
{"version":3,"file":"tests.test.js","sourceRoot":"","sources":["../s/tests.test.ts"],"names":[],"mappings":"AACA,OAAO,EAAC,OAAO,EAAC,MAAM,eAAe,CAAA;AACrC,OAAO,MAAM,MAAM,yBAAyB,CAAA;AAE5C,MAAM,OAAO,CAAC,GAAG,CAAC;IACjB,MAAM;CACN,CAAC,CAAA"}
|
|
@@ -1 +0,0 @@
|
|
|
1
|
-
export {};
|
|
@@ -1,22 +0,0 @@
|
|
|
1
|
-
import { Omni } from "./omni.js";
|
|
2
|
-
import { dummyData } from "../utils/dummy-data.js";
|
|
3
|
-
//
|
|
4
|
-
// create an omni context
|
|
5
|
-
//
|
|
6
|
-
const omni = new Omni();
|
|
7
|
-
//
|
|
8
|
-
// load in some media resources
|
|
9
|
-
//
|
|
10
|
-
const { mediaA, mediaB } = await omni.load({
|
|
11
|
-
mediaA: dummyData(),
|
|
12
|
-
mediaB: dummyData(),
|
|
13
|
-
});
|
|
14
|
-
//
|
|
15
|
-
// create a timeline
|
|
16
|
-
//
|
|
17
|
-
const timeline = omni.timeline(o => o.sequence(o.clip(mediaA), o.transition.crossfade(600), o.stack(o.clip(mediaB), o.text("hello world"))));
|
|
18
|
-
//
|
|
19
|
-
// log the timeline
|
|
20
|
-
//
|
|
21
|
-
console.log(JSON.stringify(timeline, undefined, " "));
|
|
22
|
-
//# sourceMappingURL=omni-test.js.map
|
|
@@ -1 +0,0 @@
|
|
|
1
|
-
{"version":3,"file":"omni-test.js","sourceRoot":"","sources":["../../../s/timeline/sugar/omni-test.ts"],"names":[],"mappings":"AACA,OAAO,EAAC,IAAI,EAAC,MAAM,WAAW,CAAA;AAC9B,OAAO,EAAC,SAAS,EAAC,MAAM,wBAAwB,CAAA;AAEhD,EAAE;AACF,yBAAyB;AACzB,EAAE;AAEF,MAAM,IAAI,GAAG,IAAI,IAAI,EAAE,CAAA;AAEvB,EAAE;AACF,+BAA+B;AAC/B,EAAE;AAEF,MAAM,EAAC,MAAM,EAAE,MAAM,EAAC,GAAG,MAAM,IAAI,CAAC,IAAI,CAAC;IACxC,MAAM,EAAE,SAAS,EAAE;IACnB,MAAM,EAAE,SAAS,EAAE;CACnB,CAAC,CAAA;AAEF,EAAE;AACF,oBAAoB;AACpB,EAAE;AAEF,MAAM,QAAQ,GAAG,IAAI,CAAC,QAAQ,CAAC,CAAC,CAAC,EAAE,CAAC,CAAC,CAAC,QAAQ,CAC7C,CAAC,CAAC,IAAI,CAAC,MAAM,CAAC,EACd,CAAC,CAAC,UAAU,CAAC,SAAS,CAAC,GAAG,CAAC,EAC3B,CAAC,CAAC,KAAK,CACN,CAAC,CAAC,IAAI,CAAC,MAAM,CAAC,EACd,CAAC,CAAC,IAAI,CAAC,aAAa,CAAC,CACrB,CACD,CAAC,CAAA;AAEF,EAAE;AACF,mBAAmB;AACnB,EAAE;AAEF,OAAO,CAAC,GAAG,CAAC,IAAI,CAAC,SAAS,CAAC,QAAQ,EAAE,SAAS,EAAE,IAAI,CAAC,CAAC,CAAA"}
|
|
@@ -1,19 +0,0 @@
|
|
|
1
|
-
import { Pub } from "@e280/stz";
|
|
2
|
-
import { ProgressItem } from "../speech-recognition/whisper/parts/types.js";
|
|
3
|
-
export interface LoaderEvents {
|
|
4
|
-
onModelLoadProgress: Pub<ProgressItem[]>;
|
|
5
|
-
onTpsUpdate: Pub<[number]>;
|
|
6
|
-
}
|
|
7
|
-
export declare abstract class Loader {
|
|
8
|
-
readonly name: string;
|
|
9
|
-
model: string;
|
|
10
|
-
tps: number;
|
|
11
|
-
static loaderEvents: {
|
|
12
|
-
onModelLoadProgress: Pub<ProgressItem[]>;
|
|
13
|
-
onTpsUpdate: Pub<[number]>;
|
|
14
|
-
};
|
|
15
|
-
constructor(name: string, model: string);
|
|
16
|
-
abstract init(): Promise<void>;
|
|
17
|
-
abstract setModel(model: string): void;
|
|
18
|
-
setTps(value: number): void;
|
|
19
|
-
}
|
package/x/tools/common/loader.js
DELETED
|
@@ -1,18 +0,0 @@
|
|
|
1
|
-
import { pub } from "@e280/stz";
|
|
2
|
-
export class Loader {
|
|
3
|
-
name;
|
|
4
|
-
model;
|
|
5
|
-
tps = 0;
|
|
6
|
-
static loaderEvents = {
|
|
7
|
-
onModelLoadProgress: pub(),
|
|
8
|
-
onTpsUpdate: pub()
|
|
9
|
-
};
|
|
10
|
-
constructor(name, model) {
|
|
11
|
-
this.name = name;
|
|
12
|
-
this.model = model;
|
|
13
|
-
}
|
|
14
|
-
setTps(value) {
|
|
15
|
-
this.tps = value;
|
|
16
|
-
}
|
|
17
|
-
}
|
|
18
|
-
//# sourceMappingURL=loader.js.map
|
|
@@ -1 +0,0 @@
|
|
|
1
|
-
{"version":3,"file":"loader.js","sourceRoot":"","sources":["../../../s/tools/common/loader.ts"],"names":[],"mappings":"AAAA,OAAO,EAAC,GAAG,EAAM,MAAM,WAAW,CAAA;AAQlC,MAAM,OAAgB,MAAM;IAQC;IAAqB;IAPjD,GAAG,GAAG,CAAC,CAAA;IAEP,MAAM,CAAC,YAAY,GAAG;QACrB,mBAAmB,EAAE,GAAG,EAAkB;QAC1C,WAAW,EAAE,GAAG,EAAY;KAC5B,CAAA;IAED,YAA4B,IAAY,EAAS,KAAa;QAAlC,SAAI,GAAJ,IAAI,CAAQ;QAAS,UAAK,GAAL,KAAK,CAAQ;IAAG,CAAC;IAMlE,MAAM,CAAC,KAAa;QACnB,IAAI,CAAC,GAAG,GAAG,KAAK,CAAA;IACjB,CAAC"}
|
|
@@ -1,8 +0,0 @@
|
|
|
1
|
-
import { ProgressCallback } from "../speech-recognition/whisper/parts/types.js";
|
|
2
|
-
export declare class PipelineFactory {
|
|
3
|
-
task: string;
|
|
4
|
-
instance: any;
|
|
5
|
-
model: string | null;
|
|
6
|
-
constructor(task: string);
|
|
7
|
-
createInstance(model: string, progressCallback?: ProgressCallback): Promise<any>;
|
|
8
|
-
}
|
|
@@ -1,24 +0,0 @@
|
|
|
1
|
-
//@ts-ignore
|
|
2
|
-
import { pipeline } from "https://cdn.jsdelivr.net/npm/@huggingface/transformers@3.7.0/dist/transformers.min.js";
|
|
3
|
-
export class PipelineFactory {
|
|
4
|
-
task;
|
|
5
|
-
instance = null;
|
|
6
|
-
model = null;
|
|
7
|
-
constructor(task) {
|
|
8
|
-
this.task = task;
|
|
9
|
-
}
|
|
10
|
-
async createInstance(model, progressCallback) {
|
|
11
|
-
this.model = model;
|
|
12
|
-
return this.instance = await pipeline(this.task, this.model, {
|
|
13
|
-
dtype: {
|
|
14
|
-
encoder_model: this.model === "onnx-community/whisper-large-v3-turbo"
|
|
15
|
-
? "fp16"
|
|
16
|
-
: "fp32",
|
|
17
|
-
decoder_model_merged: "q4",
|
|
18
|
-
},
|
|
19
|
-
device: "webgpu",
|
|
20
|
-
progress_callback: progressCallback,
|
|
21
|
-
});
|
|
22
|
-
}
|
|
23
|
-
}
|
|
24
|
-
//# sourceMappingURL=transformer-pipeline.js.map
|
|
@@ -1 +0,0 @@
|
|
|
1
|
-
{"version":3,"file":"transformer-pipeline.js","sourceRoot":"","sources":["../../../s/tools/common/transformer-pipeline.ts"],"names":[],"mappings":"AAAA,YAAY;AACZ,OAAO,EAAC,QAAQ,EAAC,MAAM,uFAAuF,CAAA;AAI9G,MAAM,OAAO,eAAe;IAIR;IAHnB,QAAQ,GAAQ,IAAI,CAAA;IACpB,KAAK,GAAkB,IAAI,CAAA;IAE3B,YAAmB,IAAY;QAAZ,SAAI,GAAJ,IAAI,CAAQ;IAAG,CAAC;IAEnC,KAAK,CAAC,cAAc,CAAC,KAAa,EAAE,gBAAmC;QACtE,IAAI,CAAC,KAAK,GAAG,KAAK,CAAA;QAClB,OAAO,IAAI,CAAC,QAAQ,GAAG,MAAM,QAAQ,CAAC,IAAI,CAAC,IAAI,EAAE,IAAI,CAAC,KAAK,EAAE;YAC5D,KAAK,EAAE;gBACN,aAAa,EACZ,IAAI,CAAC,KAAK,KAAK,uCAAuC;oBACrD,CAAC,CAAC,MAAM;oBACR,CAAC,CAAC,MAAM;gBACV,oBAAoB,EAAE,IAAI;aAC1B;YACD,MAAM,EAAE,QAAQ;YAChB,iBAAiB,EAAE,gBAAgB;SACnC,CAAC,CAAA;IACH,CAAC;CACD"}
|
|
@@ -1,14 +0,0 @@
|
|
|
1
|
-
import { Loader } from "../../common/loader.js";
|
|
2
|
-
import { DecoderSource } from "../../../driver/fns/schematic.js";
|
|
3
|
-
import { SpeechRecognizerModels, Word, WordGroup } from "../whisper/parts/types.js";
|
|
4
|
-
export declare abstract class SpeechRecognizer extends Loader {
|
|
5
|
-
multilingual: boolean;
|
|
6
|
-
static speechRecognizerEvents: {
|
|
7
|
-
onTranscriptionChunk: import("@e280/stz").Pub<Word[]>;
|
|
8
|
-
onTranscribeProgress: import("@e280/stz").Pub<[number]>;
|
|
9
|
-
};
|
|
10
|
-
abstract transcribe(input: DecoderSource): Promise<WordGroup>;
|
|
11
|
-
setMultilingual(value: boolean): void;
|
|
12
|
-
detectLanguage?(input: Blob | AudioBuffer): Promise<string>;
|
|
13
|
-
setModel(value: SpeechRecognizerModels): void;
|
|
14
|
-
}
|
|
@@ -1,16 +0,0 @@
|
|
|
1
|
-
import { pub } from "@e280/stz";
|
|
2
|
-
import { Loader } from "../../common/loader.js";
|
|
3
|
-
export class SpeechRecognizer extends Loader {
|
|
4
|
-
multilingual = true;
|
|
5
|
-
static speechRecognizerEvents = {
|
|
6
|
-
onTranscriptionChunk: pub(),
|
|
7
|
-
onTranscribeProgress: pub()
|
|
8
|
-
};
|
|
9
|
-
setMultilingual(value) {
|
|
10
|
-
this.multilingual = value;
|
|
11
|
-
}
|
|
12
|
-
setModel(value) {
|
|
13
|
-
this.model = value;
|
|
14
|
-
}
|
|
15
|
-
}
|
|
16
|
-
//# sourceMappingURL=model.js.map
|
|
@@ -1 +0,0 @@
|
|
|
1
|
-
{"version":3,"file":"model.js","sourceRoot":"","sources":["../../../../s/tools/speech-recognition/common/model.ts"],"names":[],"mappings":"AAAA,OAAO,EAAC,GAAG,EAAC,MAAM,WAAW,CAAA;AAE7B,OAAO,EAAC,MAAM,EAAC,MAAM,wBAAwB,CAAA;AAI7C,MAAM,OAAgB,gBAAiB,SAAQ,MAAM;IACpD,YAAY,GAAG,IAAI,CAAA;IAEnB,MAAM,CAAC,sBAAsB,GAAG;QAC/B,oBAAoB,EAAE,GAAG,EAAU;QACnC,oBAAoB,EAAE,GAAG,EAAY;KACrC,CAAA;IAID,eAAe,CAAC,KAAc;QAC7B,IAAI,CAAC,YAAY,GAAG,KAAK,CAAA;IAC1B,CAAC;IAID,QAAQ,CAAC,KAA6B;QACrC,IAAI,CAAC,KAAK,GAAG,KAAK,CAAA;IACnB,CAAC"}
|
|
@@ -1,13 +0,0 @@
|
|
|
1
|
-
import { ProgressItem } from "../parts/types.js";
|
|
2
|
-
import { SpeechRecognizerHostEvents } from "./schematic.js";
|
|
3
|
-
export declare const setupWhisperHost: (events: SpeechRecognizerHostEvents) => import("@e280/comrade").SetupHost<{
|
|
4
|
-
work: {
|
|
5
|
-
transcribe(input: import("../parts/types.js").TranscriptionMessage): Promise<import("../parts/types.js").TranscriptionResult | null>;
|
|
6
|
-
};
|
|
7
|
-
host: {
|
|
8
|
-
updateModelLoadProgress(item: ProgressItem): Promise<void>;
|
|
9
|
-
deliverTranscriptionChunk(chunk: import("../parts/types.js").TranscriptionChunk): Promise<void>;
|
|
10
|
-
updateTps(value: number): Promise<void>;
|
|
11
|
-
updateTranscribeProgress(value: number): Promise<void>;
|
|
12
|
-
};
|
|
13
|
-
}>;
|
|
@@ -1,19 +0,0 @@
|
|
|
1
|
-
import { Comrade } from "@e280/comrade";
|
|
2
|
-
export const setupWhisperHost = (events) => (Comrade.host(_shell => ({
|
|
3
|
-
async updateModelLoadProgress(item) {
|
|
4
|
-
events.onModelLoadProgress.pub(item);
|
|
5
|
-
},
|
|
6
|
-
async deliverTranscriptionChunk(chunk) {
|
|
7
|
-
events.onTranscriptionChunk.pub({
|
|
8
|
-
text: chunk.text,
|
|
9
|
-
timestamp: chunk.timestamp
|
|
10
|
-
});
|
|
11
|
-
},
|
|
12
|
-
async updateTps(value) {
|
|
13
|
-
events.onTpsUpdate.pub(value);
|
|
14
|
-
},
|
|
15
|
-
async updateTranscribeProgress(value) {
|
|
16
|
-
events.onTranscribeProgress(value);
|
|
17
|
-
}
|
|
18
|
-
})));
|
|
19
|
-
//# sourceMappingURL=host.js.map
|
|
@@ -1 +0,0 @@
|
|
|
1
|
-
{"version":3,"file":"host.js","sourceRoot":"","sources":["../../../../../s/tools/speech-recognition/whisper/fns/host.ts"],"names":[],"mappings":"AACA,OAAO,EAAC,OAAO,EAAC,MAAM,eAAe,CAAA;AAIrC,MAAM,CAAC,MAAM,gBAAgB,GAAG,CAAC,MAAkC,EAAE,EAAE,CAAC,CACvE,OAAO,CAAC,IAAI,CAAmB,MAAM,CAAC,EAAE,CAAC,CAAC;IACzC,KAAK,CAAC,uBAAuB,CAAC,IAAI;QACjC,MAAM,CAAC,mBAAmB,CAAC,GAAG,CAAC,IAAI,CAAC,CAAA;IACrC,CAAC;IACD,KAAK,CAAC,yBAAyB,CAAC,KAAK;QACpC,MAAM,CAAC,oBAAoB,CAAC,GAAG,CAAC;YAC/B,IAAI,EAAE,KAAK,CAAC,IAAI;YAChB,SAAS,EAAE,KAAK,CAAC,SAAS;SAC1B,CAAC,CAAA;IACH,CAAC;IACD,KAAK,CAAC,SAAS,CAAC,KAAK;QACpB,MAAM,CAAC,WAAW,CAAC,GAAG,CAAC,KAAK,CAAC,CAAA;IAC9B,CAAC;IACD,KAAK,CAAC,wBAAwB,CAAC,KAAK;QACnC,MAAM,CAAC,oBAAoB,CAAC,KAAK,CAAC,CAAA;IACnC,CAAC;CACD,CAAC,CAAC,CACH,CAAA"}
|
|
@@ -1,19 +0,0 @@
|
|
|
1
|
-
import { Pub } from "@e280/stz";
|
|
2
|
-
import { AsSchematic } from "@e280/comrade";
|
|
3
|
-
import { LoaderEvents } from "../../../common/loader.js";
|
|
4
|
-
import { ProgressItem, TranscriptionChunk, TranscriptionMessage, TranscriptionResult, Word } from "../parts/types.js";
|
|
5
|
-
export type WhisperSchematic = AsSchematic<{
|
|
6
|
-
work: {
|
|
7
|
-
transcribe(input: TranscriptionMessage): Promise<TranscriptionResult | null>;
|
|
8
|
-
};
|
|
9
|
-
host: {
|
|
10
|
-
updateModelLoadProgress(item: ProgressItem): Promise<void>;
|
|
11
|
-
deliverTranscriptionChunk(chunk: TranscriptionChunk): Promise<void>;
|
|
12
|
-
updateTps(value: number): Promise<void>;
|
|
13
|
-
updateTranscribeProgress(value: number): Promise<void>;
|
|
14
|
-
};
|
|
15
|
-
}>;
|
|
16
|
-
export interface SpeechRecognizerHostEvents extends LoaderEvents {
|
|
17
|
-
onTranscriptionChunk: Pub<Word[]>;
|
|
18
|
-
onTranscribeProgress: Pub<[number]>;
|
|
19
|
-
}
|
|
@@ -1 +0,0 @@
|
|
|
1
|
-
{"version":3,"file":"schematic.js","sourceRoot":"","sources":["../../../../../s/tools/speech-recognition/whisper/fns/schematic.ts"],"names":[],"mappings":""}
|
|
@@ -1,12 +0,0 @@
|
|
|
1
|
-
import { TranscriptionChunk } from "../parts/types.js";
|
|
2
|
-
export declare const setupWhisperWork: import("@e280/comrade").SetupWork<{
|
|
3
|
-
work: {
|
|
4
|
-
transcribe(input: import("../parts/types.js").TranscriptionMessage): Promise<import("../parts/types.js").TranscriptionResult | null>;
|
|
5
|
-
};
|
|
6
|
-
host: {
|
|
7
|
-
updateModelLoadProgress(item: import("../parts/types.js").ProgressItem): Promise<void>;
|
|
8
|
-
deliverTranscriptionChunk(chunk: TranscriptionChunk): Promise<void>;
|
|
9
|
-
updateTps(value: number): Promise<void>;
|
|
10
|
-
updateTranscribeProgress(value: number): Promise<void>;
|
|
11
|
-
};
|
|
12
|
-
}>;
|
|
@@ -1,74 +0,0 @@
|
|
|
1
|
-
import { Comrade } from "@e280/comrade";
|
|
2
|
-
//@ts-ignore
|
|
3
|
-
import { WhisperTextStreamer } from "https://cdn.jsdelivr.net/npm/@huggingface/transformers@3.7.0/dist/transformers.min.js";
|
|
4
|
-
import { PipelineFactory } from "../../../common/transformer-pipeline.js";
|
|
5
|
-
// TODO suspicious globals, probably bad
|
|
6
|
-
const pipeline = new PipelineFactory("automatic-speech-recognition");
|
|
7
|
-
let transcriber;
|
|
8
|
-
export const setupWhisperWork = Comrade.work(shell => ({
|
|
9
|
-
async transcribe({ audio, model, language, duration }) {
|
|
10
|
-
const isDistil = model.startsWith("distil-whisper/");
|
|
11
|
-
if (!pipeline.model || pipeline.model !== model) {
|
|
12
|
-
pipeline.instance?.dispose()?.();
|
|
13
|
-
pipeline.instance = null;
|
|
14
|
-
transcriber = await pipeline.createInstance(model, (data) => {
|
|
15
|
-
if (data.progress)
|
|
16
|
-
shell.host.updateModelLoadProgress({
|
|
17
|
-
id: data.file,
|
|
18
|
-
progress: data.progress
|
|
19
|
-
});
|
|
20
|
-
});
|
|
21
|
-
}
|
|
22
|
-
const timePrecision = transcriber.processor.feature_extractor.config.chunk_length /
|
|
23
|
-
transcriber.model.config.max_source_positions;
|
|
24
|
-
const chunkLength = isDistil ? 20 : 30;
|
|
25
|
-
const strideLength = isDistil ? 3 : 5;
|
|
26
|
-
let chunkCount = 0;
|
|
27
|
-
let startTime = null;
|
|
28
|
-
let tokenCount = 0;
|
|
29
|
-
let tps = 0;
|
|
30
|
-
const chunkDuration = chunkLength - strideLength;
|
|
31
|
-
const estimateProgress = () => {
|
|
32
|
-
const audioProgressSeconds = chunkCount * chunkDuration;
|
|
33
|
-
return Math.min(audioProgressSeconds / duration, 1);
|
|
34
|
-
};
|
|
35
|
-
const streamer = new WhisperTextStreamer(transcriber.tokenizer, {
|
|
36
|
-
time_precision: timePrecision,
|
|
37
|
-
token_callback_function: () => {
|
|
38
|
-
startTime ??= performance.now();
|
|
39
|
-
if (++tokenCount > 1) {
|
|
40
|
-
tps = (tokenCount / (performance.now() - startTime)) * 1000;
|
|
41
|
-
shell.host.updateTps(tps);
|
|
42
|
-
}
|
|
43
|
-
},
|
|
44
|
-
callback_function: (textChunk) => {
|
|
45
|
-
shell.host.deliverTranscriptionChunk(textChunk);
|
|
46
|
-
},
|
|
47
|
-
on_finalize: () => {
|
|
48
|
-
startTime = null;
|
|
49
|
-
tokenCount = 0;
|
|
50
|
-
chunkCount++;
|
|
51
|
-
const progress = estimateProgress();
|
|
52
|
-
shell.host.updateTranscribeProgress(progress);
|
|
53
|
-
},
|
|
54
|
-
});
|
|
55
|
-
const output = await transcriber(audio, {
|
|
56
|
-
top_k: 0,
|
|
57
|
-
do_sample: false,
|
|
58
|
-
chunk_length_s: chunkLength,
|
|
59
|
-
stride_length_s: strideLength,
|
|
60
|
-
language,
|
|
61
|
-
task: "transcribe",
|
|
62
|
-
return_timestamps: "word", // if using "word" the on_chunk_start & end is not called thus we cant retrieve timestamps, only after whole thing finishes
|
|
63
|
-
force_full_sequences: false,
|
|
64
|
-
streamer,
|
|
65
|
-
});
|
|
66
|
-
if (!output)
|
|
67
|
-
return null;
|
|
68
|
-
return {
|
|
69
|
-
tps,
|
|
70
|
-
...output,
|
|
71
|
-
};
|
|
72
|
-
}
|
|
73
|
-
}));
|
|
74
|
-
//# sourceMappingURL=work.js.map
|
|
@@ -1 +0,0 @@
|
|
|
1
|
-
{"version":3,"file":"work.js","sourceRoot":"","sources":["../../../../../s/tools/speech-recognition/whisper/fns/work.ts"],"names":[],"mappings":"AAAA,OAAO,EAAC,OAAO,EAAC,MAAM,eAAe,CAAA;AACrC,YAAY;AACZ,OAAO,EAAW,mBAAmB,EAAC,MAAM,uFAAuF,CAAA;AAInI,OAAO,EAAC,eAAe,EAAC,MAAM,yCAAyC,CAAA;AAEvE,wCAAwC;AACxC,MAAM,QAAQ,GAAG,IAAI,eAAe,CAAC,8BAA8B,CAAC,CAAA;AACpE,IAAI,WAAgB,CAAA;AAEpB,MAAM,CAAC,MAAM,gBAAgB,GAAG,OAAO,CAAC,IAAI,CAAmB,KAAK,CAAC,EAAE,CAAC,CAAC;IACxE,KAAK,CAAC,UAAU,CAAC,EAAC,KAAK,EAAE,KAAK,EAAE,QAAQ,EAAE,QAAQ,EAAC;QAClD,MAAM,QAAQ,GAAG,KAAK,CAAC,UAAU,CAAC,iBAAiB,CAAC,CAAA;QAEpD,IAAG,CAAC,QAAQ,CAAC,KAAK,IAAI,QAAQ,CAAC,KAAK,KAAK,KAAK,EAAE,CAAC;YAChD,QAAQ,CAAC,QAAQ,EAAE,OAAO,EAAE,EAAE,EAAE,CAAA;YAChC,QAAQ,CAAC,QAAQ,GAAG,IAAI,CAAA;YACxB,WAAW,GAAG,MAAM,QAAQ,CAAC,cAAc,CAC1C,KAAK,EACL,CAAC,IAAI,EAAE,EAAE;gBACR,IAAG,IAAI,CAAC,QAAQ;oBACf,KAAK,CAAC,IAAI,CAAC,uBAAuB,CAAC;wBAClC,EAAE,EAAE,IAAI,CAAC,IAAI;wBACb,QAAQ,EAAE,IAAI,CAAC,QAAQ;qBACvB,CAAC,CAAA;YACJ,CAAC,CACD,CAAA;QACF,CAAC;QAED,MAAM,aAAa,GAClB,WAAW,CAAC,SAAS,CAAC,iBAAiB,CAAC,MAAM,CAAC,YAAY;YAC3D,WAAW,CAAC,KAAK,CAAC,MAAM,CAAC,oBAAoB,CAAA;QAE9C,MAAM,WAAW,GAAG,QAAQ,CAAC,CAAC,CAAC,EAAE,CAAC,CAAC,CAAC,EAAE,CAAA;QACtC,MAAM,YAAY,GAAG,QAAQ,CAAC,CAAC,CAAC,CAAC,CAAC,CAAC,CAAC,CAAC,CAAA;QAErC,IAAI,UAAU,GAAG,CAAC,CAAA;QAClB,IAAI,SAAS,GAAkB,IAAI,CAAA;QACnC,IAAI,UAAU,GAAG,CAAC,CAAA;QAClB,IAAI,GAAG,GAAG,CAAC,CAAA;QAEX,MAAM,aAAa,GAAG,WAAW,GAAG,YAAY,CAAA;QAEhD,MAAM,gBAAgB,GAAG,GAAG,EAAE;YAC7B,MAAM,oBAAoB,GAAG,UAAU,GAAG,aAAa,CAAA;YACvD,OAAO,IAAI,CAAC,GAAG,CAAC,oBAAoB,GAAG,QAAQ,EAAE,CAAC,CAAC,CAAA;QACpD,CAAC,CAAA;QAED,MAAM,QAAQ,GAAG,IAAI,mBAAmB,CAAC,WAAW,CAAC,SAAS,EAAE;YAC/D,cAAc,EAAE,aAAa;YAC7B,uBAAuB,EAAE,GAAG,EAAE;gBAC7B,SAAS,KAAK,WAAW,CAAC,GAAG,EAAE,CAAA;gBAC/B,IAAI,EAAE,UAAU,GAAG,CAAC,EAAE,CAAC;oBACtB,GAAG,GAAG,CAAC,UAAU,GAAG,CAAC,WAAW,CAAC,GAAG,EAAE,GAAG,SAAS,CAAC,CAAC,GAAG,IAAI,CAAA;oBAC3D,KAAK,CAAC,IAAI,CAAC,SAAS,CAAC,GAAG,CAAC,CAAA;gBAC1B,CAAC;YACF,CAAC;YACD,iBAAiB,EAAE,CAAC,SAAc,EAAE,EAAE;gBACrC,KAAK,CAAC,IAAI,CAAC,yBAAyB,CAAC,SAAS,CAAC,CAAA;YAChD,CAAC;YACD,WAAW,EAAE,GAAG,EAAE;gBACjB,SAAS,GAAG,IAAI,CAAA;gBAChB,UAAU,GAAG,CAAC,CAAA;gBACd,UAAU,EAAE,CAAA;gBACZ,MAAM,QAAQ,GAAG,gBAAgB,EAAE,CAAA;gBACnC,KAAK,CAAC,IAAI,CAAC,wBAAwB,CAAC,QAAQ,CAAC,CAAA;YAC9C,CAAC;SACD,CAAC,CAAA;QAEF,MAAM,MAAM,GAAG,MAAM,WAAW,CAAC,KAAK,EAAE;YACvC,KAAK,EAAE,CAAC;YACR,SAAS,EAAE,KAAK;YAChB,cAAc,EAAE,WAAW;YAC3B,eAAe,EAAE,YAAY;YAC7B,QAAQ;YACR,IAAI,EAAE,YAAY;YAClB,iBAAiB,EAAE,MAAM,EAAE,2HAA2H;YACtJ,oBAAoB,EAAE,KAAK;YAC3B,QAAQ;SACR,CAAC,CAAA;QAEF,IAAI,CAAC,MAAM;YAAE,OAAO,IAAI,CAAA;QAExB,OAAO;YACN,GAAG;YACH,GAAG,MAAM;SACT,CAAA;IACF,CAAC;CACD,CAAC,CAAC,CAAA"}
|
|
@@ -1,31 +0,0 @@
|
|
|
1
|
-
export interface ProgressItem {
|
|
2
|
-
id: string;
|
|
3
|
-
progress: number;
|
|
4
|
-
}
|
|
5
|
-
export type Word = {
|
|
6
|
-
text: string;
|
|
7
|
-
timestamp: [start: number, end: number];
|
|
8
|
-
};
|
|
9
|
-
export type WordGroup = Word[];
|
|
10
|
-
export type Transcript = WordGroup[];
|
|
11
|
-
export interface TranscriptionChunk {
|
|
12
|
-
text: string;
|
|
13
|
-
offset: number;
|
|
14
|
-
timestamp: [number, number];
|
|
15
|
-
finalised: boolean;
|
|
16
|
-
}
|
|
17
|
-
export interface TranscriptionMessage {
|
|
18
|
-
audio: Float32Array;
|
|
19
|
-
model: string;
|
|
20
|
-
subtask: string | null;
|
|
21
|
-
language: string | null;
|
|
22
|
-
duration: number;
|
|
23
|
-
}
|
|
24
|
-
export interface TranscriptionResult {
|
|
25
|
-
text: string;
|
|
26
|
-
chunks: TranscriptionChunk[];
|
|
27
|
-
tps: number;
|
|
28
|
-
}
|
|
29
|
-
export type ProgressCallback = (data: any) => void;
|
|
30
|
-
export type SpeechRecognizerModels = "onnx-community/whisper-tiny_timestamped";
|
|
31
|
-
export type SpeechRecognizerSubtasks = "transcribe";
|
|
@@ -1 +0,0 @@
|
|
|
1
|
-
{"version":3,"file":"types.js","sourceRoot":"","sources":["../../../../../s/tools/speech-recognition/whisper/parts/types.ts"],"names":[],"mappings":""}
|
|
@@ -1 +0,0 @@
|
|
|
1
|
-
export {};
|