videopython 0.26.7__tar.gz → 0.26.9__tar.gz
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- {videopython-0.26.7 → videopython-0.26.9}/PKG-INFO +1 -1
- {videopython-0.26.7 → videopython-0.26.9}/pyproject.toml +1 -1
- {videopython-0.26.7 → videopython-0.26.9}/src/videopython/ai/dubbing/pipeline.py +81 -6
- {videopython-0.26.7 → videopython-0.26.9}/src/videopython/ai/generation/translation.py +27 -5
- videopython-0.26.9/src/videopython/ai/understanding/separation.py +304 -0
- videopython-0.26.7/src/videopython/ai/understanding/separation.py +0 -131
- {videopython-0.26.7 → videopython-0.26.9}/.gitignore +0 -0
- {videopython-0.26.7 → videopython-0.26.9}/LICENSE +0 -0
- {videopython-0.26.7 → videopython-0.26.9}/README.md +0 -0
- {videopython-0.26.7 → videopython-0.26.9}/src/videopython/__init__.py +0 -0
- {videopython-0.26.7 → videopython-0.26.9}/src/videopython/ai/__init__.py +0 -0
- {videopython-0.26.7 → videopython-0.26.9}/src/videopython/ai/_device.py +0 -0
- {videopython-0.26.7 → videopython-0.26.9}/src/videopython/ai/dubbing/__init__.py +0 -0
- {videopython-0.26.7 → videopython-0.26.9}/src/videopython/ai/dubbing/dubber.py +0 -0
- {videopython-0.26.7 → videopython-0.26.9}/src/videopython/ai/dubbing/models.py +0 -0
- {videopython-0.26.7 → videopython-0.26.9}/src/videopython/ai/dubbing/remux.py +0 -0
- {videopython-0.26.7 → videopython-0.26.9}/src/videopython/ai/dubbing/timing.py +0 -0
- {videopython-0.26.7 → videopython-0.26.9}/src/videopython/ai/generation/__init__.py +0 -0
- {videopython-0.26.7 → videopython-0.26.9}/src/videopython/ai/generation/audio.py +0 -0
- {videopython-0.26.7 → videopython-0.26.9}/src/videopython/ai/generation/image.py +0 -0
- {videopython-0.26.7 → videopython-0.26.9}/src/videopython/ai/generation/video.py +0 -0
- {videopython-0.26.7 → videopython-0.26.9}/src/videopython/ai/registry.py +0 -0
- {videopython-0.26.7 → videopython-0.26.9}/src/videopython/ai/swapping/__init__.py +0 -0
- {videopython-0.26.7 → videopython-0.26.9}/src/videopython/ai/swapping/inpainter.py +0 -0
- {videopython-0.26.7 → videopython-0.26.9}/src/videopython/ai/swapping/models.py +0 -0
- {videopython-0.26.7 → videopython-0.26.9}/src/videopython/ai/swapping/segmenter.py +0 -0
- {videopython-0.26.7 → videopython-0.26.9}/src/videopython/ai/swapping/swapper.py +0 -0
- {videopython-0.26.7 → videopython-0.26.9}/src/videopython/ai/transforms.py +0 -0
- {videopython-0.26.7 → videopython-0.26.9}/src/videopython/ai/understanding/__init__.py +0 -0
- {videopython-0.26.7 → videopython-0.26.9}/src/videopython/ai/understanding/audio.py +0 -0
- {videopython-0.26.7 → videopython-0.26.9}/src/videopython/ai/understanding/image.py +0 -0
- {videopython-0.26.7 → videopython-0.26.9}/src/videopython/ai/understanding/temporal.py +0 -0
- {videopython-0.26.7 → videopython-0.26.9}/src/videopython/ai/video_analysis.py +0 -0
- {videopython-0.26.7 → videopython-0.26.9}/src/videopython/base/__init__.py +0 -0
- {videopython-0.26.7 → videopython-0.26.9}/src/videopython/base/audio/__init__.py +0 -0
- {videopython-0.26.7 → videopython-0.26.9}/src/videopython/base/audio/analysis.py +0 -0
- {videopython-0.26.7 → videopython-0.26.9}/src/videopython/base/audio/audio.py +0 -0
- {videopython-0.26.7 → videopython-0.26.9}/src/videopython/base/combine.py +0 -0
- {videopython-0.26.7 → videopython-0.26.9}/src/videopython/base/description.py +0 -0
- {videopython-0.26.7 → videopython-0.26.9}/src/videopython/base/effects.py +0 -0
- {videopython-0.26.7 → videopython-0.26.9}/src/videopython/base/exceptions.py +0 -0
- {videopython-0.26.7 → videopython-0.26.9}/src/videopython/base/progress.py +0 -0
- {videopython-0.26.7 → videopython-0.26.9}/src/videopython/base/registry.py +0 -0
- {videopython-0.26.7 → videopython-0.26.9}/src/videopython/base/scene.py +0 -0
- {videopython-0.26.7 → videopython-0.26.9}/src/videopython/base/streaming.py +0 -0
- {videopython-0.26.7 → videopython-0.26.9}/src/videopython/base/text/__init__.py +0 -0
- {videopython-0.26.7 → videopython-0.26.9}/src/videopython/base/text/overlay.py +0 -0
- {videopython-0.26.7 → videopython-0.26.9}/src/videopython/base/text/transcription.py +0 -0
- {videopython-0.26.7 → videopython-0.26.9}/src/videopython/base/transforms.py +0 -0
- {videopython-0.26.7 → videopython-0.26.9}/src/videopython/base/transitions.py +0 -0
- {videopython-0.26.7 → videopython-0.26.9}/src/videopython/base/utils.py +0 -0
- {videopython-0.26.7 → videopython-0.26.9}/src/videopython/base/video.py +0 -0
- {videopython-0.26.7 → videopython-0.26.9}/src/videopython/editing/__init__.py +0 -0
- {videopython-0.26.7 → videopython-0.26.9}/src/videopython/editing/multicam.py +0 -0
- {videopython-0.26.7 → videopython-0.26.9}/src/videopython/editing/premiere_xml.py +0 -0
- {videopython-0.26.7 → videopython-0.26.9}/src/videopython/editing/video_edit.py +0 -0
- {videopython-0.26.7 → videopython-0.26.9}/src/videopython/py.typed +0 -0
|
@@ -7,12 +7,41 @@ import tempfile
|
|
|
7
7
|
from pathlib import Path
|
|
8
8
|
from typing import TYPE_CHECKING, Any, Callable, Literal
|
|
9
9
|
|
|
10
|
+
import numpy as np
|
|
11
|
+
|
|
10
12
|
from videopython.ai.dubbing.models import DubbingResult, RevoiceResult, SeparatedAudio
|
|
11
13
|
from videopython.ai.dubbing.timing import TimingSynchronizer
|
|
12
14
|
|
|
13
15
|
if TYPE_CHECKING:
|
|
14
16
|
from videopython.base.audio import Audio
|
|
15
17
|
|
|
18
|
+
|
|
19
|
+
def _peak_match(target: Audio, reference: Audio) -> Audio:
|
|
20
|
+
"""Scale ``target`` so its peak amplitude matches ``reference``.
|
|
21
|
+
|
|
22
|
+
Demucs background normalization and the timing-assembler peak guard
|
|
23
|
+
each clamp at 1.0 instead of restoring headroom, so a dubbed mix
|
|
24
|
+
typically lands quieter than the source — perceptually "thinner."
|
|
25
|
+
A single peak match recovers most of that drift without LUFS deps.
|
|
26
|
+
|
|
27
|
+
No-op when either side has zero peak (silent input or all-silent dub).
|
|
28
|
+
The new ``Audio`` shares no buffer with ``target``.
|
|
29
|
+
"""
|
|
30
|
+
from videopython.base.audio import Audio as _Audio
|
|
31
|
+
|
|
32
|
+
target_peak = float(np.max(np.abs(target.data))) if target.data.size else 0.0
|
|
33
|
+
reference_peak = float(np.max(np.abs(reference.data))) if reference.data.size else 0.0
|
|
34
|
+
|
|
35
|
+
if target_peak <= 0.0 or reference_peak <= 0.0:
|
|
36
|
+
return target
|
|
37
|
+
|
|
38
|
+
scale = reference_peak / target_peak
|
|
39
|
+
if abs(scale - 1.0) < 1e-3:
|
|
40
|
+
return target
|
|
41
|
+
|
|
42
|
+
return _Audio(target.data * scale, target.metadata)
|
|
43
|
+
|
|
44
|
+
|
|
16
45
|
WhisperModel = Literal["tiny", "base", "small", "medium", "large", "turbo"]
|
|
17
46
|
|
|
18
47
|
logger = logging.getLogger(__name__)
|
|
@@ -239,7 +268,19 @@ class LocalDubbingPipeline:
|
|
|
239
268
|
if self._separator is None:
|
|
240
269
|
self._init_separator()
|
|
241
270
|
|
|
242
|
-
|
|
271
|
+
# Limit Demucs to the speech-bearing portion of the audio. The
|
|
272
|
+
# transcription has already located every speech region; running
|
|
273
|
+
# source separation outside those is pure overhead (no vocals to
|
|
274
|
+
# isolate). On talk-heavy sources with silence/music gaps this
|
|
275
|
+
# roughly halves separation time. When speech covers most of the
|
|
276
|
+
# track separate_regions falls back to a full-track separate().
|
|
277
|
+
from videopython.ai.understanding.separation import _merge_regions
|
|
278
|
+
|
|
279
|
+
speech_regions = _merge_regions(
|
|
280
|
+
[(s.start, s.end) for s in transcription.segments],
|
|
281
|
+
audio_duration=source_audio.metadata.duration_seconds,
|
|
282
|
+
)
|
|
283
|
+
separated_audio = self._separator.separate_regions(source_audio, speech_regions)
|
|
243
284
|
self._maybe_unload("_separator")
|
|
244
285
|
vocal_audio = separated_audio.vocals
|
|
245
286
|
background_audio = separated_audio.background
|
|
@@ -296,6 +337,12 @@ class LocalDubbingPipeline:
|
|
|
296
337
|
for i, segment in enumerate(translated_segments):
|
|
297
338
|
if segment.duration < 0.1:
|
|
298
339
|
continue
|
|
340
|
+
# Translation filter (translation.py:_is_translatable_text)
|
|
341
|
+
# leaves translated_text="" for punctuation-only or empty
|
|
342
|
+
# segments. Don't TTS those — saves a model call and avoids
|
|
343
|
+
# injecting hallucinated speech into the dubbed track.
|
|
344
|
+
if not segment.translated_text.strip():
|
|
345
|
+
continue
|
|
299
346
|
|
|
300
347
|
progress = 0.50 + (0.30 * (i / len(translated_segments)))
|
|
301
348
|
report_progress(f"Generating speech ({i + 1}/{len(translated_segments)})", progress)
|
|
@@ -303,10 +350,25 @@ class LocalDubbingPipeline:
|
|
|
303
350
|
speaker = segment.speaker or "speaker_0"
|
|
304
351
|
cached_path = speaker_wav_paths.get(speaker) if voice_clone else None
|
|
305
352
|
|
|
306
|
-
|
|
307
|
-
|
|
308
|
-
|
|
309
|
-
|
|
353
|
+
try:
|
|
354
|
+
if cached_path is not None:
|
|
355
|
+
dubbed_audio = self._tts.generate_audio(segment.translated_text, voice_sample_path=cached_path)
|
|
356
|
+
else:
|
|
357
|
+
dubbed_audio = self._tts.generate_audio(segment.translated_text)
|
|
358
|
+
except Exception as e:
|
|
359
|
+
# Chatterbox occasionally crashes on short translated text
|
|
360
|
+
# (alignment_stream_analyzer indexing on tensors with <=5
|
|
361
|
+
# speech tokens). One bad segment shouldn't lose a long
|
|
362
|
+
# multi-hour run — log and skip so the rest proceeds.
|
|
363
|
+
logger.warning(
|
|
364
|
+
"TTS failed for segment %d/%d (speaker=%s, text=%r): %s — skipping",
|
|
365
|
+
i + 1,
|
|
366
|
+
len(translated_segments),
|
|
367
|
+
speaker,
|
|
368
|
+
segment.translated_text,
|
|
369
|
+
e,
|
|
370
|
+
)
|
|
371
|
+
continue
|
|
310
372
|
|
|
311
373
|
dubbed_segments.append(dubbed_audio)
|
|
312
374
|
target_durations.append(segment.duration)
|
|
@@ -343,6 +405,11 @@ class LocalDubbingPipeline:
|
|
|
343
405
|
else:
|
|
344
406
|
final_audio = dubbed_speech
|
|
345
407
|
|
|
408
|
+
# Peak-match against the source so the dub doesn't land quieter
|
|
409
|
+
# than the original. Done last so it captures both vocals+background
|
|
410
|
+
# mixes and speech-only outputs uniformly.
|
|
411
|
+
final_audio = _peak_match(final_audio, source_audio)
|
|
412
|
+
|
|
346
413
|
report_progress("Complete", 1.0)
|
|
347
414
|
|
|
348
415
|
return DubbingResult(
|
|
@@ -393,7 +460,13 @@ class LocalDubbingPipeline:
|
|
|
393
460
|
if self._separator is None:
|
|
394
461
|
self._init_separator()
|
|
395
462
|
|
|
396
|
-
|
|
463
|
+
from videopython.ai.understanding.separation import _merge_regions
|
|
464
|
+
|
|
465
|
+
speech_regions = _merge_regions(
|
|
466
|
+
[(s.start, s.end) for s in transcription.segments],
|
|
467
|
+
audio_duration=source_audio.metadata.duration_seconds,
|
|
468
|
+
)
|
|
469
|
+
separated_audio = self._separator.separate_regions(source_audio, speech_regions)
|
|
397
470
|
self._maybe_unload("_separator")
|
|
398
471
|
vocal_audio = separated_audio.vocals
|
|
399
472
|
background_audio = separated_audio.background
|
|
@@ -448,6 +521,8 @@ class LocalDubbingPipeline:
|
|
|
448
521
|
else:
|
|
449
522
|
final_audio = generated_speech
|
|
450
523
|
|
|
524
|
+
final_audio = _peak_match(final_audio, source_audio)
|
|
525
|
+
|
|
451
526
|
report_progress("Complete", 1.0)
|
|
452
527
|
|
|
453
528
|
return RevoiceResult(
|
|
@@ -8,6 +8,17 @@ from videopython.ai._device import log_device_initialization, release_device_mem
|
|
|
8
8
|
from videopython.ai.dubbing.models import TranslatedSegment
|
|
9
9
|
from videopython.base.text.transcription import TranscriptionSegment
|
|
10
10
|
|
|
11
|
+
|
|
12
|
+
def _is_translatable_text(text: str) -> bool:
|
|
13
|
+
"""Return True if text has enough content to be worth translating.
|
|
14
|
+
|
|
15
|
+
Whisper routinely emits punctuation-only or single-character segments
|
|
16
|
+
(" .", "...", "?", "♪") that MarianMT can hallucinate full sentences
|
|
17
|
+
from. Require at least 2 alphanumeric characters to filter these out.
|
|
18
|
+
"""
|
|
19
|
+
return sum(1 for c in text if c.isalnum()) >= 2
|
|
20
|
+
|
|
21
|
+
|
|
11
22
|
LANGUAGE_NAMES = {
|
|
12
23
|
"en": "English",
|
|
13
24
|
"es": "Spanish",
|
|
@@ -159,17 +170,28 @@ class TextTranslator:
|
|
|
159
170
|
target_lang: str,
|
|
160
171
|
source_lang: str | None = None,
|
|
161
172
|
) -> list[TranslatedSegment]:
|
|
162
|
-
"""Translate transcription segments while preserving timing/speaker info.
|
|
173
|
+
"""Translate transcription segments while preserving timing/speaker info.
|
|
174
|
+
|
|
175
|
+
Segments whose text is empty or contains fewer than 2 alphanumeric
|
|
176
|
+
characters are not sent to the model — they receive
|
|
177
|
+
``translated_text=""`` instead. This avoids MarianMT hallucinating
|
|
178
|
+
full sentences from " .", "...", or single-token Whisper segments,
|
|
179
|
+
which would otherwise be TTS'd into the dubbed track.
|
|
180
|
+
"""
|
|
163
181
|
effective_source = source_lang or "en"
|
|
164
|
-
|
|
165
|
-
|
|
182
|
+
|
|
183
|
+
translatable_indices = [i for i, segment in enumerate(segments) if _is_translatable_text(segment.text)]
|
|
184
|
+
translatable_texts = [segments[i].text for i in translatable_indices]
|
|
185
|
+
translated_texts = self.translate_batch(translatable_texts, target_lang, source_lang)
|
|
186
|
+
|
|
187
|
+
translation_map: dict[int, str] = dict(zip(translatable_indices, translated_texts))
|
|
166
188
|
|
|
167
189
|
translated_segments = []
|
|
168
|
-
for
|
|
190
|
+
for i, segment in enumerate(segments):
|
|
169
191
|
translated_segments.append(
|
|
170
192
|
TranslatedSegment(
|
|
171
193
|
original_segment=segment,
|
|
172
|
-
translated_text=
|
|
194
|
+
translated_text=translation_map.get(i, ""),
|
|
173
195
|
source_lang=effective_source,
|
|
174
196
|
target_lang=target_lang,
|
|
175
197
|
speaker=segment.speaker,
|
|
@@ -0,0 +1,304 @@
|
|
|
1
|
+
"""Audio source separation using local Demucs models."""
|
|
2
|
+
|
|
3
|
+
from __future__ import annotations
|
|
4
|
+
|
|
5
|
+
import logging
|
|
6
|
+
from typing import Any
|
|
7
|
+
|
|
8
|
+
from videopython.ai._device import log_device_initialization, release_device_memory, select_device
|
|
9
|
+
from videopython.ai.dubbing.models import SeparatedAudio
|
|
10
|
+
from videopython.base.audio import Audio, AudioMetadata
|
|
11
|
+
|
|
12
|
+
logger = logging.getLogger(__name__)
|
|
13
|
+
|
|
14
|
+
|
|
15
|
+
def _merge_regions(
|
|
16
|
+
regions: list[tuple[float, float]],
|
|
17
|
+
audio_duration: float,
|
|
18
|
+
pad: float = 0.5,
|
|
19
|
+
merge_gap: float = 1.0,
|
|
20
|
+
) -> list[tuple[float, float]]:
|
|
21
|
+
"""Merge overlapping/adjacent (start, end) ranges and pad each side.
|
|
22
|
+
|
|
23
|
+
Args:
|
|
24
|
+
regions: Speech regions in seconds. Order does not matter.
|
|
25
|
+
audio_duration: Total audio duration; output is clamped to ``[0, audio_duration]``.
|
|
26
|
+
pad: Seconds added to each side. Demucs needs context to separate
|
|
27
|
+
cleanly at boundaries; 0.5s avoids clipped onsets/decays.
|
|
28
|
+
merge_gap: Adjacent regions whose padded edges are within this
|
|
29
|
+
many seconds are merged. Avoids running Demucs on very short
|
|
30
|
+
slices (where its temporal context isn't there).
|
|
31
|
+
|
|
32
|
+
Returns:
|
|
33
|
+
Sorted list of non-overlapping (start, end) regions covering the
|
|
34
|
+
speech-bearing portion of the audio.
|
|
35
|
+
"""
|
|
36
|
+
if not regions:
|
|
37
|
+
return []
|
|
38
|
+
|
|
39
|
+
sorted_regions = sorted(regions)
|
|
40
|
+
|
|
41
|
+
merged: list[tuple[float, float]] = []
|
|
42
|
+
for start, end in sorted_regions:
|
|
43
|
+
if end <= start:
|
|
44
|
+
continue
|
|
45
|
+
padded_start = max(0.0, start - pad)
|
|
46
|
+
padded_end = min(audio_duration, end + pad)
|
|
47
|
+
if padded_start >= audio_duration or padded_end <= 0.0:
|
|
48
|
+
continue
|
|
49
|
+
|
|
50
|
+
if merged and padded_start - merged[-1][1] <= merge_gap:
|
|
51
|
+
merged[-1] = (merged[-1][0], max(merged[-1][1], padded_end))
|
|
52
|
+
else:
|
|
53
|
+
merged.append((padded_start, padded_end))
|
|
54
|
+
|
|
55
|
+
return merged
|
|
56
|
+
|
|
57
|
+
|
|
58
|
+
class AudioSeparator:
|
|
59
|
+
"""Separates audio into vocals and background components using Demucs."""
|
|
60
|
+
|
|
61
|
+
SUPPORTED_MODELS: list[str] = ["htdemucs", "htdemucs_ft", "htdemucs_6s", "mdx_extra"]
|
|
62
|
+
STEM_NAMES = ["drums", "bass", "other", "vocals"]
|
|
63
|
+
STEM_NAMES_6S = ["drums", "bass", "other", "vocals", "guitar", "piano"]
|
|
64
|
+
|
|
65
|
+
def __init__(self, model_name: str = "htdemucs", device: str | None = None):
|
|
66
|
+
if model_name not in self.SUPPORTED_MODELS:
|
|
67
|
+
raise ValueError(f"Model '{model_name}' not supported. Supported: {self.SUPPORTED_MODELS}")
|
|
68
|
+
|
|
69
|
+
self.model_name = model_name
|
|
70
|
+
self.device = device
|
|
71
|
+
self._model: Any = None
|
|
72
|
+
|
|
73
|
+
def _init_local(self) -> None:
|
|
74
|
+
"""Initialize local Demucs model."""
|
|
75
|
+
from demucs.pretrained import get_model
|
|
76
|
+
|
|
77
|
+
requested_device = self.device
|
|
78
|
+
device = select_device(self.device, mps_allowed=False)
|
|
79
|
+
|
|
80
|
+
self._model = get_model(self.model_name)
|
|
81
|
+
self._model.to(device)
|
|
82
|
+
self._model.eval()
|
|
83
|
+
self.device = device
|
|
84
|
+
log_device_initialization(
|
|
85
|
+
"AudioSeparator",
|
|
86
|
+
requested_device=requested_device,
|
|
87
|
+
resolved_device=device,
|
|
88
|
+
)
|
|
89
|
+
|
|
90
|
+
def _separate_local(self, audio: Audio) -> SeparatedAudio:
|
|
91
|
+
"""Separate audio using local Demucs model.
|
|
92
|
+
|
|
93
|
+
Keeps the input tensor on CPU and passes ``device=self.device`` to
|
|
94
|
+
``apply_model`` so per-chunk compute runs on GPU while the full
|
|
95
|
+
``(stems, channels, samples)`` output is stored in CPU RAM. For long
|
|
96
|
+
sources this is the difference between OOM-on-GPU and running cleanly:
|
|
97
|
+
a 2h stereo @ 44.1kHz output is ~10 GB — too big for an 8 GB card but
|
|
98
|
+
comfortable on a 32 GB host.
|
|
99
|
+
"""
|
|
100
|
+
import numpy as np
|
|
101
|
+
import torch
|
|
102
|
+
from demucs.apply import apply_model
|
|
103
|
+
|
|
104
|
+
if self._model is None:
|
|
105
|
+
self._init_local()
|
|
106
|
+
|
|
107
|
+
target_sr = self._model.samplerate
|
|
108
|
+
|
|
109
|
+
if audio.metadata.channels == 1:
|
|
110
|
+
audio = audio._to_stereo()
|
|
111
|
+
|
|
112
|
+
if audio.metadata.sample_rate != target_sr:
|
|
113
|
+
audio = audio.resample(target_sr)
|
|
114
|
+
|
|
115
|
+
audio_data = audio.data
|
|
116
|
+
if audio_data.ndim == 1:
|
|
117
|
+
audio_data = np.stack([audio_data, audio_data])
|
|
118
|
+
elif audio_data.ndim == 2:
|
|
119
|
+
audio_data = audio_data.T
|
|
120
|
+
|
|
121
|
+
wav = torch.tensor(audio_data, dtype=torch.float32).unsqueeze(0)
|
|
122
|
+
|
|
123
|
+
with torch.no_grad():
|
|
124
|
+
sources = apply_model(self._model, wav, device=self.device)
|
|
125
|
+
|
|
126
|
+
sources_np = sources[0].cpu().numpy()
|
|
127
|
+
del sources
|
|
128
|
+
|
|
129
|
+
stem_names = self.STEM_NAMES_6S if self.model_name == "htdemucs_6s" else self.STEM_NAMES
|
|
130
|
+
vocals_idx = stem_names.index("vocals")
|
|
131
|
+
non_vocal_indices = [i for i in range(len(stem_names)) if i != vocals_idx]
|
|
132
|
+
|
|
133
|
+
vocals_data = sources_np[vocals_idx].T
|
|
134
|
+
background_data = sources_np[non_vocal_indices].sum(axis=0).T
|
|
135
|
+
del sources_np
|
|
136
|
+
|
|
137
|
+
max_val = np.max(np.abs(background_data))
|
|
138
|
+
if max_val > 1.0:
|
|
139
|
+
background_data /= max_val
|
|
140
|
+
|
|
141
|
+
metadata = AudioMetadata(
|
|
142
|
+
sample_rate=target_sr,
|
|
143
|
+
channels=2,
|
|
144
|
+
sample_width=2,
|
|
145
|
+
duration_seconds=vocals_data.shape[0] / target_sr,
|
|
146
|
+
frame_count=vocals_data.shape[0],
|
|
147
|
+
)
|
|
148
|
+
vocals = Audio(np.ascontiguousarray(vocals_data, dtype=np.float32), metadata)
|
|
149
|
+
background = Audio(np.ascontiguousarray(background_data, dtype=np.float32), metadata)
|
|
150
|
+
|
|
151
|
+
return SeparatedAudio(
|
|
152
|
+
vocals=vocals,
|
|
153
|
+
background=background,
|
|
154
|
+
original=audio,
|
|
155
|
+
music=None,
|
|
156
|
+
effects=None,
|
|
157
|
+
)
|
|
158
|
+
|
|
159
|
+
def separate(self, audio: Audio) -> SeparatedAudio:
|
|
160
|
+
"""Separate audio into vocals and background components."""
|
|
161
|
+
return self._separate_local(audio)
|
|
162
|
+
|
|
163
|
+
def separate_regions(
|
|
164
|
+
self,
|
|
165
|
+
audio: Audio,
|
|
166
|
+
regions: list[tuple[float, float]],
|
|
167
|
+
full_separation_threshold: float = 0.9,
|
|
168
|
+
) -> SeparatedAudio:
|
|
169
|
+
"""Separate only the given (start, end) regions; pass the rest through.
|
|
170
|
+
|
|
171
|
+
Demucs is the slowest stage of the dubbing pipeline. On talk-heavy
|
|
172
|
+
sources (podcasts, interviews) most of the track is speech, but
|
|
173
|
+
long pauses, silence, or music-only stretches don't need vocal
|
|
174
|
+
isolation — there's nothing to isolate. We run Demucs only on the
|
|
175
|
+
speech-bearing regions and treat the rest as pure background.
|
|
176
|
+
|
|
177
|
+
Output is full-length: vocals are silent outside the given
|
|
178
|
+
regions; background is the original audio outside the given
|
|
179
|
+
regions and the Demucs-separated background inside.
|
|
180
|
+
|
|
181
|
+
Args:
|
|
182
|
+
audio: Source audio (typically the full track).
|
|
183
|
+
regions: List of ``(start, end)`` second pairs marking
|
|
184
|
+
speech-bearing portions. Caller is responsible for
|
|
185
|
+
merging/padding (use ``_merge_regions``).
|
|
186
|
+
full_separation_threshold: If the regions cover more than
|
|
187
|
+
this fraction of the audio, fall back to full-track
|
|
188
|
+
``separate()`` since per-region slicing+stitching
|
|
189
|
+
overhead would exceed the savings. Default 0.9.
|
|
190
|
+
|
|
191
|
+
Returns:
|
|
192
|
+
``SeparatedAudio`` with full-length vocals and background.
|
|
193
|
+
"""
|
|
194
|
+
import numpy as np
|
|
195
|
+
|
|
196
|
+
if not regions:
|
|
197
|
+
logger.info("separate_regions: no regions, returning silent vocals over original audio")
|
|
198
|
+
return self._passthrough_separation(audio)
|
|
199
|
+
|
|
200
|
+
total_duration = audio.metadata.duration_seconds
|
|
201
|
+
speech_duration = sum(end - start for start, end in regions)
|
|
202
|
+
if total_duration > 0 and speech_duration / total_duration >= full_separation_threshold:
|
|
203
|
+
logger.info(
|
|
204
|
+
"separate_regions: speech covers %.0f%% of audio (>=%.0f%%), using full-track separation",
|
|
205
|
+
speech_duration / total_duration * 100,
|
|
206
|
+
full_separation_threshold * 100,
|
|
207
|
+
)
|
|
208
|
+
return self._separate_local(audio)
|
|
209
|
+
|
|
210
|
+
logger.info(
|
|
211
|
+
"separate_regions: separating %.1fs of speech across %d region(s) (full duration: %.1fs)",
|
|
212
|
+
speech_duration,
|
|
213
|
+
len(regions),
|
|
214
|
+
total_duration,
|
|
215
|
+
)
|
|
216
|
+
|
|
217
|
+
# Build full-length output buffers. Background defaults to the
|
|
218
|
+
# original audio (so non-speech gaps pass through unchanged); vocals
|
|
219
|
+
# default to silence (no speech to isolate outside the regions).
|
|
220
|
+
# Both are stereo to match the full-track separation contract.
|
|
221
|
+
sr = audio.metadata.sample_rate
|
|
222
|
+
stereo_audio = audio if audio.metadata.channels == 2 else audio._to_stereo()
|
|
223
|
+
|
|
224
|
+
total_samples = len(stereo_audio.data)
|
|
225
|
+
vocals_full = np.zeros((total_samples, 2), dtype=np.float32)
|
|
226
|
+
background_full = stereo_audio.data.astype(np.float32, copy=True)
|
|
227
|
+
|
|
228
|
+
for start, end in regions:
|
|
229
|
+
chunk = audio.slice(start, end)
|
|
230
|
+
separated_chunk = self._separate_local(chunk)
|
|
231
|
+
chunk_vocals = separated_chunk.vocals.data
|
|
232
|
+
chunk_background = separated_chunk.background.data
|
|
233
|
+
|
|
234
|
+
# Demucs operates at its model sample rate (typically 44.1 kHz)
|
|
235
|
+
# and returns stereo. The slice of `audio` we passed in may have
|
|
236
|
+
# been resampled inside _separate_local, so resample the chunk
|
|
237
|
+
# outputs back to the source sample rate before splicing.
|
|
238
|
+
chunk_sr = separated_chunk.vocals.metadata.sample_rate
|
|
239
|
+
if chunk_sr != sr:
|
|
240
|
+
chunk_vocals = separated_chunk.vocals.resample(sr).data
|
|
241
|
+
chunk_background = separated_chunk.background.resample(sr).data
|
|
242
|
+
|
|
243
|
+
start_sample = int(start * sr)
|
|
244
|
+
end_sample = min(start_sample + len(chunk_vocals), total_samples)
|
|
245
|
+
length = end_sample - start_sample
|
|
246
|
+
if length <= 0:
|
|
247
|
+
continue
|
|
248
|
+
|
|
249
|
+
vocals_full[start_sample:end_sample] = chunk_vocals[:length]
|
|
250
|
+
background_full[start_sample:end_sample] = chunk_background[:length]
|
|
251
|
+
|
|
252
|
+
metadata = AudioMetadata(
|
|
253
|
+
sample_rate=sr,
|
|
254
|
+
channels=2,
|
|
255
|
+
sample_width=audio.metadata.sample_width,
|
|
256
|
+
duration_seconds=total_samples / sr,
|
|
257
|
+
frame_count=total_samples,
|
|
258
|
+
)
|
|
259
|
+
vocals = Audio(np.ascontiguousarray(vocals_full, dtype=np.float32), metadata)
|
|
260
|
+
background = Audio(np.ascontiguousarray(background_full, dtype=np.float32), metadata)
|
|
261
|
+
|
|
262
|
+
return SeparatedAudio(
|
|
263
|
+
vocals=vocals,
|
|
264
|
+
background=background,
|
|
265
|
+
original=stereo_audio,
|
|
266
|
+
music=None,
|
|
267
|
+
effects=None,
|
|
268
|
+
)
|
|
269
|
+
|
|
270
|
+
def _passthrough_separation(self, audio: Audio) -> SeparatedAudio:
|
|
271
|
+
"""Return the original audio as background with silent vocals.
|
|
272
|
+
|
|
273
|
+
Used when no speech regions are present — there's nothing to
|
|
274
|
+
separate, so the entire signal is background by definition.
|
|
275
|
+
"""
|
|
276
|
+
import numpy as np
|
|
277
|
+
|
|
278
|
+
stereo_audio = audio if audio.metadata.channels == 2 else audio._to_stereo()
|
|
279
|
+
silent_vocals_data = np.zeros_like(stereo_audio.data, dtype=np.float32)
|
|
280
|
+
vocals = Audio(silent_vocals_data, stereo_audio.metadata)
|
|
281
|
+
|
|
282
|
+
return SeparatedAudio(
|
|
283
|
+
vocals=vocals,
|
|
284
|
+
background=stereo_audio,
|
|
285
|
+
original=stereo_audio,
|
|
286
|
+
music=None,
|
|
287
|
+
effects=None,
|
|
288
|
+
)
|
|
289
|
+
|
|
290
|
+
def extract_vocals(self, audio: Audio) -> Audio:
|
|
291
|
+
"""Convenience method to extract only vocals from audio."""
|
|
292
|
+
return self.separate(audio).vocals
|
|
293
|
+
|
|
294
|
+
def extract_background(self, audio: Audio) -> Audio:
|
|
295
|
+
"""Convenience method to extract only background from audio."""
|
|
296
|
+
return self.separate(audio).background
|
|
297
|
+
|
|
298
|
+
def unload(self) -> None:
|
|
299
|
+
"""Release the Demucs model so the next separate() re-initializes.
|
|
300
|
+
|
|
301
|
+
Used by low-memory dubbing to free VRAM between pipeline stages.
|
|
302
|
+
"""
|
|
303
|
+
self._model = None
|
|
304
|
+
release_device_memory(self.device)
|
|
@@ -1,131 +0,0 @@
|
|
|
1
|
-
"""Audio source separation using local Demucs models."""
|
|
2
|
-
|
|
3
|
-
from __future__ import annotations
|
|
4
|
-
|
|
5
|
-
from typing import Any
|
|
6
|
-
|
|
7
|
-
from videopython.ai._device import log_device_initialization, release_device_memory, select_device
|
|
8
|
-
from videopython.ai.dubbing.models import SeparatedAudio
|
|
9
|
-
from videopython.base.audio import Audio, AudioMetadata
|
|
10
|
-
|
|
11
|
-
|
|
12
|
-
class AudioSeparator:
|
|
13
|
-
"""Separates audio into vocals and background components using Demucs."""
|
|
14
|
-
|
|
15
|
-
SUPPORTED_MODELS: list[str] = ["htdemucs", "htdemucs_ft", "htdemucs_6s", "mdx_extra"]
|
|
16
|
-
STEM_NAMES = ["drums", "bass", "other", "vocals"]
|
|
17
|
-
STEM_NAMES_6S = ["drums", "bass", "other", "vocals", "guitar", "piano"]
|
|
18
|
-
|
|
19
|
-
def __init__(self, model_name: str = "htdemucs", device: str | None = None):
|
|
20
|
-
if model_name not in self.SUPPORTED_MODELS:
|
|
21
|
-
raise ValueError(f"Model '{model_name}' not supported. Supported: {self.SUPPORTED_MODELS}")
|
|
22
|
-
|
|
23
|
-
self.model_name = model_name
|
|
24
|
-
self.device = device
|
|
25
|
-
self._model: Any = None
|
|
26
|
-
|
|
27
|
-
def _init_local(self) -> None:
|
|
28
|
-
"""Initialize local Demucs model."""
|
|
29
|
-
from demucs.pretrained import get_model
|
|
30
|
-
|
|
31
|
-
requested_device = self.device
|
|
32
|
-
device = select_device(self.device, mps_allowed=False)
|
|
33
|
-
|
|
34
|
-
self._model = get_model(self.model_name)
|
|
35
|
-
self._model.to(device)
|
|
36
|
-
self._model.eval()
|
|
37
|
-
self.device = device
|
|
38
|
-
log_device_initialization(
|
|
39
|
-
"AudioSeparator",
|
|
40
|
-
requested_device=requested_device,
|
|
41
|
-
resolved_device=device,
|
|
42
|
-
)
|
|
43
|
-
|
|
44
|
-
def _separate_local(self, audio: Audio) -> SeparatedAudio:
|
|
45
|
-
"""Separate audio using local Demucs model.
|
|
46
|
-
|
|
47
|
-
Keeps the input tensor on CPU and passes ``device=self.device`` to
|
|
48
|
-
``apply_model`` so per-chunk compute runs on GPU while the full
|
|
49
|
-
``(stems, channels, samples)`` output is stored in CPU RAM. For long
|
|
50
|
-
sources this is the difference between OOM-on-GPU and running cleanly:
|
|
51
|
-
a 2h stereo @ 44.1kHz output is ~10 GB — too big for an 8 GB card but
|
|
52
|
-
comfortable on a 32 GB host.
|
|
53
|
-
"""
|
|
54
|
-
import numpy as np
|
|
55
|
-
import torch
|
|
56
|
-
from demucs.apply import apply_model
|
|
57
|
-
|
|
58
|
-
if self._model is None:
|
|
59
|
-
self._init_local()
|
|
60
|
-
|
|
61
|
-
target_sr = self._model.samplerate
|
|
62
|
-
|
|
63
|
-
if audio.metadata.channels == 1:
|
|
64
|
-
audio = audio._to_stereo()
|
|
65
|
-
|
|
66
|
-
if audio.metadata.sample_rate != target_sr:
|
|
67
|
-
audio = audio.resample(target_sr)
|
|
68
|
-
|
|
69
|
-
audio_data = audio.data
|
|
70
|
-
if audio_data.ndim == 1:
|
|
71
|
-
audio_data = np.stack([audio_data, audio_data])
|
|
72
|
-
elif audio_data.ndim == 2:
|
|
73
|
-
audio_data = audio_data.T
|
|
74
|
-
|
|
75
|
-
wav = torch.tensor(audio_data, dtype=torch.float32).unsqueeze(0)
|
|
76
|
-
|
|
77
|
-
with torch.no_grad():
|
|
78
|
-
sources = apply_model(self._model, wav, device=self.device)
|
|
79
|
-
|
|
80
|
-
sources_np = sources[0].cpu().numpy()
|
|
81
|
-
del sources
|
|
82
|
-
|
|
83
|
-
stem_names = self.STEM_NAMES_6S if self.model_name == "htdemucs_6s" else self.STEM_NAMES
|
|
84
|
-
vocals_idx = stem_names.index("vocals")
|
|
85
|
-
non_vocal_indices = [i for i in range(len(stem_names)) if i != vocals_idx]
|
|
86
|
-
|
|
87
|
-
vocals_data = sources_np[vocals_idx].T
|
|
88
|
-
background_data = sources_np[non_vocal_indices].sum(axis=0).T
|
|
89
|
-
del sources_np
|
|
90
|
-
|
|
91
|
-
max_val = np.max(np.abs(background_data))
|
|
92
|
-
if max_val > 1.0:
|
|
93
|
-
background_data /= max_val
|
|
94
|
-
|
|
95
|
-
metadata = AudioMetadata(
|
|
96
|
-
sample_rate=target_sr,
|
|
97
|
-
channels=2,
|
|
98
|
-
sample_width=2,
|
|
99
|
-
duration_seconds=vocals_data.shape[0] / target_sr,
|
|
100
|
-
frame_count=vocals_data.shape[0],
|
|
101
|
-
)
|
|
102
|
-
vocals = Audio(np.ascontiguousarray(vocals_data, dtype=np.float32), metadata)
|
|
103
|
-
background = Audio(np.ascontiguousarray(background_data, dtype=np.float32), metadata)
|
|
104
|
-
|
|
105
|
-
return SeparatedAudio(
|
|
106
|
-
vocals=vocals,
|
|
107
|
-
background=background,
|
|
108
|
-
original=audio,
|
|
109
|
-
music=None,
|
|
110
|
-
effects=None,
|
|
111
|
-
)
|
|
112
|
-
|
|
113
|
-
def separate(self, audio: Audio) -> SeparatedAudio:
|
|
114
|
-
"""Separate audio into vocals and background components."""
|
|
115
|
-
return self._separate_local(audio)
|
|
116
|
-
|
|
117
|
-
def extract_vocals(self, audio: Audio) -> Audio:
|
|
118
|
-
"""Convenience method to extract only vocals from audio."""
|
|
119
|
-
return self.separate(audio).vocals
|
|
120
|
-
|
|
121
|
-
def extract_background(self, audio: Audio) -> Audio:
|
|
122
|
-
"""Convenience method to extract only background from audio."""
|
|
123
|
-
return self.separate(audio).background
|
|
124
|
-
|
|
125
|
-
def unload(self) -> None:
|
|
126
|
-
"""Release the Demucs model so the next separate() re-initializes.
|
|
127
|
-
|
|
128
|
-
Used by low-memory dubbing to free VRAM between pipeline stages.
|
|
129
|
-
"""
|
|
130
|
-
self._model = None
|
|
131
|
-
release_device_memory(self.device)
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|