xinference 1.10.0__py3-none-any.whl → 1.11.0__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of xinference might be problematic. Click here for more details.
- xinference/_version.py +3 -3
- xinference/api/restful_api.py +473 -31
- xinference/client/restful/async_restful_client.py +178 -8
- xinference/client/restful/restful_client.py +151 -3
- xinference/core/supervisor.py +99 -53
- xinference/core/worker.py +10 -0
- xinference/deploy/cmdline.py +15 -0
- xinference/model/audio/core.py +21 -6
- xinference/model/audio/indextts2.py +166 -0
- xinference/model/audio/model_spec.json +58 -21
- xinference/model/image/model_spec.json +159 -90
- xinference/model/image/stable_diffusion/core.py +13 -4
- xinference/model/llm/__init__.py +6 -2
- xinference/model/llm/llm_family.json +1299 -174
- xinference/model/llm/mlx/distributed_models/core.py +41 -0
- xinference/model/llm/mlx/distributed_models/qwen2.py +1 -2
- xinference/model/llm/sglang/core.py +44 -11
- xinference/model/llm/tool_parsers/deepseek_r1_tool_parser.py +94 -32
- xinference/model/llm/tool_parsers/qwen_tool_parser.py +29 -4
- xinference/model/llm/transformers/chatglm.py +3 -0
- xinference/model/llm/transformers/core.py +129 -36
- xinference/model/llm/transformers/multimodal/minicpmv45.py +340 -0
- xinference/model/llm/transformers/multimodal/qwen2_vl.py +34 -8
- xinference/model/llm/transformers/utils.py +23 -0
- xinference/model/llm/utils.py +48 -32
- xinference/model/llm/vllm/core.py +207 -72
- xinference/model/utils.py +74 -31
- xinference/thirdparty/audiotools/__init__.py +10 -0
- xinference/thirdparty/audiotools/core/__init__.py +4 -0
- xinference/thirdparty/audiotools/core/audio_signal.py +1682 -0
- xinference/thirdparty/audiotools/core/display.py +194 -0
- xinference/thirdparty/audiotools/core/dsp.py +390 -0
- xinference/thirdparty/audiotools/core/effects.py +647 -0
- xinference/thirdparty/audiotools/core/ffmpeg.py +211 -0
- xinference/thirdparty/audiotools/core/loudness.py +320 -0
- xinference/thirdparty/audiotools/core/playback.py +252 -0
- xinference/thirdparty/audiotools/core/templates/__init__.py +0 -0
- xinference/thirdparty/audiotools/core/templates/headers.html +322 -0
- xinference/thirdparty/audiotools/core/templates/pandoc.css +407 -0
- xinference/thirdparty/audiotools/core/templates/widget.html +52 -0
- xinference/thirdparty/audiotools/core/util.py +671 -0
- xinference/thirdparty/audiotools/core/whisper.py +97 -0
- xinference/thirdparty/audiotools/data/__init__.py +3 -0
- xinference/thirdparty/audiotools/data/datasets.py +517 -0
- xinference/thirdparty/audiotools/data/preprocess.py +81 -0
- xinference/thirdparty/audiotools/data/transforms.py +1592 -0
- xinference/thirdparty/audiotools/metrics/__init__.py +6 -0
- xinference/thirdparty/audiotools/metrics/distance.py +131 -0
- xinference/thirdparty/audiotools/metrics/quality.py +159 -0
- xinference/thirdparty/audiotools/metrics/spectral.py +247 -0
- xinference/thirdparty/audiotools/ml/__init__.py +5 -0
- xinference/thirdparty/audiotools/ml/accelerator.py +184 -0
- xinference/thirdparty/audiotools/ml/decorators.py +440 -0
- xinference/thirdparty/audiotools/ml/experiment.py +90 -0
- xinference/thirdparty/audiotools/ml/layers/__init__.py +2 -0
- xinference/thirdparty/audiotools/ml/layers/base.py +328 -0
- xinference/thirdparty/audiotools/ml/layers/spectral_gate.py +127 -0
- xinference/thirdparty/audiotools/post.py +140 -0
- xinference/thirdparty/audiotools/preference.py +600 -0
- xinference/thirdparty/fish_speech/fish_speech/text/chn_text_norm/text.py +1 -1
- xinference/thirdparty/indextts/BigVGAN/ECAPA_TDNN.py +656 -0
- xinference/thirdparty/indextts/BigVGAN/__init__.py +0 -0
- xinference/thirdparty/indextts/BigVGAN/activations.py +122 -0
- xinference/thirdparty/indextts/BigVGAN/alias_free_activation/__init__.py +0 -0
- xinference/thirdparty/indextts/BigVGAN/alias_free_activation/cuda/.gitignore +1 -0
- xinference/thirdparty/indextts/BigVGAN/alias_free_activation/cuda/__init__.py +0 -0
- xinference/thirdparty/indextts/BigVGAN/alias_free_activation/cuda/activation1d.py +76 -0
- xinference/thirdparty/indextts/BigVGAN/alias_free_activation/cuda/anti_alias_activation.cpp +23 -0
- xinference/thirdparty/indextts/BigVGAN/alias_free_activation/cuda/anti_alias_activation_cuda.cu +256 -0
- xinference/thirdparty/indextts/BigVGAN/alias_free_activation/cuda/compat.h +29 -0
- xinference/thirdparty/indextts/BigVGAN/alias_free_activation/cuda/load.py +121 -0
- xinference/thirdparty/indextts/BigVGAN/alias_free_activation/cuda/type_shim.h +92 -0
- xinference/thirdparty/indextts/BigVGAN/alias_free_activation/torch/__init__.py +6 -0
- xinference/thirdparty/indextts/BigVGAN/alias_free_activation/torch/act.py +31 -0
- xinference/thirdparty/indextts/BigVGAN/alias_free_activation/torch/filter.py +102 -0
- xinference/thirdparty/indextts/BigVGAN/alias_free_activation/torch/resample.py +58 -0
- xinference/thirdparty/indextts/BigVGAN/alias_free_torch/__init__.py +6 -0
- xinference/thirdparty/indextts/BigVGAN/alias_free_torch/act.py +29 -0
- xinference/thirdparty/indextts/BigVGAN/alias_free_torch/filter.py +96 -0
- xinference/thirdparty/indextts/BigVGAN/alias_free_torch/resample.py +49 -0
- xinference/thirdparty/indextts/BigVGAN/bigvgan.py +534 -0
- xinference/thirdparty/indextts/BigVGAN/models.py +451 -0
- xinference/thirdparty/indextts/BigVGAN/nnet/CNN.py +546 -0
- xinference/thirdparty/indextts/BigVGAN/nnet/__init__.py +0 -0
- xinference/thirdparty/indextts/BigVGAN/nnet/linear.py +89 -0
- xinference/thirdparty/indextts/BigVGAN/nnet/normalization.py +670 -0
- xinference/thirdparty/indextts/BigVGAN/utils.py +101 -0
- xinference/thirdparty/indextts/__init__.py +0 -0
- xinference/thirdparty/indextts/cli.py +65 -0
- xinference/thirdparty/indextts/gpt/__init__.py +0 -0
- xinference/thirdparty/indextts/gpt/conformer/__init__.py +0 -0
- xinference/thirdparty/indextts/gpt/conformer/attention.py +312 -0
- xinference/thirdparty/indextts/gpt/conformer/embedding.py +163 -0
- xinference/thirdparty/indextts/gpt/conformer/subsampling.py +348 -0
- xinference/thirdparty/indextts/gpt/conformer_encoder.py +520 -0
- xinference/thirdparty/indextts/gpt/model.py +713 -0
- xinference/thirdparty/indextts/gpt/model_v2.py +747 -0
- xinference/thirdparty/indextts/gpt/perceiver.py +317 -0
- xinference/thirdparty/indextts/gpt/transformers_beam_search.py +1013 -0
- xinference/thirdparty/indextts/gpt/transformers_generation_utils.py +4747 -0
- xinference/thirdparty/indextts/gpt/transformers_gpt2.py +1878 -0
- xinference/thirdparty/indextts/gpt/transformers_modeling_utils.py +5525 -0
- xinference/thirdparty/indextts/infer.py +690 -0
- xinference/thirdparty/indextts/infer_v2.py +739 -0
- xinference/thirdparty/indextts/s2mel/dac/__init__.py +16 -0
- xinference/thirdparty/indextts/s2mel/dac/__main__.py +36 -0
- xinference/thirdparty/indextts/s2mel/dac/model/__init__.py +4 -0
- xinference/thirdparty/indextts/s2mel/dac/model/base.py +294 -0
- xinference/thirdparty/indextts/s2mel/dac/model/dac.py +400 -0
- xinference/thirdparty/indextts/s2mel/dac/model/discriminator.py +228 -0
- xinference/thirdparty/indextts/s2mel/dac/model/encodec.py +320 -0
- xinference/thirdparty/indextts/s2mel/dac/nn/__init__.py +3 -0
- xinference/thirdparty/indextts/s2mel/dac/nn/layers.py +33 -0
- xinference/thirdparty/indextts/s2mel/dac/nn/loss.py +368 -0
- xinference/thirdparty/indextts/s2mel/dac/nn/quantize.py +339 -0
- xinference/thirdparty/indextts/s2mel/dac/utils/__init__.py +123 -0
- xinference/thirdparty/indextts/s2mel/dac/utils/decode.py +95 -0
- xinference/thirdparty/indextts/s2mel/dac/utils/encode.py +94 -0
- xinference/thirdparty/indextts/s2mel/hf_utils.py +12 -0
- xinference/thirdparty/indextts/s2mel/modules/alias_free_torch/__init__.py +5 -0
- xinference/thirdparty/indextts/s2mel/modules/alias_free_torch/act.py +29 -0
- xinference/thirdparty/indextts/s2mel/modules/alias_free_torch/filter.py +96 -0
- xinference/thirdparty/indextts/s2mel/modules/alias_free_torch/resample.py +57 -0
- xinference/thirdparty/indextts/s2mel/modules/audio.py +82 -0
- xinference/thirdparty/indextts/s2mel/modules/bigvgan/activations.py +120 -0
- xinference/thirdparty/indextts/s2mel/modules/bigvgan/alias_free_activation/cuda/__init__.py +0 -0
- xinference/thirdparty/indextts/s2mel/modules/bigvgan/alias_free_activation/cuda/activation1d.py +77 -0
- xinference/thirdparty/indextts/s2mel/modules/bigvgan/alias_free_activation/cuda/anti_alias_activation.cpp +23 -0
- xinference/thirdparty/indextts/s2mel/modules/bigvgan/alias_free_activation/cuda/anti_alias_activation_cuda.cu +246 -0
- xinference/thirdparty/indextts/s2mel/modules/bigvgan/alias_free_activation/cuda/compat.h +29 -0
- xinference/thirdparty/indextts/s2mel/modules/bigvgan/alias_free_activation/cuda/load.py +86 -0
- xinference/thirdparty/indextts/s2mel/modules/bigvgan/alias_free_activation/cuda/type_shim.h +92 -0
- xinference/thirdparty/indextts/s2mel/modules/bigvgan/alias_free_activation/torch/__init__.py +6 -0
- xinference/thirdparty/indextts/s2mel/modules/bigvgan/alias_free_activation/torch/act.py +30 -0
- xinference/thirdparty/indextts/s2mel/modules/bigvgan/alias_free_activation/torch/filter.py +101 -0
- xinference/thirdparty/indextts/s2mel/modules/bigvgan/alias_free_activation/torch/resample.py +58 -0
- xinference/thirdparty/indextts/s2mel/modules/bigvgan/bigvgan.py +492 -0
- xinference/thirdparty/indextts/s2mel/modules/bigvgan/config.json +63 -0
- xinference/thirdparty/indextts/s2mel/modules/bigvgan/env.py +18 -0
- xinference/thirdparty/indextts/s2mel/modules/bigvgan/meldataset.py +354 -0
- xinference/thirdparty/indextts/s2mel/modules/bigvgan/utils.py +99 -0
- xinference/thirdparty/indextts/s2mel/modules/campplus/DTDNN.py +115 -0
- xinference/thirdparty/indextts/s2mel/modules/campplus/classifier.py +70 -0
- xinference/thirdparty/indextts/s2mel/modules/campplus/layers.py +253 -0
- xinference/thirdparty/indextts/s2mel/modules/commons.py +632 -0
- xinference/thirdparty/indextts/s2mel/modules/diffusion_transformer.py +257 -0
- xinference/thirdparty/indextts/s2mel/modules/encodec.py +292 -0
- xinference/thirdparty/indextts/s2mel/modules/flow_matching.py +171 -0
- xinference/thirdparty/indextts/s2mel/modules/gpt_fast/generate.py +436 -0
- xinference/thirdparty/indextts/s2mel/modules/gpt_fast/model.py +360 -0
- xinference/thirdparty/indextts/s2mel/modules/gpt_fast/quantize.py +622 -0
- xinference/thirdparty/indextts/s2mel/modules/hifigan/f0_predictor.py +55 -0
- xinference/thirdparty/indextts/s2mel/modules/hifigan/generator.py +454 -0
- xinference/thirdparty/indextts/s2mel/modules/layers.py +354 -0
- xinference/thirdparty/indextts/s2mel/modules/length_regulator.py +141 -0
- xinference/thirdparty/indextts/s2mel/modules/openvoice/__init__.py +0 -0
- xinference/thirdparty/indextts/s2mel/modules/openvoice/api.py +186 -0
- xinference/thirdparty/indextts/s2mel/modules/openvoice/attentions.py +465 -0
- xinference/thirdparty/indextts/s2mel/modules/openvoice/checkpoints_v2/converter/config.json +57 -0
- xinference/thirdparty/indextts/s2mel/modules/openvoice/commons.py +160 -0
- xinference/thirdparty/indextts/s2mel/modules/openvoice/mel_processing.py +183 -0
- xinference/thirdparty/indextts/s2mel/modules/openvoice/models.py +499 -0
- xinference/thirdparty/indextts/s2mel/modules/openvoice/modules.py +598 -0
- xinference/thirdparty/indextts/s2mel/modules/openvoice/openvoice_app.py +275 -0
- xinference/thirdparty/indextts/s2mel/modules/openvoice/se_extractor.py +153 -0
- xinference/thirdparty/indextts/s2mel/modules/openvoice/transforms.py +209 -0
- xinference/thirdparty/indextts/s2mel/modules/openvoice/utils.py +194 -0
- xinference/thirdparty/indextts/s2mel/modules/quantize.py +229 -0
- xinference/thirdparty/indextts/s2mel/modules/rmvpe.py +631 -0
- xinference/thirdparty/indextts/s2mel/modules/vocos/__init__.py +4 -0
- xinference/thirdparty/indextts/s2mel/modules/vocos/heads.py +164 -0
- xinference/thirdparty/indextts/s2mel/modules/vocos/helpers.py +71 -0
- xinference/thirdparty/indextts/s2mel/modules/vocos/loss.py +114 -0
- xinference/thirdparty/indextts/s2mel/modules/vocos/models.py +118 -0
- xinference/thirdparty/indextts/s2mel/modules/vocos/modules.py +213 -0
- xinference/thirdparty/indextts/s2mel/modules/vocos/pretrained.py +51 -0
- xinference/thirdparty/indextts/s2mel/modules/vocos/spectral_ops.py +192 -0
- xinference/thirdparty/indextts/s2mel/modules/wavenet.py +174 -0
- xinference/thirdparty/indextts/s2mel/optimizers.py +96 -0
- xinference/thirdparty/indextts/s2mel/wav2vecbert_extract.py +148 -0
- xinference/thirdparty/indextts/utils/__init__.py +0 -0
- xinference/thirdparty/indextts/utils/arch_util.py +120 -0
- xinference/thirdparty/indextts/utils/checkpoint.py +34 -0
- xinference/thirdparty/indextts/utils/common.py +121 -0
- xinference/thirdparty/indextts/utils/feature_extractors.py +50 -0
- xinference/thirdparty/indextts/utils/front.py +536 -0
- xinference/thirdparty/indextts/utils/maskgct/models/codec/__init__.py +0 -0
- xinference/thirdparty/indextts/utils/maskgct/models/codec/amphion_codec/codec.py +427 -0
- xinference/thirdparty/indextts/utils/maskgct/models/codec/amphion_codec/quantize/__init__.py +11 -0
- xinference/thirdparty/indextts/utils/maskgct/models/codec/amphion_codec/quantize/factorized_vector_quantize.py +150 -0
- xinference/thirdparty/indextts/utils/maskgct/models/codec/amphion_codec/quantize/lookup_free_quantize.py +77 -0
- xinference/thirdparty/indextts/utils/maskgct/models/codec/amphion_codec/quantize/residual_vq.py +177 -0
- xinference/thirdparty/indextts/utils/maskgct/models/codec/amphion_codec/quantize/vector_quantize.py +401 -0
- xinference/thirdparty/indextts/utils/maskgct/models/codec/amphion_codec/vocos.py +881 -0
- xinference/thirdparty/indextts/utils/maskgct/models/codec/codec_dataset.py +264 -0
- xinference/thirdparty/indextts/utils/maskgct/models/codec/codec_inference.py +515 -0
- xinference/thirdparty/indextts/utils/maskgct/models/codec/codec_sampler.py +126 -0
- xinference/thirdparty/indextts/utils/maskgct/models/codec/codec_trainer.py +166 -0
- xinference/thirdparty/indextts/utils/maskgct/models/codec/facodec/__init__.py +0 -0
- xinference/thirdparty/indextts/utils/maskgct/models/codec/facodec/alias_free_torch/__init__.py +5 -0
- xinference/thirdparty/indextts/utils/maskgct/models/codec/facodec/alias_free_torch/act.py +29 -0
- xinference/thirdparty/indextts/utils/maskgct/models/codec/facodec/alias_free_torch/filter.py +96 -0
- xinference/thirdparty/indextts/utils/maskgct/models/codec/facodec/alias_free_torch/resample.py +57 -0
- xinference/thirdparty/indextts/utils/maskgct/models/codec/facodec/facodec_dataset.py +98 -0
- xinference/thirdparty/indextts/utils/maskgct/models/codec/facodec/facodec_inference.py +137 -0
- xinference/thirdparty/indextts/utils/maskgct/models/codec/facodec/facodec_trainer.py +776 -0
- xinference/thirdparty/indextts/utils/maskgct/models/codec/facodec/modules/JDC/__init__.py +1 -0
- xinference/thirdparty/indextts/utils/maskgct/models/codec/facodec/modules/JDC/bst.t7 +0 -0
- xinference/thirdparty/indextts/utils/maskgct/models/codec/facodec/modules/JDC/model.py +219 -0
- xinference/thirdparty/indextts/utils/maskgct/models/codec/facodec/modules/attentions.py +437 -0
- xinference/thirdparty/indextts/utils/maskgct/models/codec/facodec/modules/commons.py +331 -0
- xinference/thirdparty/indextts/utils/maskgct/models/codec/facodec/modules/gradient_reversal.py +35 -0
- xinference/thirdparty/indextts/utils/maskgct/models/codec/facodec/modules/layers.py +460 -0
- xinference/thirdparty/indextts/utils/maskgct/models/codec/facodec/modules/quantize.py +741 -0
- xinference/thirdparty/indextts/utils/maskgct/models/codec/facodec/modules/style_encoder.py +110 -0
- xinference/thirdparty/indextts/utils/maskgct/models/codec/facodec/modules/wavenet.py +224 -0
- xinference/thirdparty/indextts/utils/maskgct/models/codec/facodec/optimizer.py +104 -0
- xinference/thirdparty/indextts/utils/maskgct/models/codec/kmeans/repcodec_model.py +210 -0
- xinference/thirdparty/indextts/utils/maskgct/models/codec/kmeans/vocos.py +850 -0
- xinference/thirdparty/indextts/utils/maskgct/models/codec/melvqgan/melspec.py +108 -0
- xinference/thirdparty/indextts/utils/maskgct/models/codec/ns3_codec/README.md +216 -0
- xinference/thirdparty/indextts/utils/maskgct/models/codec/ns3_codec/__init__.py +6 -0
- xinference/thirdparty/indextts/utils/maskgct/models/codec/ns3_codec/alias_free_torch/__init__.py +5 -0
- xinference/thirdparty/indextts/utils/maskgct/models/codec/ns3_codec/alias_free_torch/act.py +29 -0
- xinference/thirdparty/indextts/utils/maskgct/models/codec/ns3_codec/alias_free_torch/filter.py +96 -0
- xinference/thirdparty/indextts/utils/maskgct/models/codec/ns3_codec/alias_free_torch/resample.py +57 -0
- xinference/thirdparty/indextts/utils/maskgct/models/codec/ns3_codec/facodec.py +1222 -0
- xinference/thirdparty/indextts/utils/maskgct/models/codec/ns3_codec/gradient_reversal.py +35 -0
- xinference/thirdparty/indextts/utils/maskgct/models/codec/ns3_codec/melspec.py +102 -0
- xinference/thirdparty/indextts/utils/maskgct/models/codec/ns3_codec/quantize/__init__.py +7 -0
- xinference/thirdparty/indextts/utils/maskgct/models/codec/ns3_codec/quantize/fvq.py +116 -0
- xinference/thirdparty/indextts/utils/maskgct/models/codec/ns3_codec/quantize/rvq.py +87 -0
- xinference/thirdparty/indextts/utils/maskgct/models/codec/ns3_codec/transformer.py +234 -0
- xinference/thirdparty/indextts/utils/maskgct/models/codec/speechtokenizer/model.py +184 -0
- xinference/thirdparty/indextts/utils/maskgct/models/codec/speechtokenizer/modules/__init__.py +27 -0
- xinference/thirdparty/indextts/utils/maskgct/models/codec/speechtokenizer/modules/conv.py +346 -0
- xinference/thirdparty/indextts/utils/maskgct/models/codec/speechtokenizer/modules/lstm.py +46 -0
- xinference/thirdparty/indextts/utils/maskgct/models/codec/speechtokenizer/modules/norm.py +37 -0
- xinference/thirdparty/indextts/utils/maskgct/models/codec/speechtokenizer/modules/quantization/__init__.py +14 -0
- xinference/thirdparty/indextts/utils/maskgct/models/codec/speechtokenizer/modules/quantization/ac.py +317 -0
- xinference/thirdparty/indextts/utils/maskgct/models/codec/speechtokenizer/modules/quantization/core_vq.py +388 -0
- xinference/thirdparty/indextts/utils/maskgct/models/codec/speechtokenizer/modules/quantization/distrib.py +135 -0
- xinference/thirdparty/indextts/utils/maskgct/models/codec/speechtokenizer/modules/quantization/vq.py +125 -0
- xinference/thirdparty/indextts/utils/maskgct/models/codec/speechtokenizer/modules/seanet.py +414 -0
- xinference/thirdparty/indextts/utils/maskgct/models/codec/vevo/vevo_repcodec.py +592 -0
- xinference/thirdparty/indextts/utils/maskgct/models/tts/maskgct/ckpt/wav2vec2bert_stats.pt +0 -0
- xinference/thirdparty/indextts/utils/maskgct/models/tts/maskgct/llama_nar.py +650 -0
- xinference/thirdparty/indextts/utils/maskgct/models/tts/maskgct/maskgct_s2a.py +503 -0
- xinference/thirdparty/indextts/utils/maskgct_utils.py +259 -0
- xinference/thirdparty/indextts/utils/text_utils.py +41 -0
- xinference/thirdparty/indextts/utils/typical_sampling.py +30 -0
- xinference/thirdparty/indextts/utils/utils.py +93 -0
- xinference/thirdparty/indextts/utils/webui_utils.py +42 -0
- xinference/thirdparty/indextts/utils/xtransformers.py +1247 -0
- xinference/thirdparty/indextts/vqvae/__init__.py +0 -0
- xinference/thirdparty/indextts/vqvae/xtts_dvae.py +395 -0
- xinference/thirdparty/melo/text/chinese_mix.py +2 -2
- xinference/types.py +9 -0
- xinference/ui/gradio/media_interface.py +66 -8
- xinference/ui/web/ui/build/asset-manifest.json +6 -6
- xinference/ui/web/ui/build/index.html +1 -1
- xinference/ui/web/ui/build/static/css/main.5ea97072.css +2 -0
- xinference/ui/web/ui/build/static/css/main.5ea97072.css.map +1 -0
- xinference/ui/web/ui/build/static/js/main.45e78536.js +3 -0
- xinference/ui/web/ui/build/static/js/{main.1086c759.js.LICENSE.txt → main.45e78536.js.LICENSE.txt} +0 -7
- xinference/ui/web/ui/build/static/js/main.45e78536.js.map +1 -0
- xinference/ui/web/ui/node_modules/.cache/babel-loader/089c38df5f52348d212ed868dda5c518a42e0c2762caed4175487c0405830c35.json +1 -0
- xinference/ui/web/ui/node_modules/.cache/babel-loader/2b6e3a5b6eb2c5c5f2d007e68cd46c372721cd52bf63508adcdb21ecf79241d8.json +1 -0
- xinference/ui/web/ui/node_modules/.cache/babel-loader/2d887825fd07a56f872eda4420da25fba0b5b62a23bdcc6c6da1a5281887f618.json +1 -0
- xinference/ui/web/ui/node_modules/.cache/babel-loader/4001f9c3e64e73a4f2158826650c174a59d5e3f89ddecddf17cbb6bb688cc4ca.json +1 -0
- xinference/ui/web/ui/node_modules/.cache/babel-loader/4a7018a69e6b7f90fc313248c2aa86f2a8f1eb1db120df586047a8023549b44b.json +1 -0
- xinference/ui/web/ui/node_modules/.cache/babel-loader/64b12aaa1c1d1bf53820ada8a63769067c0ccc5aab46b32348eb1917ae7f2a11.json +1 -0
- xinference/ui/web/ui/node_modules/.cache/babel-loader/7275b67c78ec76ce38a686bb8a576d8c9cecf54e1573614c84859d538efb9be5.json +1 -0
- xinference/ui/web/ui/node_modules/.cache/babel-loader/a68b6ee3b31eadc051fb95ce8f8ccb9c2e8b52c60f290dbab545a1917e065282.json +1 -0
- xinference/ui/web/ui/node_modules/.cache/babel-loader/ae8771cc37693feb160fa8727231312a0c54ef2d1d1ca893be568cd70016ca7e.json +1 -0
- xinference/ui/web/ui/node_modules/.cache/babel-loader/bb4e8722d2d41d87f1fce3661bc8937bffe9448e231fc5f0462630849e851592.json +1 -0
- xinference/ui/web/ui/node_modules/.cache/babel-loader/be6aada1ee4adc2bbf65dbe56d17db32bb3b5478be05d6b527805a8ba6cfb2b9.json +1 -0
- xinference/ui/web/ui/node_modules/.cache/babel-loader/de91c352653c233cf0cb6674e6e04049a44fd0e1156560de65d5c4620521391e.json +1 -0
- xinference/ui/web/ui/node_modules/.cache/babel-loader/e85f7002fc325c83b9c9cd8a1619e5b3ebc701d30e811afc284b88e6ae710cb5.json +1 -0
- xinference/ui/web/ui/node_modules/.cache/babel-loader/e8b603c78944bf3d213639078bfe155ff5c0dfa4048a93cbb967cad6a4eb4ff3.json +1 -0
- xinference/ui/web/ui/node_modules/.cache/babel-loader/ea2a26361204e70cf1018d6990fb6354bed82b3ac69690391e0f100385e7abb7.json +1 -0
- xinference/ui/web/ui/node_modules/.cache/babel-loader/f05535160a508b2a312de546a6de234776c613db276479ea4253c0b1bdeeb7d6.json +1 -0
- xinference/ui/web/ui/node_modules/.cache/babel-loader/f09ba9e11106bd59a0de10cc85c55084097729dcab575f43dfcf07375961ed87.json +1 -0
- xinference/ui/web/ui/node_modules/.package-lock.json +0 -33
- xinference/ui/web/ui/package-lock.json +0 -34
- xinference/ui/web/ui/package.json +0 -1
- xinference/ui/web/ui/src/locales/en.json +9 -3
- xinference/ui/web/ui/src/locales/ja.json +9 -3
- xinference/ui/web/ui/src/locales/ko.json +9 -3
- xinference/ui/web/ui/src/locales/zh.json +9 -3
- {xinference-1.10.0.dist-info → xinference-1.11.0.dist-info}/METADATA +24 -6
- {xinference-1.10.0.dist-info → xinference-1.11.0.dist-info}/RECORD +296 -77
- xinference/ui/web/ui/build/static/css/main.013f296b.css +0 -2
- xinference/ui/web/ui/build/static/css/main.013f296b.css.map +0 -1
- xinference/ui/web/ui/build/static/js/main.1086c759.js +0 -3
- xinference/ui/web/ui/build/static/js/main.1086c759.js.map +0 -1
- xinference/ui/web/ui/node_modules/.cache/babel-loader/0b0f77000cc1b482ca091cfbcae511dfe02f08916971645fad21d0b1234d04a2.json +0 -1
- xinference/ui/web/ui/node_modules/.cache/babel-loader/1c5f8ff423a7c9202bea60b15680f04b1e9964b445b0da3f86c6ff70cf24e797.json +0 -1
- xinference/ui/web/ui/node_modules/.cache/babel-loader/44ce7993e344980e3ed4f13e8f69237d4a5dfc60e37ca6b54f51f8ee1357bd67.json +0 -1
- xinference/ui/web/ui/node_modules/.cache/babel-loader/4aec1cc414ac3ebb3481d3d915e4db597d9127de813291346eacb8554ab170d4.json +0 -1
- xinference/ui/web/ui/node_modules/.cache/babel-loader/644cfec52f3c57a6e222ce60f112237a1efefe9835efd9aad857a685f53d8eed.json +0 -1
- xinference/ui/web/ui/node_modules/.cache/babel-loader/663436f72af53fe0d72394f56d003fa4e0bba489e5bb4e483fd34b00f84637f7.json +0 -1
- xinference/ui/web/ui/node_modules/.cache/babel-loader/69db82ca9bfe27fe417cc6cf2b1716b09be9c6f0cd198530f12bfc60e801bbcf.json +0 -1
- xinference/ui/web/ui/node_modules/.cache/babel-loader/85087e27618d740c236bf159f30e0219db443ab55f0997388eed5fde6f9e90cc.json +0 -1
- xinference/ui/web/ui/node_modules/.cache/babel-loader/88b07838348864aa86c672be3bbca1e9f58f6f3a2881b32070ec27f4e7b449d1.json +0 -1
- xinference/ui/web/ui/node_modules/.cache/babel-loader/8b8cd408ccfbe115acef27ccfa5b233da8597131a2a5712add13e1e4d5d4504b.json +0 -1
- xinference/ui/web/ui/node_modules/.cache/babel-loader/a23824fe746b9c6ca5eee9159b5764d1ff1653c1d856288c0f75c742bbb0023b.json +0 -1
- xinference/ui/web/ui/node_modules/.cache/babel-loader/a3eb18af328280b139693c9092dff2a0ef8c9a967e6c8956ceee0996611f1984.json +0 -1
- xinference/ui/web/ui/node_modules/.cache/babel-loader/bc1aacc65a102db325ca61bcd2f681e1ae22c36a1f1d98a6ff5e4ad49dc7544f.json +0 -1
- xinference/ui/web/ui/node_modules/.cache/babel-loader/c682fd521747c19dae437d83ce3235a306ce6b68e24a117bc57c27ebb8d1f1ca.json +0 -1
- xinference/ui/web/ui/node_modules/.cache/babel-loader/d5c224be7081f18cba1678b7874a9782eba895df004874ff8f243f94ba79942a.json +0 -1
- xinference/ui/web/ui/node_modules/.cache/babel-loader/f7f18bfb539b036a6a342176dd98a85df5057a884a8da978d679f2a0264883d0.json +0 -1
- xinference/ui/web/ui/node_modules/clipboard/.babelrc.json +0 -11
- xinference/ui/web/ui/node_modules/clipboard/.eslintrc.json +0 -24
- xinference/ui/web/ui/node_modules/clipboard/.prettierrc.json +0 -9
- xinference/ui/web/ui/node_modules/clipboard/bower.json +0 -18
- xinference/ui/web/ui/node_modules/clipboard/composer.json +0 -25
- xinference/ui/web/ui/node_modules/clipboard/package.json +0 -63
- xinference/ui/web/ui/node_modules/delegate/package.json +0 -31
- xinference/ui/web/ui/node_modules/good-listener/bower.json +0 -11
- xinference/ui/web/ui/node_modules/good-listener/package.json +0 -35
- xinference/ui/web/ui/node_modules/select/bower.json +0 -13
- xinference/ui/web/ui/node_modules/select/package.json +0 -29
- xinference/ui/web/ui/node_modules/tiny-emitter/package.json +0 -53
- {xinference-1.10.0.dist-info → xinference-1.11.0.dist-info}/WHEEL +0 -0
- {xinference-1.10.0.dist-info → xinference-1.11.0.dist-info}/entry_points.txt +0 -0
- {xinference-1.10.0.dist-info → xinference-1.11.0.dist-info}/licenses/LICENSE +0 -0
- {xinference-1.10.0.dist-info → xinference-1.11.0.dist-info}/top_level.txt +0 -0
|
@@ -0,0 +1,108 @@
|
|
|
1
|
+
# Copyright (c) 2023 Amphion.
|
|
2
|
+
#
|
|
3
|
+
# This source code is licensed under the MIT license found in the
|
|
4
|
+
# LICENSE file in the root directory of this source tree.
|
|
5
|
+
|
|
6
|
+
import torch
|
|
7
|
+
import pyworld as pw
|
|
8
|
+
import numpy as np
|
|
9
|
+
import soundfile as sf
|
|
10
|
+
import os
|
|
11
|
+
from torchaudio.functional import pitch_shift
|
|
12
|
+
import librosa
|
|
13
|
+
from librosa.filters import mel as librosa_mel_fn
|
|
14
|
+
import torch.nn as nn
|
|
15
|
+
import torch.nn.functional as F
|
|
16
|
+
import tqdm
|
|
17
|
+
|
|
18
|
+
|
|
19
|
+
def dynamic_range_compression(x, C=1, clip_val=1e-5):
|
|
20
|
+
return np.log(np.clip(x, a_min=clip_val, a_max=None) * C)
|
|
21
|
+
|
|
22
|
+
|
|
23
|
+
def dynamic_range_decompression(x, C=1):
|
|
24
|
+
return np.exp(x) / C
|
|
25
|
+
|
|
26
|
+
|
|
27
|
+
def dynamic_range_compression_torch(x, C=1, clip_val=1e-5):
|
|
28
|
+
return torch.log(torch.clamp(x, min=clip_val) * C)
|
|
29
|
+
|
|
30
|
+
|
|
31
|
+
def dynamic_range_decompression_torch(x, C=1):
|
|
32
|
+
return torch.exp(x) / C
|
|
33
|
+
|
|
34
|
+
|
|
35
|
+
def spectral_normalize_torch(magnitudes):
|
|
36
|
+
output = dynamic_range_compression_torch(magnitudes)
|
|
37
|
+
return output
|
|
38
|
+
|
|
39
|
+
|
|
40
|
+
def spectral_de_normalize_torch(magnitudes):
|
|
41
|
+
output = dynamic_range_decompression_torch(magnitudes)
|
|
42
|
+
return output
|
|
43
|
+
|
|
44
|
+
|
|
45
|
+
class MelSpectrogram(nn.Module):
|
|
46
|
+
def __init__(
|
|
47
|
+
self,
|
|
48
|
+
n_fft,
|
|
49
|
+
num_mels,
|
|
50
|
+
sampling_rate,
|
|
51
|
+
hop_size,
|
|
52
|
+
win_size,
|
|
53
|
+
fmin,
|
|
54
|
+
fmax,
|
|
55
|
+
center=False,
|
|
56
|
+
):
|
|
57
|
+
super(MelSpectrogram, self).__init__()
|
|
58
|
+
self.n_fft = n_fft
|
|
59
|
+
self.hop_size = hop_size
|
|
60
|
+
self.win_size = win_size
|
|
61
|
+
self.sampling_rate = sampling_rate
|
|
62
|
+
self.num_mels = num_mels
|
|
63
|
+
self.fmin = fmin
|
|
64
|
+
self.fmax = fmax
|
|
65
|
+
self.center = center
|
|
66
|
+
|
|
67
|
+
mel_basis = {}
|
|
68
|
+
hann_window = {}
|
|
69
|
+
|
|
70
|
+
mel = librosa_mel_fn(
|
|
71
|
+
sr=sampling_rate, n_fft=n_fft, n_mels=num_mels, fmin=fmin, fmax=fmax
|
|
72
|
+
)
|
|
73
|
+
mel_basis = torch.from_numpy(mel).float()
|
|
74
|
+
hann_window = torch.hann_window(win_size)
|
|
75
|
+
|
|
76
|
+
self.register_buffer("mel_basis", mel_basis)
|
|
77
|
+
self.register_buffer("hann_window", hann_window)
|
|
78
|
+
|
|
79
|
+
def forward(self, y):
|
|
80
|
+
y = torch.nn.functional.pad(
|
|
81
|
+
y.unsqueeze(1),
|
|
82
|
+
(
|
|
83
|
+
int((self.n_fft - self.hop_size) / 2),
|
|
84
|
+
int((self.n_fft - self.hop_size) / 2),
|
|
85
|
+
),
|
|
86
|
+
mode="reflect",
|
|
87
|
+
)
|
|
88
|
+
y = y.squeeze(1)
|
|
89
|
+
spec = torch.stft(
|
|
90
|
+
y,
|
|
91
|
+
self.n_fft,
|
|
92
|
+
hop_length=self.hop_size,
|
|
93
|
+
win_length=self.win_size,
|
|
94
|
+
window=self.hann_window,
|
|
95
|
+
center=self.center,
|
|
96
|
+
pad_mode="reflect",
|
|
97
|
+
normalized=False,
|
|
98
|
+
onesided=True,
|
|
99
|
+
return_complex=True,
|
|
100
|
+
)
|
|
101
|
+
spec = torch.view_as_real(spec)
|
|
102
|
+
|
|
103
|
+
spec = torch.sqrt(spec.pow(2).sum(-1) + (1e-9))
|
|
104
|
+
|
|
105
|
+
spec = torch.matmul(self.mel_basis, spec)
|
|
106
|
+
spec = spectral_normalize_torch(spec)
|
|
107
|
+
|
|
108
|
+
return spec
|
|
@@ -0,0 +1,216 @@
|
|
|
1
|
+
## FACodec: Speech Codec with Attribute Factorization used for NaturalSpeech 3
|
|
2
|
+
|
|
3
|
+
[](https://arxiv.org/pdf/2403.03100.pdf)
|
|
4
|
+
[](https://speechresearch.github.io/naturalspeech3/)
|
|
5
|
+
[](https://huggingface.co/amphion/naturalspeech3_facodec)
|
|
6
|
+
[](https://huggingface.co/spaces/amphion/naturalspeech3_facodec)
|
|
7
|
+
|
|
8
|
+
## Overview
|
|
9
|
+
|
|
10
|
+
FACodec is a core component of the advanced text-to-speech (TTS) model NaturalSpeech 3. FACodec converts complex speech waveform into disentangled subspaces representing speech attributes of content, prosody, timbre, and acoustic details and reconstruct high-quality speech waveform from these attributes. FACodec decomposes complex speech into subspaces representing different attributes, thus simplifying the modeling of speech representation.
|
|
11
|
+
|
|
12
|
+
Research can use FACodec to develop different modes of TTS models, such as non-autoregressive based discrete diffusion (NaturalSpeech 3) or autoregressive models (like VALL-E).
|
|
13
|
+
|
|
14
|
+
<br>
|
|
15
|
+
<div align="center">
|
|
16
|
+
<img src="../../../imgs/ns3/ns3_overview.png" width="65%">
|
|
17
|
+
</div>
|
|
18
|
+
<br>
|
|
19
|
+
|
|
20
|
+
<br>
|
|
21
|
+
<div align="center">
|
|
22
|
+
<img src="../../../imgs/ns3/ns3_facodec.png" width="100%">
|
|
23
|
+
</div>
|
|
24
|
+
<br>
|
|
25
|
+
|
|
26
|
+
## Useage
|
|
27
|
+
|
|
28
|
+
Download the pre-trained FACodec model from HuggingFace: [Pretrained FACodec checkpoint](https://huggingface.co/amphion/naturalspeech3_facodec)
|
|
29
|
+
|
|
30
|
+
Install Amphion
|
|
31
|
+
```bash
|
|
32
|
+
git clone https://github.com/open-mmlab/Amphion.git
|
|
33
|
+
```
|
|
34
|
+
|
|
35
|
+
Few lines of code to use the pre-trained FACodec model
|
|
36
|
+
```python
|
|
37
|
+
from Amphion.models.codec.ns3_codec import FACodecEncoder, FACodecDecoder
|
|
38
|
+
from huggingface_hub import hf_hub_download
|
|
39
|
+
|
|
40
|
+
fa_encoder = FACodecEncoder(
|
|
41
|
+
ngf=32,
|
|
42
|
+
up_ratios=[2, 4, 5, 5],
|
|
43
|
+
out_channels=256,
|
|
44
|
+
)
|
|
45
|
+
|
|
46
|
+
fa_decoder = FACodecDecoder(
|
|
47
|
+
in_channels=256,
|
|
48
|
+
upsample_initial_channel=1024,
|
|
49
|
+
ngf=32,
|
|
50
|
+
up_ratios=[5, 5, 4, 2],
|
|
51
|
+
vq_num_q_c=2,
|
|
52
|
+
vq_num_q_p=1,
|
|
53
|
+
vq_num_q_r=3,
|
|
54
|
+
vq_dim=256,
|
|
55
|
+
codebook_dim=8,
|
|
56
|
+
codebook_size_prosody=10,
|
|
57
|
+
codebook_size_content=10,
|
|
58
|
+
codebook_size_residual=10,
|
|
59
|
+
use_gr_x_timbre=True,
|
|
60
|
+
use_gr_residual_f0=True,
|
|
61
|
+
use_gr_residual_phone=True,
|
|
62
|
+
)
|
|
63
|
+
|
|
64
|
+
encoder_ckpt = hf_hub_download(repo_id="amphion/naturalspeech3_facodec", filename="ns3_facodec_encoder.bin")
|
|
65
|
+
decoder_ckpt = hf_hub_download(repo_id="amphion/naturalspeech3_facodec", filename="ns3_facodec_decoder.bin")
|
|
66
|
+
|
|
67
|
+
fa_encoder.load_state_dict(torch.load(encoder_ckpt))
|
|
68
|
+
fa_decoder.load_state_dict(torch.load(decoder_ckpt))
|
|
69
|
+
|
|
70
|
+
fa_encoder.eval()
|
|
71
|
+
fa_decoder.eval()
|
|
72
|
+
|
|
73
|
+
```
|
|
74
|
+
|
|
75
|
+
Inference
|
|
76
|
+
```python
|
|
77
|
+
test_wav_path = "test.wav"
|
|
78
|
+
test_wav = librosa.load(test_wav_path, sr=16000)[0]
|
|
79
|
+
test_wav = torch.from_numpy(test_wav).float()
|
|
80
|
+
test_wav = test_wav.unsqueeze(0).unsqueeze(0)
|
|
81
|
+
|
|
82
|
+
with torch.no_grad():
|
|
83
|
+
|
|
84
|
+
# encode
|
|
85
|
+
enc_out = fa_encoder(test_wav)
|
|
86
|
+
print(enc_out.shape)
|
|
87
|
+
|
|
88
|
+
# quantize
|
|
89
|
+
vq_post_emb, vq_id, _, quantized, spk_embs = fa_decoder(enc_out, eval_vq=False, vq=True)
|
|
90
|
+
|
|
91
|
+
# latent after quantization
|
|
92
|
+
print(vq_post_emb.shape)
|
|
93
|
+
|
|
94
|
+
# codes
|
|
95
|
+
print("vq id shape:", vq_id.shape)
|
|
96
|
+
|
|
97
|
+
# get prosody code
|
|
98
|
+
prosody_code = vq_id[:1]
|
|
99
|
+
print("prosody code shape:", prosody_code.shape)
|
|
100
|
+
|
|
101
|
+
# get content code
|
|
102
|
+
cotent_code = vq_id[1:3]
|
|
103
|
+
print("content code shape:", cotent_code.shape)
|
|
104
|
+
|
|
105
|
+
# get residual code (acoustic detail codes)
|
|
106
|
+
residual_code = vq_id[3:]
|
|
107
|
+
print("residual code shape:", residual_code.shape)
|
|
108
|
+
|
|
109
|
+
# speaker embedding
|
|
110
|
+
print("speaker embedding shape:", spk_embs.shape)
|
|
111
|
+
|
|
112
|
+
# decode (recommand)
|
|
113
|
+
recon_wav = fa_decoder.inference(vq_post_emb, spk_embs)
|
|
114
|
+
print(recon_wav.shape)
|
|
115
|
+
sf.write("recon.wav", recon_wav[0][0].cpu().numpy(), 16000)
|
|
116
|
+
```
|
|
117
|
+
|
|
118
|
+
FACodec can achieve zero-shot voice conversion with FACodecEncoderV2/FACodecDecoderV2 or FACodecRedecoder
|
|
119
|
+
```python
|
|
120
|
+
from Amphion.models.codec.ns3_codec import FACodecEncoderV2, FACodecDecoderV2
|
|
121
|
+
|
|
122
|
+
# Same parameters as FACodecEncoder/FACodecDecoder
|
|
123
|
+
fa_encoder_v2 = FACodecEncoderV2(...)
|
|
124
|
+
fa_decoder_v2 = FACodecDecoderV2(...)
|
|
125
|
+
|
|
126
|
+
encoder_v2_ckpt = hf_hub_download(repo_id="amphion/naturalspeech3_facodec", filename="ns3_facodec_encoder_v2.bin")
|
|
127
|
+
decoder_v2_ckpt = hf_hub_download(repo_id="amphion/naturalspeech3_facodec", filename="ns3_facodec_decoder_v2.bin")
|
|
128
|
+
|
|
129
|
+
fa_encoder_v2.load_state_dict(torch.load(encoder_v2_ckpt))
|
|
130
|
+
fa_decoder_v2.load_state_dict(torch.load(decoder_v2_ckpt))
|
|
131
|
+
|
|
132
|
+
with torch.no_grad():
|
|
133
|
+
enc_out_a = fa_encoder_v2(wav_a)
|
|
134
|
+
prosody_a = fa_encoder_v2.get_prosody_feature(wav_a)
|
|
135
|
+
enc_out_b = fa_encoder_v2(wav_b)
|
|
136
|
+
prosody_b = fa_encoder_v2.get_prosody_feature(wav_b)
|
|
137
|
+
|
|
138
|
+
vq_post_emb_a, vq_id_a, _, quantized, spk_embs_a = fa_decoder_v2(
|
|
139
|
+
enc_out_a, prosody_a, eval_vq=False, vq=True
|
|
140
|
+
)
|
|
141
|
+
vq_post_emb_b, vq_id_b, _, quantized, spk_embs_b = fa_decoder_v2(
|
|
142
|
+
enc_out_b, prosody_b, eval_vq=False, vq=True
|
|
143
|
+
)
|
|
144
|
+
|
|
145
|
+
vq_post_emb_a_to_b = fa_decoder_v2.vq2emb(vq_id_a, use_residual=False)
|
|
146
|
+
recon_wav_a_to_b = fa_decoder_v2.inference(vq_post_emb_a_to_b, spk_embs_b)
|
|
147
|
+
```
|
|
148
|
+
|
|
149
|
+
or
|
|
150
|
+
|
|
151
|
+
```python
|
|
152
|
+
from Amphion.models.codec.ns3_codec import FACodecRedecoder
|
|
153
|
+
|
|
154
|
+
fa_redecoder = FACodecRedecoder()
|
|
155
|
+
|
|
156
|
+
redecoder_ckpt = hf_hub_download(repo_id="amphion/naturalspeech3_facodec", filename="ns3_facodec_redecoder.bin")
|
|
157
|
+
|
|
158
|
+
fa_redecoder.load_state_dict(torch.load(redecoder_ckpt))
|
|
159
|
+
|
|
160
|
+
with torch.no_grad():
|
|
161
|
+
enc_out_a = fa_encoder(wav_a)
|
|
162
|
+
enc_out_b = fa_encoder(wav_b)
|
|
163
|
+
|
|
164
|
+
vq_post_emb_a, vq_id_a, _, quantized_a, spk_embs_a = fa_decoder(enc_out_a, eval_vq=False, vq=True)
|
|
165
|
+
vq_post_emb_b, vq_id_b, _, quantized_b, spk_embs_b = fa_decoder(enc_out_b, eval_vq=False, vq=True)
|
|
166
|
+
|
|
167
|
+
# convert speaker
|
|
168
|
+
vq_post_emb_a_to_b = fa_redecoder.vq2emb(vq_id_a, spk_embs_b, use_residual=False)
|
|
169
|
+
recon_wav_a_to_b = fa_redecoder.inference(vq_post_emb_a_to_b, spk_embs_b)
|
|
170
|
+
|
|
171
|
+
sf.write("recon_a_to_b.wav", recon_wav_a_to_b[0][0].cpu().numpy(), 16000)
|
|
172
|
+
```
|
|
173
|
+
|
|
174
|
+
## Q&A
|
|
175
|
+
|
|
176
|
+
Q1: What audio sample rate does FACodec support? What is the hop size? How many codes will be generated for each frame?
|
|
177
|
+
|
|
178
|
+
A1: FACodec supports 16KHz speech audio. The hop size is 200 samples, and (16000/200) * 6 (total number of codebooks) codes will be generated for each frame.
|
|
179
|
+
|
|
180
|
+
Q2: Is it possible to train an autoregressive TTS model like VALL-E using FACodec?
|
|
181
|
+
|
|
182
|
+
A2: Yes. In fact, the authors of NaturalSpeech 3 have already employ explore the autoregressive generative model for discrete token generation with FACodec. They use an autoregressive language model to generate prosody codes, followed by a non-autoregressive model to generate the remaining content and acoustic details codes.
|
|
183
|
+
|
|
184
|
+
Q3: Is it possible to train a latent diffusion TTS model like NaturalSpeech2 using FACodec?
|
|
185
|
+
|
|
186
|
+
A3: Yes. You can use the latent getted after quanzaition as the modelling target for the latent diffusion model.
|
|
187
|
+
|
|
188
|
+
Q4: Can FACodec compress and reconstruct audio from other domains? Such as sound effects, music, etc.
|
|
189
|
+
|
|
190
|
+
A4: Since FACodec is designed for speech, it may not be suitable for other audio domains. However, it is possible to use the FACodec model to compress and reconstruct audio from other domains, but the quality may not be as good as the original audio.
|
|
191
|
+
|
|
192
|
+
Q5: Can FACodec be used for content feature for some other tasks like voice conversion?
|
|
193
|
+
|
|
194
|
+
A5: I think the answer is yes. Researchers can use the content code of FACodec as the content feature for voice conversion. We hope to see more research in this direction.
|
|
195
|
+
|
|
196
|
+
## Citations
|
|
197
|
+
|
|
198
|
+
If you use our FACodec model, please cite the following paper:
|
|
199
|
+
|
|
200
|
+
```bibtex
|
|
201
|
+
@article{ju2024naturalspeech,
|
|
202
|
+
title={NaturalSpeech 3: Zero-Shot Speech Synthesis with Factorized Codec and Diffusion Models},
|
|
203
|
+
author={Ju, Zeqian and Wang, Yuancheng and Shen, Kai and Tan, Xu and Xin, Detai and Yang, Dongchao and Liu, Yanqing and Leng, Yichong and Song, Kaitao and Tang, Siliang and others},
|
|
204
|
+
journal={arXiv preprint arXiv:2403.03100},
|
|
205
|
+
year={2024}
|
|
206
|
+
}
|
|
207
|
+
|
|
208
|
+
@article{zhang2023amphion,
|
|
209
|
+
title={Amphion: An Open-Source Audio, Music and Speech Generation Toolkit},
|
|
210
|
+
author={Xueyao Zhang and Liumeng Xue and Yicheng Gu and Yuancheng Wang and Haorui He and Chaoren Wang and Xi Chen and Zihao Fang and Haopeng Chen and Junan Zhang and Tze Ying Tang and Lexiao Zou and Mingxuan Wang and Jun Han and Kai Chen and Haizhou Li and Zhizheng Wu},
|
|
211
|
+
journal={arXiv},
|
|
212
|
+
year={2024},
|
|
213
|
+
volume={abs/2312.09911}
|
|
214
|
+
}
|
|
215
|
+
```
|
|
216
|
+
|
|
@@ -0,0 +1,29 @@
|
|
|
1
|
+
# Adapted from https://github.com/junjun3518/alias-free-torch under the Apache License 2.0
|
|
2
|
+
|
|
3
|
+
import torch.nn as nn
|
|
4
|
+
from .resample import UpSample1d, DownSample1d
|
|
5
|
+
|
|
6
|
+
|
|
7
|
+
class Activation1d(nn.Module):
|
|
8
|
+
def __init__(
|
|
9
|
+
self,
|
|
10
|
+
activation,
|
|
11
|
+
up_ratio: int = 2,
|
|
12
|
+
down_ratio: int = 2,
|
|
13
|
+
up_kernel_size: int = 12,
|
|
14
|
+
down_kernel_size: int = 12,
|
|
15
|
+
):
|
|
16
|
+
super().__init__()
|
|
17
|
+
self.up_ratio = up_ratio
|
|
18
|
+
self.down_ratio = down_ratio
|
|
19
|
+
self.act = activation
|
|
20
|
+
self.upsample = UpSample1d(up_ratio, up_kernel_size)
|
|
21
|
+
self.downsample = DownSample1d(down_ratio, down_kernel_size)
|
|
22
|
+
|
|
23
|
+
# x: [B,C,T]
|
|
24
|
+
def forward(self, x):
|
|
25
|
+
x = self.upsample(x)
|
|
26
|
+
x = self.act(x)
|
|
27
|
+
x = self.downsample(x)
|
|
28
|
+
|
|
29
|
+
return x
|
xinference/thirdparty/indextts/utils/maskgct/models/codec/ns3_codec/alias_free_torch/filter.py
ADDED
|
@@ -0,0 +1,96 @@
|
|
|
1
|
+
# Adapted from https://github.com/junjun3518/alias-free-torch under the Apache License 2.0
|
|
2
|
+
|
|
3
|
+
import torch
|
|
4
|
+
import torch.nn as nn
|
|
5
|
+
import torch.nn.functional as F
|
|
6
|
+
import math
|
|
7
|
+
|
|
8
|
+
if "sinc" in dir(torch):
|
|
9
|
+
sinc = torch.sinc
|
|
10
|
+
else:
|
|
11
|
+
# This code is adopted from adefossez's julius.core.sinc under the MIT License
|
|
12
|
+
# https://adefossez.github.io/julius/julius/core.html
|
|
13
|
+
def sinc(x: torch.Tensor):
|
|
14
|
+
"""
|
|
15
|
+
Implementation of sinc, i.e. sin(pi * x) / (pi * x)
|
|
16
|
+
__Warning__: Different to julius.sinc, the input is multiplied by `pi`!
|
|
17
|
+
"""
|
|
18
|
+
return torch.where(
|
|
19
|
+
x == 0,
|
|
20
|
+
torch.tensor(1.0, device=x.device, dtype=x.dtype),
|
|
21
|
+
torch.sin(math.pi * x) / math.pi / x,
|
|
22
|
+
)
|
|
23
|
+
|
|
24
|
+
|
|
25
|
+
# This code is adopted from adefossez's julius.lowpass.LowPassFilters under the MIT License
|
|
26
|
+
# https://adefossez.github.io/julius/julius/lowpass.html
|
|
27
|
+
def kaiser_sinc_filter1d(
|
|
28
|
+
cutoff, half_width, kernel_size
|
|
29
|
+
): # return filter [1,1,kernel_size]
|
|
30
|
+
even = kernel_size % 2 == 0
|
|
31
|
+
half_size = kernel_size // 2
|
|
32
|
+
|
|
33
|
+
# For kaiser window
|
|
34
|
+
delta_f = 4 * half_width
|
|
35
|
+
A = 2.285 * (half_size - 1) * math.pi * delta_f + 7.95
|
|
36
|
+
if A > 50.0:
|
|
37
|
+
beta = 0.1102 * (A - 8.7)
|
|
38
|
+
elif A >= 21.0:
|
|
39
|
+
beta = 0.5842 * (A - 21) ** 0.4 + 0.07886 * (A - 21.0)
|
|
40
|
+
else:
|
|
41
|
+
beta = 0.0
|
|
42
|
+
window = torch.kaiser_window(kernel_size, beta=beta, periodic=False)
|
|
43
|
+
|
|
44
|
+
# ratio = 0.5/cutoff -> 2 * cutoff = 1 / ratio
|
|
45
|
+
if even:
|
|
46
|
+
time = torch.arange(-half_size, half_size) + 0.5
|
|
47
|
+
else:
|
|
48
|
+
time = torch.arange(kernel_size) - half_size
|
|
49
|
+
if cutoff == 0:
|
|
50
|
+
filter_ = torch.zeros_like(time)
|
|
51
|
+
else:
|
|
52
|
+
filter_ = 2 * cutoff * window * sinc(2 * cutoff * time)
|
|
53
|
+
# Normalize filter to have sum = 1, otherwise we will have a small leakage
|
|
54
|
+
# of the constant component in the input signal.
|
|
55
|
+
filter_ /= filter_.sum()
|
|
56
|
+
filter = filter_.view(1, 1, kernel_size)
|
|
57
|
+
|
|
58
|
+
return filter
|
|
59
|
+
|
|
60
|
+
|
|
61
|
+
class LowPassFilter1d(nn.Module):
|
|
62
|
+
def __init__(
|
|
63
|
+
self,
|
|
64
|
+
cutoff=0.5,
|
|
65
|
+
half_width=0.6,
|
|
66
|
+
stride: int = 1,
|
|
67
|
+
padding: bool = True,
|
|
68
|
+
padding_mode: str = "replicate",
|
|
69
|
+
kernel_size: int = 12,
|
|
70
|
+
):
|
|
71
|
+
# kernel_size should be even number for stylegan3 setup,
|
|
72
|
+
# in this implementation, odd number is also possible.
|
|
73
|
+
super().__init__()
|
|
74
|
+
if cutoff < -0.0:
|
|
75
|
+
raise ValueError("Minimum cutoff must be larger than zero.")
|
|
76
|
+
if cutoff > 0.5:
|
|
77
|
+
raise ValueError("A cutoff above 0.5 does not make sense.")
|
|
78
|
+
self.kernel_size = kernel_size
|
|
79
|
+
self.even = kernel_size % 2 == 0
|
|
80
|
+
self.pad_left = kernel_size // 2 - int(self.even)
|
|
81
|
+
self.pad_right = kernel_size // 2
|
|
82
|
+
self.stride = stride
|
|
83
|
+
self.padding = padding
|
|
84
|
+
self.padding_mode = padding_mode
|
|
85
|
+
filter = kaiser_sinc_filter1d(cutoff, half_width, kernel_size)
|
|
86
|
+
self.register_buffer("filter", filter)
|
|
87
|
+
|
|
88
|
+
# input [B, C, T]
|
|
89
|
+
def forward(self, x):
|
|
90
|
+
_, C, _ = x.shape
|
|
91
|
+
|
|
92
|
+
if self.padding:
|
|
93
|
+
x = F.pad(x, (self.pad_left, self.pad_right), mode=self.padding_mode)
|
|
94
|
+
out = F.conv1d(x, self.filter.expand(C, -1, -1), stride=self.stride, groups=C)
|
|
95
|
+
|
|
96
|
+
return out
|
xinference/thirdparty/indextts/utils/maskgct/models/codec/ns3_codec/alias_free_torch/resample.py
ADDED
|
@@ -0,0 +1,57 @@
|
|
|
1
|
+
# Adapted from https://github.com/junjun3518/alias-free-torch under the Apache License 2.0
|
|
2
|
+
|
|
3
|
+
import torch.nn as nn
|
|
4
|
+
from torch.nn import functional as F
|
|
5
|
+
from .filter import LowPassFilter1d
|
|
6
|
+
from .filter import kaiser_sinc_filter1d
|
|
7
|
+
|
|
8
|
+
|
|
9
|
+
class UpSample1d(nn.Module):
|
|
10
|
+
def __init__(self, ratio=2, kernel_size=None):
|
|
11
|
+
super().__init__()
|
|
12
|
+
self.ratio = ratio
|
|
13
|
+
self.kernel_size = (
|
|
14
|
+
int(6 * ratio // 2) * 2 if kernel_size is None else kernel_size
|
|
15
|
+
)
|
|
16
|
+
self.stride = ratio
|
|
17
|
+
self.pad = self.kernel_size // ratio - 1
|
|
18
|
+
self.pad_left = self.pad * self.stride + (self.kernel_size - self.stride) // 2
|
|
19
|
+
self.pad_right = (
|
|
20
|
+
self.pad * self.stride + (self.kernel_size - self.stride + 1) // 2
|
|
21
|
+
)
|
|
22
|
+
filter = kaiser_sinc_filter1d(
|
|
23
|
+
cutoff=0.5 / ratio, half_width=0.6 / ratio, kernel_size=self.kernel_size
|
|
24
|
+
)
|
|
25
|
+
self.register_buffer("filter", filter)
|
|
26
|
+
|
|
27
|
+
# x: [B, C, T]
|
|
28
|
+
def forward(self, x):
|
|
29
|
+
_, C, _ = x.shape
|
|
30
|
+
|
|
31
|
+
x = F.pad(x, (self.pad, self.pad), mode="replicate")
|
|
32
|
+
x = self.ratio * F.conv_transpose1d(
|
|
33
|
+
x, self.filter.expand(C, -1, -1), stride=self.stride, groups=C
|
|
34
|
+
)
|
|
35
|
+
x = x[..., self.pad_left : -self.pad_right]
|
|
36
|
+
|
|
37
|
+
return x
|
|
38
|
+
|
|
39
|
+
|
|
40
|
+
class DownSample1d(nn.Module):
|
|
41
|
+
def __init__(self, ratio=2, kernel_size=None):
|
|
42
|
+
super().__init__()
|
|
43
|
+
self.ratio = ratio
|
|
44
|
+
self.kernel_size = (
|
|
45
|
+
int(6 * ratio // 2) * 2 if kernel_size is None else kernel_size
|
|
46
|
+
)
|
|
47
|
+
self.lowpass = LowPassFilter1d(
|
|
48
|
+
cutoff=0.5 / ratio,
|
|
49
|
+
half_width=0.6 / ratio,
|
|
50
|
+
stride=ratio,
|
|
51
|
+
kernel_size=self.kernel_size,
|
|
52
|
+
)
|
|
53
|
+
|
|
54
|
+
def forward(self, x):
|
|
55
|
+
xx = self.lowpass(x)
|
|
56
|
+
|
|
57
|
+
return xx
|