qwen-tts 0.0.5__tar.gz → 0.1.0__tar.gz

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (25) hide show
  1. {qwen_tts-0.0.5 → qwen_tts-0.1.0}/PKG-INFO +7 -1
  2. {qwen_tts-0.0.5 → qwen_tts-0.1.0}/README.md +6 -0
  3. {qwen_tts-0.0.5 → qwen_tts-0.1.0}/pyproject.toml +1 -1
  4. {qwen_tts-0.0.5 → qwen_tts-0.1.0}/qwen_tts/core/models/modeling_qwen3_tts.py +8 -3
  5. {qwen_tts-0.0.5 → qwen_tts-0.1.0}/qwen_tts/core/tokenizer_12hz/modeling_qwen3_tts_tokenizer_v2.py +4 -3
  6. {qwen_tts-0.0.5 → qwen_tts-0.1.0}/LICENSE +0 -0
  7. {qwen_tts-0.0.5 → qwen_tts-0.1.0}/MANIFEST.in +0 -0
  8. {qwen_tts-0.0.5 → qwen_tts-0.1.0}/qwen_tts/__init__.py +0 -0
  9. {qwen_tts-0.0.5 → qwen_tts-0.1.0}/qwen_tts/__main__.py +0 -0
  10. {qwen_tts-0.0.5 → qwen_tts-0.1.0}/qwen_tts/cli/demo.py +0 -0
  11. {qwen_tts-0.0.5 → qwen_tts-0.1.0}/qwen_tts/core/__init__.py +0 -0
  12. {qwen_tts-0.0.5 → qwen_tts-0.1.0}/qwen_tts/core/models/__init__.py +0 -0
  13. {qwen_tts-0.0.5 → qwen_tts-0.1.0}/qwen_tts/core/models/configuration_qwen3_tts.py +0 -0
  14. {qwen_tts-0.0.5 → qwen_tts-0.1.0}/qwen_tts/core/models/processing_qwen3_tts.py +0 -0
  15. {qwen_tts-0.0.5 → qwen_tts-0.1.0}/qwen_tts/core/tokenizer_12hz/configuration_qwen3_tts_tokenizer_v2.py +0 -0
  16. {qwen_tts-0.0.5 → qwen_tts-0.1.0}/qwen_tts/core/tokenizer_25hz/configuration_qwen3_tts_tokenizer_v1.py +0 -0
  17. {qwen_tts-0.0.5 → qwen_tts-0.1.0}/qwen_tts/core/tokenizer_25hz/modeling_qwen3_tts_tokenizer_v1.py +0 -0
  18. {qwen_tts-0.0.5 → qwen_tts-0.1.0}/qwen_tts/core/tokenizer_25hz/vq/assets/mel_filters.npz +0 -0
  19. {qwen_tts-0.0.5 → qwen_tts-0.1.0}/qwen_tts/core/tokenizer_25hz/vq/core_vq.py +0 -0
  20. {qwen_tts-0.0.5 → qwen_tts-0.1.0}/qwen_tts/core/tokenizer_25hz/vq/speech_vq.py +0 -0
  21. {qwen_tts-0.0.5 → qwen_tts-0.1.0}/qwen_tts/core/tokenizer_25hz/vq/whisper_encoder.py +0 -0
  22. {qwen_tts-0.0.5 → qwen_tts-0.1.0}/qwen_tts/inference/qwen3_tts_model.py +0 -0
  23. {qwen_tts-0.0.5 → qwen_tts-0.1.0}/qwen_tts/inference/qwen3_tts_tokenizer.py +0 -0
  24. {qwen_tts-0.0.5 → qwen_tts-0.1.0}/qwen_tts.egg-info/SOURCES.txt +0 -0
  25. {qwen_tts-0.0.5 → qwen_tts-0.1.0}/setup.cfg +0 -0
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.4
2
2
  Name: qwen-tts
3
- Version: 0.0.5
3
+ Version: 0.1.0
4
4
  Summary: Qwen-TTS python package
5
5
  Author: Alibaba Qwen Team
6
6
  License: Apache-2.0
@@ -1380,4 +1380,10 @@ If you find our paper and code useful in your research, please consider giving a
1380
1380
  }
1381
1381
  ```
1382
1382
 
1383
+
1384
+ ## Star History
1385
+
1386
+ [![Star History Chart](https://api.star-history.com/svg?repos=QwenLM/Qwen3-TTS&type=Date)](https://star-history.com/#QwenLM/Qwen3-TTS&Date)
1387
+
1388
+
1383
1389
  <br>
@@ -1352,4 +1352,10 @@ If you find our paper and code useful in your research, please consider giving a
1352
1352
  }
1353
1353
  ```
1354
1354
 
1355
+
1356
+ ## Star History
1357
+
1358
+ [![Star History Chart](https://api.star-history.com/svg?repos=QwenLM/Qwen3-TTS&type=Date)](https://star-history.com/#QwenLM/Qwen3-TTS&Date)
1359
+
1360
+
1355
1361
  <br>
@@ -4,7 +4,7 @@ build-backend = "setuptools.build_meta"
4
4
 
5
5
  [project]
6
6
  name = "qwen-tts"
7
- version = "0.0.5"
7
+ version = "0.1.0"
8
8
  description = "Qwen-TTS python package"
9
9
  readme = "README.md"
10
10
  requires-python = ">=3.9"
@@ -470,7 +470,7 @@ class Qwen3TTSPreTrainedModel(PreTrainedModel):
470
470
  supports_gradient_checkpointing = True
471
471
  _no_split_modules = ["Qwen3TTSDecoderLayer"]
472
472
  _skip_keys_device_placement = "past_key_values"
473
- _supports_flash_attn_2 = True
473
+ _supports_flash_attn = True
474
474
  _supports_sdpa = True
475
475
  _supports_cache_class = True
476
476
  _supports_static_cache = False
@@ -501,8 +501,7 @@ class Qwen3TTSTalkerTextPreTrainedModel(PreTrainedModel):
501
501
  supports_gradient_checkpointing = True
502
502
  _no_split_modules = []
503
503
  _skip_keys_device_placement = ["past_key_values"]
504
- _supports_flash_attn_3 = True
505
- _supports_flash_attn_2 = True
504
+ _supports_flash_attn = True
506
505
  _supports_sdpa = True
507
506
  _supports_flex_attn = True
508
507
  _supports_cache_class = True
@@ -1869,6 +1868,11 @@ class Qwen3TTSForConditionalGeneration(Qwen3TTSPreTrainedModel, GenerationMixin)
1869
1868
  weights_only=True,
1870
1869
  **kwargs,
1871
1870
  ):
1871
+ # Hotfix to enable passing the correct attn implementation which is stored in the config but not in kwargs
1872
+ requested_attn_implementation = kwargs.pop("attn_implementation", None)
1873
+ if requested_attn_implementation is None and config and config._attn_implementation:
1874
+ requested_attn_implementation = config._attn_implementation
1875
+
1872
1876
  model = super().from_pretrained(
1873
1877
  pretrained_model_name_or_path,
1874
1878
  *model_args,
@@ -1881,6 +1885,7 @@ class Qwen3TTSForConditionalGeneration(Qwen3TTSPreTrainedModel, GenerationMixin)
1881
1885
  revision=revision,
1882
1886
  use_safetensors=use_safetensors,
1883
1887
  weights_only=weights_only,
1888
+ attn_implementation=requested_attn_implementation,
1884
1889
  **kwargs,
1885
1890
  )
1886
1891
  if not local_files_only and not os.path.isdir(pretrained_model_name_or_path):
@@ -198,12 +198,13 @@ class Qwen3TTSTokenizerV2CausalTransConvNet(nn.Module):
198
198
  self.conv = nn.ConvTranspose1d(in_channels, out_channels, kernel_size, stride=stride)
199
199
 
200
200
  pad = kernel_size - stride
201
- self.left_pad = math.ceil(pad)
202
- self.right_pad = pad = self.left_pad
201
+ self.left_pad = 0
202
+ self.right_pad = int(pad)
203
203
 
204
204
  def forward(self, hidden_state):
205
205
  hidden_state = self.conv(hidden_state)
206
- hidden_state = hidden_state[..., self.left_pad : hidden_state.shape[-1] - self.right_pad]
206
+ if self.right_pad > 0:
207
+ hidden_state = hidden_state[..., : hidden_state.shape[-1] - self.right_pad]
207
208
  return hidden_state.contiguous()
208
209
 
209
210
 
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes