rxnn 0.1.36__tar.gz → 0.1.37__tar.gz

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (31) hide show
  1. {rxnn-0.1.36 → rxnn-0.1.37}/PKG-INFO +1 -1
  2. {rxnn-0.1.36 → rxnn-0.1.37}/pyproject.toml +1 -1
  3. {rxnn-0.1.36 → rxnn-0.1.37}/src/rxnn/transformers/attention.py +6 -2
  4. {rxnn-0.1.36 → rxnn-0.1.37}/LICENSE +0 -0
  5. {rxnn-0.1.36 → rxnn-0.1.37}/README.md +0 -0
  6. {rxnn-0.1.36 → rxnn-0.1.37}/src/rxnn/__init__.py +0 -0
  7. {rxnn-0.1.36 → rxnn-0.1.37}/src/rxnn/experimental/__init__.py +0 -0
  8. {rxnn-0.1.36 → rxnn-0.1.37}/src/rxnn/experimental/attention.py +0 -0
  9. {rxnn-0.1.36 → rxnn-0.1.37}/src/rxnn/experimental/models.py +0 -0
  10. {rxnn-0.1.36 → rxnn-0.1.37}/src/rxnn/experimental/moe.py +0 -0
  11. {rxnn-0.1.36 → rxnn-0.1.37}/src/rxnn/memory/__init__.py +0 -0
  12. {rxnn-0.1.36 → rxnn-0.1.37}/src/rxnn/memory/norm.py +0 -0
  13. {rxnn-0.1.36 → rxnn-0.1.37}/src/rxnn/memory/stm.py +0 -0
  14. {rxnn-0.1.36 → rxnn-0.1.37}/src/rxnn/rxt/__init__.py +0 -0
  15. {rxnn-0.1.36 → rxnn-0.1.37}/src/rxnn/rxt/models.py +0 -0
  16. {rxnn-0.1.36 → rxnn-0.1.37}/src/rxnn/training/__init__.py +0 -0
  17. {rxnn-0.1.36 → rxnn-0.1.37}/src/rxnn/training/base.py +0 -0
  18. {rxnn-0.1.36 → rxnn-0.1.37}/src/rxnn/training/bml.py +0 -0
  19. {rxnn-0.1.36 → rxnn-0.1.37}/src/rxnn/training/callbacks.py +0 -0
  20. {rxnn-0.1.36 → rxnn-0.1.37}/src/rxnn/training/dataset.py +0 -0
  21. {rxnn-0.1.36 → rxnn-0.1.37}/src/rxnn/training/scheduler.py +0 -0
  22. {rxnn-0.1.36 → rxnn-0.1.37}/src/rxnn/training/tokenizer.py +0 -0
  23. {rxnn-0.1.36 → rxnn-0.1.37}/src/rxnn/transformers/__init__.py +0 -0
  24. {rxnn-0.1.36 → rxnn-0.1.37}/src/rxnn/transformers/ff.py +0 -0
  25. {rxnn-0.1.36 → rxnn-0.1.37}/src/rxnn/transformers/layers.py +0 -0
  26. {rxnn-0.1.36 → rxnn-0.1.37}/src/rxnn/transformers/mask.py +0 -0
  27. {rxnn-0.1.36 → rxnn-0.1.37}/src/rxnn/transformers/models.py +0 -0
  28. {rxnn-0.1.36 → rxnn-0.1.37}/src/rxnn/transformers/moe.py +0 -0
  29. {rxnn-0.1.36 → rxnn-0.1.37}/src/rxnn/transformers/positional.py +0 -0
  30. {rxnn-0.1.36 → rxnn-0.1.37}/src/rxnn/transformers/sampler.py +0 -0
  31. {rxnn-0.1.36 → rxnn-0.1.37}/src/rxnn/utils.py +0 -0
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.3
2
2
  Name: rxnn
3
- Version: 0.1.36
3
+ Version: 0.1.37
4
4
  Summary: RxNN: Reactive Neural Networks Platform
5
5
  License: Apache-2.0
6
6
  Keywords: deep-learning,ai,machine-learning
@@ -4,7 +4,7 @@ build-backend = "poetry.core.masonry.api"
4
4
 
5
5
  [tool.poetry]
6
6
  name = "rxnn"
7
- version = "0.1.36"
7
+ version = "0.1.37"
8
8
  description = "RxNN: Reactive Neural Networks Platform"
9
9
 
10
10
  license = "Apache-2.0"
@@ -102,8 +102,12 @@ class MultiHeadAttention(nn.Module):
102
102
 
103
103
  def _flash_attention(self, q: torch.Tensor, k: torch.Tensor, v: torch.Tensor, b: int, t: int, d: int,
104
104
  mask: torch.Tensor = None, enable_gqa: bool = False):
105
- with sdpa_kernel(backends=[SDPBackend.FLASH_ATTENTION]):
106
- return self._torch_attention(q, k, v, b, t, d, mask=mask, enable_gqa=enable_gqa)
105
+ # After ~6h of fighthing, PyTorch based is still now working so I decided to use FlashAttention directly
106
+ # with sdpa_kernel(backends=[SDPBackend.FLASH_ATTENTION]):
107
+ # return self._torch_attention(q, k, v, b, t, d, mask=mask, enable_gqa=enable_gqa)
108
+ from flash_attn import flash_attn_func
109
+ attn_output = flash_attn_func(q, k, v, dropout_p=self.dropout.p if self.training else 0.0, is_causal=self.is_causal)
110
+ return self._transpose_output(attn_output, b, t, d)
107
111
 
108
112
  def _torch_attention(self, q: torch.Tensor, k: torch.Tensor, v: torch.Tensor, b: int, t: int, d: int,
109
113
  mask: torch.Tensor = None, enable_gqa: bool = False):
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes