rxnn 0.1.33__tar.gz → 0.1.34__tar.gz

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (31) hide show
  1. {rxnn-0.1.33 → rxnn-0.1.34}/PKG-INFO +1 -1
  2. {rxnn-0.1.33 → rxnn-0.1.34}/pyproject.toml +1 -1
  3. {rxnn-0.1.33 → rxnn-0.1.34}/src/rxnn/transformers/attention.py +2 -2
  4. {rxnn-0.1.33 → rxnn-0.1.34}/LICENSE +0 -0
  5. {rxnn-0.1.33 → rxnn-0.1.34}/README.md +0 -0
  6. {rxnn-0.1.33 → rxnn-0.1.34}/src/rxnn/__init__.py +0 -0
  7. {rxnn-0.1.33 → rxnn-0.1.34}/src/rxnn/experimental/__init__.py +0 -0
  8. {rxnn-0.1.33 → rxnn-0.1.34}/src/rxnn/experimental/attention.py +0 -0
  9. {rxnn-0.1.33 → rxnn-0.1.34}/src/rxnn/experimental/models.py +0 -0
  10. {rxnn-0.1.33 → rxnn-0.1.34}/src/rxnn/experimental/moe.py +0 -0
  11. {rxnn-0.1.33 → rxnn-0.1.34}/src/rxnn/memory/__init__.py +0 -0
  12. {rxnn-0.1.33 → rxnn-0.1.34}/src/rxnn/memory/norm.py +0 -0
  13. {rxnn-0.1.33 → rxnn-0.1.34}/src/rxnn/memory/stm.py +0 -0
  14. {rxnn-0.1.33 → rxnn-0.1.34}/src/rxnn/rxt/__init__.py +0 -0
  15. {rxnn-0.1.33 → rxnn-0.1.34}/src/rxnn/rxt/models.py +0 -0
  16. {rxnn-0.1.33 → rxnn-0.1.34}/src/rxnn/training/__init__.py +0 -0
  17. {rxnn-0.1.33 → rxnn-0.1.34}/src/rxnn/training/base.py +0 -0
  18. {rxnn-0.1.33 → rxnn-0.1.34}/src/rxnn/training/bml.py +0 -0
  19. {rxnn-0.1.33 → rxnn-0.1.34}/src/rxnn/training/callbacks.py +0 -0
  20. {rxnn-0.1.33 → rxnn-0.1.34}/src/rxnn/training/dataset.py +0 -0
  21. {rxnn-0.1.33 → rxnn-0.1.34}/src/rxnn/training/scheduler.py +0 -0
  22. {rxnn-0.1.33 → rxnn-0.1.34}/src/rxnn/training/tokenizer.py +0 -0
  23. {rxnn-0.1.33 → rxnn-0.1.34}/src/rxnn/transformers/__init__.py +0 -0
  24. {rxnn-0.1.33 → rxnn-0.1.34}/src/rxnn/transformers/ff.py +0 -0
  25. {rxnn-0.1.33 → rxnn-0.1.34}/src/rxnn/transformers/layers.py +0 -0
  26. {rxnn-0.1.33 → rxnn-0.1.34}/src/rxnn/transformers/mask.py +0 -0
  27. {rxnn-0.1.33 → rxnn-0.1.34}/src/rxnn/transformers/models.py +0 -0
  28. {rxnn-0.1.33 → rxnn-0.1.34}/src/rxnn/transformers/moe.py +0 -0
  29. {rxnn-0.1.33 → rxnn-0.1.34}/src/rxnn/transformers/positional.py +0 -0
  30. {rxnn-0.1.33 → rxnn-0.1.34}/src/rxnn/transformers/sampler.py +0 -0
  31. {rxnn-0.1.33 → rxnn-0.1.34}/src/rxnn/utils.py +0 -0
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.3
2
2
  Name: rxnn
3
- Version: 0.1.33
3
+ Version: 0.1.34
4
4
  Summary: RxNN: Reactive Neural Networks Platform
5
5
  License: Apache-2.0
6
6
  Keywords: deep-learning,ai,machine-learning
@@ -4,7 +4,7 @@ build-backend = "poetry.core.masonry.api"
4
4
 
5
5
  [tool.poetry]
6
6
  name = "rxnn"
7
- version = "0.1.33"
7
+ version = "0.1.34"
8
8
  description = "RxNN: Reactive Neural Networks Platform"
9
9
 
10
10
  license = "Apache-2.0"
@@ -1,7 +1,7 @@
1
1
  import torch
2
2
  import torch.nn as nn
3
3
  import torch.nn.functional as F
4
- from torch.backends.cuda import sdp_kernel, SDPBackend
4
+ from torch.nn.attention import sdpa_kernel, SDPBackend
5
5
  import math
6
6
  from .positional import RotaryPositionalEmbedding, RelativePositionalEmbedding
7
7
 
@@ -102,7 +102,7 @@ class MultiHeadAttention(nn.Module):
102
102
 
103
103
  def _flash_attention(self, q: torch.Tensor, k: torch.Tensor, v: torch.Tensor, b: int, t: int, d: int,
104
104
  mask: torch.Tensor = None, enable_gqa: bool = False):
105
- with sdp_kernel(backends=[SDPBackend.FLASH_ATTENTION]):
105
+ with sdpa_kernel(backends=[SDPBackend.FLASH_ATTENTION]):
106
106
  attn_output = F.scaled_dot_product_attention(
107
107
  q, k, v,
108
108
  attn_mask=mask if not self.is_causal else None,
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes