rxnn 0.1.55__tar.gz → 0.1.56__tar.gz

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (31) hide show
  1. {rxnn-0.1.55 → rxnn-0.1.56}/PKG-INFO +1 -1
  2. {rxnn-0.1.55 → rxnn-0.1.56}/pyproject.toml +1 -1
  3. {rxnn-0.1.55 → rxnn-0.1.56}/src/rxnn/experimental/attention.py +2 -2
  4. {rxnn-0.1.55 → rxnn-0.1.56}/LICENSE +0 -0
  5. {rxnn-0.1.55 → rxnn-0.1.56}/README.md +0 -0
  6. {rxnn-0.1.55 → rxnn-0.1.56}/src/rxnn/__init__.py +0 -0
  7. {rxnn-0.1.55 → rxnn-0.1.56}/src/rxnn/experimental/__init__.py +0 -0
  8. {rxnn-0.1.55 → rxnn-0.1.56}/src/rxnn/experimental/models.py +0 -0
  9. {rxnn-0.1.55 → rxnn-0.1.56}/src/rxnn/experimental/moe.py +0 -0
  10. {rxnn-0.1.55 → rxnn-0.1.56}/src/rxnn/memory/__init__.py +0 -0
  11. {rxnn-0.1.55 → rxnn-0.1.56}/src/rxnn/memory/norm.py +0 -0
  12. {rxnn-0.1.55 → rxnn-0.1.56}/src/rxnn/memory/stm.py +0 -0
  13. {rxnn-0.1.55 → rxnn-0.1.56}/src/rxnn/rxt/__init__.py +0 -0
  14. {rxnn-0.1.55 → rxnn-0.1.56}/src/rxnn/rxt/models.py +0 -0
  15. {rxnn-0.1.55 → rxnn-0.1.56}/src/rxnn/training/__init__.py +0 -0
  16. {rxnn-0.1.55 → rxnn-0.1.56}/src/rxnn/training/base.py +0 -0
  17. {rxnn-0.1.55 → rxnn-0.1.56}/src/rxnn/training/bml.py +0 -0
  18. {rxnn-0.1.55 → rxnn-0.1.56}/src/rxnn/training/callbacks.py +0 -0
  19. {rxnn-0.1.55 → rxnn-0.1.56}/src/rxnn/training/dataset.py +0 -0
  20. {rxnn-0.1.55 → rxnn-0.1.56}/src/rxnn/training/scheduler.py +0 -0
  21. {rxnn-0.1.55 → rxnn-0.1.56}/src/rxnn/training/tokenizer.py +0 -0
  22. {rxnn-0.1.55 → rxnn-0.1.56}/src/rxnn/transformers/__init__.py +0 -0
  23. {rxnn-0.1.55 → rxnn-0.1.56}/src/rxnn/transformers/attention.py +0 -0
  24. {rxnn-0.1.55 → rxnn-0.1.56}/src/rxnn/transformers/ff.py +0 -0
  25. {rxnn-0.1.55 → rxnn-0.1.56}/src/rxnn/transformers/layers.py +0 -0
  26. {rxnn-0.1.55 → rxnn-0.1.56}/src/rxnn/transformers/mask.py +0 -0
  27. {rxnn-0.1.55 → rxnn-0.1.56}/src/rxnn/transformers/models.py +0 -0
  28. {rxnn-0.1.55 → rxnn-0.1.56}/src/rxnn/transformers/moe.py +0 -0
  29. {rxnn-0.1.55 → rxnn-0.1.56}/src/rxnn/transformers/positional.py +0 -0
  30. {rxnn-0.1.55 → rxnn-0.1.56}/src/rxnn/transformers/sampler.py +0 -0
  31. {rxnn-0.1.55 → rxnn-0.1.56}/src/rxnn/utils.py +0 -0
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.3
2
2
  Name: rxnn
3
- Version: 0.1.55
3
+ Version: 0.1.56
4
4
  Summary: RxNN: Reactive Neural Networks Platform
5
5
  License: Apache-2.0
6
6
  Keywords: deep-learning,ai,machine-learning
@@ -4,7 +4,7 @@ build-backend = "poetry.core.masonry.api"
4
4
 
5
5
  [tool.poetry]
6
6
  name = "rxnn"
7
- version = "0.1.55"
7
+ version = "0.1.56"
8
8
  description = "RxNN: Reactive Neural Networks Platform"
9
9
 
10
10
  license = "Apache-2.0"
@@ -283,12 +283,12 @@ class SparseQueryAttention(MultiHeadAttention):
283
283
  """Override query, key, and value projections for GQA case - split data into heads and groups"""
284
284
  head_dim = d // self.num_heads
285
285
  if not self.rel_embed:
286
- q = self.q_proj(query).view(b, t, self.num_query_heads, head_dim).transpose(1, 2)
286
+ q = self.q_proj(query).view(b, t, self.num_query_groups, head_dim).transpose(1, 2)
287
287
  k = self.k_proj(key).view(b, -1, self.num_groups, head_dim).transpose(1, 2)
288
288
  v = self.v_proj(value).view(b, -1, self.num_groups, head_dim).transpose(1, 2)
289
289
  else:
290
290
  group_heads = self.num_heads // self.num_groups
291
- query_heads = self.num_query_heads // self.num_query_groups
291
+ query_heads = self.num_heads // self.num_query_groups
292
292
  # Process Q
293
293
  q = self.q_proj(query).view(b, -1, self.num_query_groups, head_dim).transpose(1, 2) # (B, Q_G, T, head_dim)
294
294
 
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes