x-transformers 1.30.4__py3-none-any.whl → 1.30.6__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
x_transformers/attend.py CHANGED
@@ -22,6 +22,7 @@ class Intermediates:
22
22
  pre_softmax_attn: Tensor | None = None
23
23
  post_softmax_attn: Tensor | None = None
24
24
  cached_kv: Tuple[Tensor, Tensor] | None = None
25
+ layer_type: str | None = None
25
26
 
26
27
  def to_tuple(self):
27
28
  return (self.qk_similarities, self.pre_softmax_attn, self.post_softmax_attn)
@@ -1410,6 +1410,7 @@ class AttentionLayers(Module):
1410
1410
  x = residual_fn(out, inner_residual)
1411
1411
 
1412
1412
  if layer_type in ('a', 'c') and return_hiddens:
1413
+ inter.layer_type = layer_type
1413
1414
  intermediates.append(inter)
1414
1415
 
1415
1416
  if layer_type == 'a' and self.residual_attn:
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.1
2
2
  Name: x-transformers
3
- Version: 1.30.4
3
+ Version: 1.30.6
4
4
  Summary: X-Transformers - Pytorch
5
5
  Home-page: https://github.com/lucidrains/x-transformers
6
6
  Author: Phil Wang
@@ -1,14 +1,14 @@
1
1
  x_transformers/__init__.py,sha256=8LQl-dNL6vj8VHRx5LMSOlRDTXQvYOuM21PDXz8WdiI,703
2
- x_transformers/attend.py,sha256=2SPHjXS_QAAZt04lHWGtdOypTExmo3BrbFhgcIQTk-Y,10671
2
+ x_transformers/attend.py,sha256=8opOeCQddi440WcH73B_wB5vtL0jaEQwBL-DIWq2lCs,10713
3
3
  x_transformers/autoregressive_wrapper.py,sha256=uX8Mb0zLsQrZECt_9UGt35g7tC05Rk3nPqO6xp2FFCc,9619
4
4
  x_transformers/continuous.py,sha256=WO52n9lFAXv5-SGadi2cApGF8dkouN8QSTEOuC7erj8,6180
5
5
  x_transformers/dpo.py,sha256=LjvWgCkqTl-UuehrzQ8nkX5guLr4whYwsmm7SKSwdls,3450
6
6
  x_transformers/nonautoregressive_wrapper.py,sha256=ys_p8obc7lTeeodCqvkRKxOXQ1C9T3j5Jwr-JbVgnXk,10432
7
- x_transformers/x_transformers.py,sha256=P4rqlYGS9j9Gz00B4NPM7L6mhvamSYdBy5nG0ggOIMM,66342
7
+ x_transformers/x_transformers.py,sha256=pXckFcDL6kTghYEUjIamZiR5H8dV6aIEPQTIYAGgqxA,66388
8
8
  x_transformers/xl_autoregressive_wrapper.py,sha256=DCx4n0_c1tFai4nOqaWVnqx2p9eutsZsDMiMP1ckxNU,4117
9
9
  x_transformers/xval.py,sha256=QE1ltYZTR_eGgIHPP2BrMWVWVLqMW-OpDZh87BSmQEg,8563
10
- x_transformers-1.30.4.dist-info/LICENSE,sha256=As9u198X-U-vph5noInuUfqsAG2zX_oXPHDmdjwlPPY,1066
11
- x_transformers-1.30.4.dist-info/METADATA,sha256=VwdrJaRjocQXIAxdGzq4rByPGvaA4jsogostzCysdjI,661
12
- x_transformers-1.30.4.dist-info/WHEEL,sha256=GJ7t_kWBFywbagK5eo9IoUwLW6oyOeTKmQ-9iHFVNxQ,92
13
- x_transformers-1.30.4.dist-info/top_level.txt,sha256=hO6KGpFuGucRNEtRfme4A_rGcM53AKwGP7RVlRIxS5Q,15
14
- x_transformers-1.30.4.dist-info/RECORD,,
10
+ x_transformers-1.30.6.dist-info/LICENSE,sha256=As9u198X-U-vph5noInuUfqsAG2zX_oXPHDmdjwlPPY,1066
11
+ x_transformers-1.30.6.dist-info/METADATA,sha256=1Xq9oSctaCQ5TOpdM3j6lJENYStLuda5VqEzOtq1B0c,661
12
+ x_transformers-1.30.6.dist-info/WHEEL,sha256=GJ7t_kWBFywbagK5eo9IoUwLW6oyOeTKmQ-9iHFVNxQ,92
13
+ x_transformers-1.30.6.dist-info/top_level.txt,sha256=hO6KGpFuGucRNEtRfme4A_rGcM53AKwGP7RVlRIxS5Q,15
14
+ x_transformers-1.30.6.dist-info/RECORD,,