x-transformers 1.31.5__py3-none-any.whl → 1.31.6__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1267,7 +1267,8 @@ class AttentionLayers(Module):
1267
1267
  scale_residual_constant = 1.,
1268
1268
  shift_tokens = 0,
1269
1269
  sandwich_norm = False,
1270
- softclamp_output_value: float | None = None,
1270
+ softclamp_output = False,
1271
+ softclamp_output_value = 50.,
1271
1272
  resi_dual = False,
1272
1273
  resi_dual_scale = 1.,
1273
1274
  zero_init_branch_output = False,
@@ -1484,6 +1485,7 @@ class AttentionLayers(Module):
1484
1485
  # optional soft clamping just before the final norm
1485
1486
  # used in gemma 2
1486
1487
 
1488
+ self.softclamp_output = softclamp_output
1487
1489
  self.softclamp_output_value = softclamp_output_value
1488
1490
 
1489
1491
  # whether it has post norm
@@ -1717,7 +1719,7 @@ class AttentionLayers(Module):
1717
1719
  if return_hiddens:
1718
1720
  layer_hiddens.append(x)
1719
1721
 
1720
- if exists(self.softclamp_output_value):
1722
+ if self.softclamp_output:
1721
1723
  x = softclamp(x, self.softclamp_output_value)
1722
1724
 
1723
1725
  final_norm = self.final_norm
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.1
2
2
  Name: x-transformers
3
- Version: 1.31.5
3
+ Version: 1.31.6
4
4
  Summary: X-Transformers - Pytorch
5
5
  Home-page: https://github.com/lucidrains/x-transformers
6
6
  Author: Phil Wang
@@ -4,11 +4,11 @@ x_transformers/autoregressive_wrapper.py,sha256=uX8Mb0zLsQrZECt_9UGt35g7tC05Rk3n
4
4
  x_transformers/continuous.py,sha256=WO52n9lFAXv5-SGadi2cApGF8dkouN8QSTEOuC7erj8,6180
5
5
  x_transformers/dpo.py,sha256=LjvWgCkqTl-UuehrzQ8nkX5guLr4whYwsmm7SKSwdls,3450
6
6
  x_transformers/nonautoregressive_wrapper.py,sha256=ys_p8obc7lTeeodCqvkRKxOXQ1C9T3j5Jwr-JbVgnXk,10432
7
- x_transformers/x_transformers.py,sha256=VL9Dm8L5jnpgyt_V6DWGtLs9MeiTn7ZMCQdcHSFLVo8,75871
7
+ x_transformers/x_transformers.py,sha256=xPwzR3bd8BS_ChEcz0UxsNtx99u4UbP8jg1fFIRDGUw,75925
8
8
  x_transformers/xl_autoregressive_wrapper.py,sha256=DCx4n0_c1tFai4nOqaWVnqx2p9eutsZsDMiMP1ckxNU,4117
9
9
  x_transformers/xval.py,sha256=QE1ltYZTR_eGgIHPP2BrMWVWVLqMW-OpDZh87BSmQEg,8563
10
- x_transformers-1.31.5.dist-info/LICENSE,sha256=As9u198X-U-vph5noInuUfqsAG2zX_oXPHDmdjwlPPY,1066
11
- x_transformers-1.31.5.dist-info/METADATA,sha256=o7z9SmDOf9BAU9FyMRuAbX4LIH8nIACA0acjGHPMz8s,661
12
- x_transformers-1.31.5.dist-info/WHEEL,sha256=mguMlWGMX-VHnMpKOjjQidIo1ssRlCFu4a4mBpz1s2M,91
13
- x_transformers-1.31.5.dist-info/top_level.txt,sha256=hO6KGpFuGucRNEtRfme4A_rGcM53AKwGP7RVlRIxS5Q,15
14
- x_transformers-1.31.5.dist-info/RECORD,,
10
+ x_transformers-1.31.6.dist-info/LICENSE,sha256=As9u198X-U-vph5noInuUfqsAG2zX_oXPHDmdjwlPPY,1066
11
+ x_transformers-1.31.6.dist-info/METADATA,sha256=O2MZXNuX-jqrAdpcxkIDH4J0T63t5nt9utI8FHrRIA0,661
12
+ x_transformers-1.31.6.dist-info/WHEEL,sha256=mguMlWGMX-VHnMpKOjjQidIo1ssRlCFu4a4mBpz1s2M,91
13
+ x_transformers-1.31.6.dist-info/top_level.txt,sha256=hO6KGpFuGucRNEtRfme4A_rGcM53AKwGP7RVlRIxS5Q,15
14
+ x_transformers-1.31.6.dist-info/RECORD,,