x-transformers 1.31.5__py3-none-any.whl → 1.31.6__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- x_transformers/x_transformers.py +4 -2
- {x_transformers-1.31.5.dist-info → x_transformers-1.31.6.dist-info}/METADATA +1 -1
- {x_transformers-1.31.5.dist-info → x_transformers-1.31.6.dist-info}/RECORD +6 -6
- {x_transformers-1.31.5.dist-info → x_transformers-1.31.6.dist-info}/LICENSE +0 -0
- {x_transformers-1.31.5.dist-info → x_transformers-1.31.6.dist-info}/WHEEL +0 -0
- {x_transformers-1.31.5.dist-info → x_transformers-1.31.6.dist-info}/top_level.txt +0 -0
x_transformers/x_transformers.py
CHANGED
@@ -1267,7 +1267,8 @@ class AttentionLayers(Module):
|
|
1267
1267
|
scale_residual_constant = 1.,
|
1268
1268
|
shift_tokens = 0,
|
1269
1269
|
sandwich_norm = False,
|
1270
|
-
|
1270
|
+
softclamp_output = False,
|
1271
|
+
softclamp_output_value = 50.,
|
1271
1272
|
resi_dual = False,
|
1272
1273
|
resi_dual_scale = 1.,
|
1273
1274
|
zero_init_branch_output = False,
|
@@ -1484,6 +1485,7 @@ class AttentionLayers(Module):
|
|
1484
1485
|
# optional soft clamping just before the final norm
|
1485
1486
|
# used in gemma 2
|
1486
1487
|
|
1488
|
+
self.softclamp_output = softclamp_output
|
1487
1489
|
self.softclamp_output_value = softclamp_output_value
|
1488
1490
|
|
1489
1491
|
# whether it has post norm
|
@@ -1717,7 +1719,7 @@ class AttentionLayers(Module):
|
|
1717
1719
|
if return_hiddens:
|
1718
1720
|
layer_hiddens.append(x)
|
1719
1721
|
|
1720
|
-
if
|
1722
|
+
if self.softclamp_output:
|
1721
1723
|
x = softclamp(x, self.softclamp_output_value)
|
1722
1724
|
|
1723
1725
|
final_norm = self.final_norm
|
@@ -4,11 +4,11 @@ x_transformers/autoregressive_wrapper.py,sha256=uX8Mb0zLsQrZECt_9UGt35g7tC05Rk3n
|
|
4
4
|
x_transformers/continuous.py,sha256=WO52n9lFAXv5-SGadi2cApGF8dkouN8QSTEOuC7erj8,6180
|
5
5
|
x_transformers/dpo.py,sha256=LjvWgCkqTl-UuehrzQ8nkX5guLr4whYwsmm7SKSwdls,3450
|
6
6
|
x_transformers/nonautoregressive_wrapper.py,sha256=ys_p8obc7lTeeodCqvkRKxOXQ1C9T3j5Jwr-JbVgnXk,10432
|
7
|
-
x_transformers/x_transformers.py,sha256=
|
7
|
+
x_transformers/x_transformers.py,sha256=xPwzR3bd8BS_ChEcz0UxsNtx99u4UbP8jg1fFIRDGUw,75925
|
8
8
|
x_transformers/xl_autoregressive_wrapper.py,sha256=DCx4n0_c1tFai4nOqaWVnqx2p9eutsZsDMiMP1ckxNU,4117
|
9
9
|
x_transformers/xval.py,sha256=QE1ltYZTR_eGgIHPP2BrMWVWVLqMW-OpDZh87BSmQEg,8563
|
10
|
-
x_transformers-1.31.
|
11
|
-
x_transformers-1.31.
|
12
|
-
x_transformers-1.31.
|
13
|
-
x_transformers-1.31.
|
14
|
-
x_transformers-1.31.
|
10
|
+
x_transformers-1.31.6.dist-info/LICENSE,sha256=As9u198X-U-vph5noInuUfqsAG2zX_oXPHDmdjwlPPY,1066
|
11
|
+
x_transformers-1.31.6.dist-info/METADATA,sha256=O2MZXNuX-jqrAdpcxkIDH4J0T63t5nt9utI8FHrRIA0,661
|
12
|
+
x_transformers-1.31.6.dist-info/WHEEL,sha256=mguMlWGMX-VHnMpKOjjQidIo1ssRlCFu4a4mBpz1s2M,91
|
13
|
+
x_transformers-1.31.6.dist-info/top_level.txt,sha256=hO6KGpFuGucRNEtRfme4A_rGcM53AKwGP7RVlRIxS5Q,15
|
14
|
+
x_transformers-1.31.6.dist-info/RECORD,,
|
File without changes
|
File without changes
|
File without changes
|