x-transformers 1.40.9__py3-none-any.whl → 1.40.10__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- x_transformers/x_transformers.py +1 -1
- {x_transformers-1.40.9.dist-info → x_transformers-1.40.10.dist-info}/METADATA +1 -1
- {x_transformers-1.40.9.dist-info → x_transformers-1.40.10.dist-info}/RECORD +6 -6
- {x_transformers-1.40.9.dist-info → x_transformers-1.40.10.dist-info}/LICENSE +0 -0
- {x_transformers-1.40.9.dist-info → x_transformers-1.40.10.dist-info}/WHEEL +0 -0
- {x_transformers-1.40.9.dist-info → x_transformers-1.40.10.dist-info}/top_level.txt +0 -0
x_transformers/x_transformers.py
CHANGED
@@ -1019,7 +1019,7 @@ class Attention(Module):
|
|
1019
1019
|
self.qk_norm_q_scale = self.qk_norm_k_scale = 1
|
1020
1020
|
if qk_norm and qk_norm_dim_scale:
|
1021
1021
|
self.qk_norm_q_scale = nn.Parameter(torch.ones(heads, 1, dim_head))
|
1022
|
-
self.qk_norm_k_scale = nn.Parameter(torch.ones(
|
1022
|
+
self.qk_norm_k_scale = nn.Parameter(torch.ones(kv_heads, 1, dim_head))
|
1023
1023
|
|
1024
1024
|
assert (not qk_norm) or divisible_by(dim_head, qk_norm_groups), 'dimension per attention head must be divisible by the qk norm groups'
|
1025
1025
|
assert not (qk_norm and (dim_head // qk_norm_groups) <= 2), 'the group dimension may be too small (2 was too small in my tests, but 4 still works, surprisingly)'
|
@@ -5,11 +5,11 @@ x_transformers/continuous.py,sha256=cIVEdhfei258__ziV7kQBrJMxCel54bExBTDrO9rfCI,
|
|
5
5
|
x_transformers/dpo.py,sha256=LjvWgCkqTl-UuehrzQ8nkX5guLr4whYwsmm7SKSwdls,3450
|
6
6
|
x_transformers/multi_input.py,sha256=tCh-fTJDj2ib4SMGtsa-AM8MxKzJAQSwqAXOu3HU2mg,9252
|
7
7
|
x_transformers/nonautoregressive_wrapper.py,sha256=2NU58hYMgn-4Jzg3mie-mXb0XH_dCN7fjlzd3K1rLUY,10510
|
8
|
-
x_transformers/x_transformers.py,sha256=
|
8
|
+
x_transformers/x_transformers.py,sha256=R3kIKNEw6wOoH3RKSx0ceHC33Hj3qVrDxaEfVbGKzHk,88075
|
9
9
|
x_transformers/xl_autoregressive_wrapper.py,sha256=CvZMJ6A6PA-Y_bQAhnORwjJBSl6Vjq2IdW5KTdk8NI8,4195
|
10
10
|
x_transformers/xval.py,sha256=7S00kCuab4tWQa-vf-z-XfzADjVj48MoFIr7VSIvttg,8575
|
11
|
-
x_transformers-1.40.
|
12
|
-
x_transformers-1.40.
|
13
|
-
x_transformers-1.40.
|
14
|
-
x_transformers-1.40.
|
15
|
-
x_transformers-1.40.
|
11
|
+
x_transformers-1.40.10.dist-info/LICENSE,sha256=As9u198X-U-vph5noInuUfqsAG2zX_oXPHDmdjwlPPY,1066
|
12
|
+
x_transformers-1.40.10.dist-info/METADATA,sha256=ve1-9D3mwdz2euWxr6ijTkaDAv7yvJfhflsF9IaBUDw,662
|
13
|
+
x_transformers-1.40.10.dist-info/WHEEL,sha256=OVMc5UfuAQiSplgO0_WdW7vXVGAt9Hdd6qtN4HotdyA,91
|
14
|
+
x_transformers-1.40.10.dist-info/top_level.txt,sha256=hO6KGpFuGucRNEtRfme4A_rGcM53AKwGP7RVlRIxS5Q,15
|
15
|
+
x_transformers-1.40.10.dist-info/RECORD,,
|
File without changes
|
File without changes
|
File without changes
|