x-transformers 1.44.4__py3-none-any.whl → 1.44.5__py3-none-any.whl
Sign up to get free protection for your applications and to get access to all the features.
- x_transformers/x_transformers.py +2 -2
- {x_transformers-1.44.4.dist-info → x_transformers-1.44.5.dist-info}/METADATA +11 -2
- {x_transformers-1.44.4.dist-info → x_transformers-1.44.5.dist-info}/RECORD +6 -6
- {x_transformers-1.44.4.dist-info → x_transformers-1.44.5.dist-info}/WHEEL +1 -1
- {x_transformers-1.44.4.dist-info → x_transformers-1.44.5.dist-info}/LICENSE +0 -0
- {x_transformers-1.44.4.dist-info → x_transformers-1.44.5.dist-info}/top_level.txt +0 -0
x_transformers/x_transformers.py
CHANGED
@@ -2224,7 +2224,7 @@ class AttentionLayers(Module):
|
|
2224
2224
|
layer_variables = tuple(tuple(layer_variable[i] for i in layers_execute_order) for layer_variable in layer_variables)
|
2225
2225
|
|
2226
2226
|
# derived input for reinjection if needed
|
2227
|
-
|
2227
|
+
inp_inject = None
|
2228
2228
|
if self.reinject_input:
|
2229
2229
|
assert not exists(in_attn_cond)
|
2230
2230
|
inp_inject = self.reinject_input_proj(x)
|
@@ -2282,7 +2282,7 @@ class AttentionLayers(Module):
|
|
2282
2282
|
post_branch_norm = maybe(partial)(post_branch_norm, **norm_kwargs)
|
2283
2283
|
post_main_norm = maybe(partial)(post_main_norm, **norm_kwargs)
|
2284
2284
|
|
2285
|
-
if
|
2285
|
+
if exists(inp_inject):
|
2286
2286
|
x = x + inp_inject
|
2287
2287
|
|
2288
2288
|
if exists(pre_norm):
|
@@ -1,6 +1,6 @@
|
|
1
|
-
Metadata-Version: 2.
|
1
|
+
Metadata-Version: 2.2
|
2
2
|
Name: x-transformers
|
3
|
-
Version: 1.44.
|
3
|
+
Version: 1.44.5
|
4
4
|
Summary: X-Transformers - Pytorch
|
5
5
|
Home-page: https://github.com/lucidrains/x-transformers
|
6
6
|
Author: Phil Wang
|
@@ -19,3 +19,12 @@ Requires-Dist: einops>=0.8.0
|
|
19
19
|
Requires-Dist: loguru
|
20
20
|
Requires-Dist: packaging>=21.0
|
21
21
|
Requires-Dist: torch>=2.0
|
22
|
+
Dynamic: author
|
23
|
+
Dynamic: author-email
|
24
|
+
Dynamic: classifier
|
25
|
+
Dynamic: description-content-type
|
26
|
+
Dynamic: home-page
|
27
|
+
Dynamic: keywords
|
28
|
+
Dynamic: license
|
29
|
+
Dynamic: requires-dist
|
30
|
+
Dynamic: summary
|
@@ -6,11 +6,11 @@ x_transformers/dpo.py,sha256=xt4OuOWhU8pN3OKN2LZAaC2NC8iiEnchqqcrPWVqf0o,3521
|
|
6
6
|
x_transformers/multi_input.py,sha256=tCh-fTJDj2ib4SMGtsa-AM8MxKzJAQSwqAXOu3HU2mg,9252
|
7
7
|
x_transformers/neo_mlp.py,sha256=XCNnnop9WLarcxap1kGuYc1x8GHvwkZiDRnXOxSl3Po,3452
|
8
8
|
x_transformers/nonautoregressive_wrapper.py,sha256=2NU58hYMgn-4Jzg3mie-mXb0XH_dCN7fjlzd3K1rLUY,10510
|
9
|
-
x_transformers/x_transformers.py,sha256
|
9
|
+
x_transformers/x_transformers.py,sha256=-B4xutlod09EwaA1tE24GgD-5ioi4HAXIwmI5N4MYio,103124
|
10
10
|
x_transformers/xl_autoregressive_wrapper.py,sha256=CvZMJ6A6PA-Y_bQAhnORwjJBSl6Vjq2IdW5KTdk8NI8,4195
|
11
11
|
x_transformers/xval.py,sha256=7S00kCuab4tWQa-vf-z-XfzADjVj48MoFIr7VSIvttg,8575
|
12
|
-
x_transformers-1.44.
|
13
|
-
x_transformers-1.44.
|
14
|
-
x_transformers-1.44.
|
15
|
-
x_transformers-1.44.
|
16
|
-
x_transformers-1.44.
|
12
|
+
x_transformers-1.44.5.dist-info/LICENSE,sha256=As9u198X-U-vph5noInuUfqsAG2zX_oXPHDmdjwlPPY,1066
|
13
|
+
x_transformers-1.44.5.dist-info/METADATA,sha256=j_8nbdft6Qx1NtIjSd1NyERsrpu-W9L3-CuR0hug_S0,924
|
14
|
+
x_transformers-1.44.5.dist-info/WHEEL,sha256=In9FTNxeP60KnTkGw7wk6mJPYd_dQSjEZmXdBdMCI-8,91
|
15
|
+
x_transformers-1.44.5.dist-info/top_level.txt,sha256=hO6KGpFuGucRNEtRfme4A_rGcM53AKwGP7RVlRIxS5Q,15
|
16
|
+
x_transformers-1.44.5.dist-info/RECORD,,
|
File without changes
|
File without changes
|