x-transformers 1.43.1__py3-none-any.whl → 1.43.4__py3-none-any.whl
Sign up to get free protection for your applications and to get access to all the features.
- x_transformers/x_transformers.py +6 -6
- {x_transformers-1.43.1.dist-info → x_transformers-1.43.4.dist-info}/METADATA +1 -1
- {x_transformers-1.43.1.dist-info → x_transformers-1.43.4.dist-info}/RECORD +6 -6
- {x_transformers-1.43.1.dist-info → x_transformers-1.43.4.dist-info}/LICENSE +0 -0
- {x_transformers-1.43.1.dist-info → x_transformers-1.43.4.dist-info}/WHEEL +0 -0
- {x_transformers-1.43.1.dist-info → x_transformers-1.43.4.dist-info}/top_level.txt +0 -0
x_transformers/x_transformers.py
CHANGED
@@ -666,7 +666,7 @@ class RotaryEmbedding(Module):
|
|
666
666
|
return freqs, 1.
|
667
667
|
|
668
668
|
power = (t - (max_pos // 2)) / self.scale_base
|
669
|
-
scale = self.scale ** rearrange(power, 'n -> n 1')
|
669
|
+
scale = self.scale ** rearrange(power, '... n -> ... n 1')
|
670
670
|
scale = torch.stack((scale, scale), dim = -1)
|
671
671
|
scale = rearrange(scale, '... d r -> ... (d r)')
|
672
672
|
|
@@ -2270,16 +2270,16 @@ class AttentionLayers(Module):
|
|
2270
2270
|
if self.need_condition:
|
2271
2271
|
final_norm = maybe(partial)(final_norm, **norm_kwargs)
|
2272
2272
|
|
2273
|
-
if self.resi_dual:
|
2274
|
-
x = x + final_norm(outer_residual)
|
2275
|
-
else:
|
2276
|
-
x = final_norm(x)
|
2277
|
-
|
2278
2273
|
# take care of multistreams if needed, use sum for now
|
2279
2274
|
|
2280
2275
|
if is_multistream:
|
2281
2276
|
x = reduce(x, '(b s) n d -> b n d', 'sum', s = streams)
|
2282
2277
|
|
2278
|
+
if self.resi_dual:
|
2279
|
+
x = x + final_norm(outer_residual)
|
2280
|
+
else:
|
2281
|
+
x = final_norm(x)
|
2282
|
+
|
2283
2283
|
if not return_hiddens:
|
2284
2284
|
return x
|
2285
2285
|
|
@@ -6,11 +6,11 @@ x_transformers/dpo.py,sha256=xt4OuOWhU8pN3OKN2LZAaC2NC8iiEnchqqcrPWVqf0o,3521
|
|
6
6
|
x_transformers/multi_input.py,sha256=tCh-fTJDj2ib4SMGtsa-AM8MxKzJAQSwqAXOu3HU2mg,9252
|
7
7
|
x_transformers/neo_mlp.py,sha256=XCNnnop9WLarcxap1kGuYc1x8GHvwkZiDRnXOxSl3Po,3452
|
8
8
|
x_transformers/nonautoregressive_wrapper.py,sha256=2NU58hYMgn-4Jzg3mie-mXb0XH_dCN7fjlzd3K1rLUY,10510
|
9
|
-
x_transformers/x_transformers.py,sha256=
|
9
|
+
x_transformers/x_transformers.py,sha256=gn0vRtwbjBA67T-Z8tkU-k3Xte0PaMTxZlmzdK8UsFw,100392
|
10
10
|
x_transformers/xl_autoregressive_wrapper.py,sha256=CvZMJ6A6PA-Y_bQAhnORwjJBSl6Vjq2IdW5KTdk8NI8,4195
|
11
11
|
x_transformers/xval.py,sha256=7S00kCuab4tWQa-vf-z-XfzADjVj48MoFIr7VSIvttg,8575
|
12
|
-
x_transformers-1.43.
|
13
|
-
x_transformers-1.43.
|
14
|
-
x_transformers-1.43.
|
15
|
-
x_transformers-1.43.
|
16
|
-
x_transformers-1.43.
|
12
|
+
x_transformers-1.43.4.dist-info/LICENSE,sha256=As9u198X-U-vph5noInuUfqsAG2zX_oXPHDmdjwlPPY,1066
|
13
|
+
x_transformers-1.43.4.dist-info/METADATA,sha256=Nlj9DcMqnMxJH-xR4Dwd4aU1U-UQIUshpQaMDcggVes,738
|
14
|
+
x_transformers-1.43.4.dist-info/WHEEL,sha256=PZUExdf71Ui_so67QXpySuHtCi3-J3wvF4ORK6k_S8U,91
|
15
|
+
x_transformers-1.43.4.dist-info/top_level.txt,sha256=hO6KGpFuGucRNEtRfme4A_rGcM53AKwGP7RVlRIxS5Q,15
|
16
|
+
x_transformers-1.43.4.dist-info/RECORD,,
|
File without changes
|
File without changes
|
File without changes
|