x-transformers 1.43.1__py3-none-any.whl → 1.43.4__py3-none-any.whl

Sign up to get free protection for your applications and to get access to all the features.
@@ -666,7 +666,7 @@ class RotaryEmbedding(Module):
666
666
  return freqs, 1.
667
667
 
668
668
  power = (t - (max_pos // 2)) / self.scale_base
669
- scale = self.scale ** rearrange(power, 'n -> n 1')
669
+ scale = self.scale ** rearrange(power, '... n -> ... n 1')
670
670
  scale = torch.stack((scale, scale), dim = -1)
671
671
  scale = rearrange(scale, '... d r -> ... (d r)')
672
672
 
@@ -2270,16 +2270,16 @@ class AttentionLayers(Module):
2270
2270
  if self.need_condition:
2271
2271
  final_norm = maybe(partial)(final_norm, **norm_kwargs)
2272
2272
 
2273
- if self.resi_dual:
2274
- x = x + final_norm(outer_residual)
2275
- else:
2276
- x = final_norm(x)
2277
-
2278
2273
  # take care of multistreams if needed, use sum for now
2279
2274
 
2280
2275
  if is_multistream:
2281
2276
  x = reduce(x, '(b s) n d -> b n d', 'sum', s = streams)
2282
2277
 
2278
+ if self.resi_dual:
2279
+ x = x + final_norm(outer_residual)
2280
+ else:
2281
+ x = final_norm(x)
2282
+
2283
2283
  if not return_hiddens:
2284
2284
  return x
2285
2285
 
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.1
2
2
  Name: x-transformers
3
- Version: 1.43.1
3
+ Version: 1.43.4
4
4
  Summary: X-Transformers - Pytorch
5
5
  Home-page: https://github.com/lucidrains/x-transformers
6
6
  Author: Phil Wang
@@ -6,11 +6,11 @@ x_transformers/dpo.py,sha256=xt4OuOWhU8pN3OKN2LZAaC2NC8iiEnchqqcrPWVqf0o,3521
6
6
  x_transformers/multi_input.py,sha256=tCh-fTJDj2ib4SMGtsa-AM8MxKzJAQSwqAXOu3HU2mg,9252
7
7
  x_transformers/neo_mlp.py,sha256=XCNnnop9WLarcxap1kGuYc1x8GHvwkZiDRnXOxSl3Po,3452
8
8
  x_transformers/nonautoregressive_wrapper.py,sha256=2NU58hYMgn-4Jzg3mie-mXb0XH_dCN7fjlzd3K1rLUY,10510
9
- x_transformers/x_transformers.py,sha256=JG38kcXdhRBKT5_FHMhV5dQabSGrAHsuIQkHjPalDiI,100384
9
+ x_transformers/x_transformers.py,sha256=gn0vRtwbjBA67T-Z8tkU-k3Xte0PaMTxZlmzdK8UsFw,100392
10
10
  x_transformers/xl_autoregressive_wrapper.py,sha256=CvZMJ6A6PA-Y_bQAhnORwjJBSl6Vjq2IdW5KTdk8NI8,4195
11
11
  x_transformers/xval.py,sha256=7S00kCuab4tWQa-vf-z-XfzADjVj48MoFIr7VSIvttg,8575
12
- x_transformers-1.43.1.dist-info/LICENSE,sha256=As9u198X-U-vph5noInuUfqsAG2zX_oXPHDmdjwlPPY,1066
13
- x_transformers-1.43.1.dist-info/METADATA,sha256=V57c6Bps0GjG0GLEBpxkHdbvxIWzXss2Xu5_KQJJXPc,738
14
- x_transformers-1.43.1.dist-info/WHEEL,sha256=PZUExdf71Ui_so67QXpySuHtCi3-J3wvF4ORK6k_S8U,91
15
- x_transformers-1.43.1.dist-info/top_level.txt,sha256=hO6KGpFuGucRNEtRfme4A_rGcM53AKwGP7RVlRIxS5Q,15
16
- x_transformers-1.43.1.dist-info/RECORD,,
12
+ x_transformers-1.43.4.dist-info/LICENSE,sha256=As9u198X-U-vph5noInuUfqsAG2zX_oXPHDmdjwlPPY,1066
13
+ x_transformers-1.43.4.dist-info/METADATA,sha256=Nlj9DcMqnMxJH-xR4Dwd4aU1U-UQIUshpQaMDcggVes,738
14
+ x_transformers-1.43.4.dist-info/WHEEL,sha256=PZUExdf71Ui_so67QXpySuHtCi3-J3wvF4ORK6k_S8U,91
15
+ x_transformers-1.43.4.dist-info/top_level.txt,sha256=hO6KGpFuGucRNEtRfme4A_rGcM53AKwGP7RVlRIxS5Q,15
16
+ x_transformers-1.43.4.dist-info/RECORD,,