x-transformers 1.30.22__py3-none-any.whl → 1.30.23__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1201,8 +1201,8 @@ class AttentionLayers(Module):
1201
1201
  rotary_xpos_scale_base = 512,
1202
1202
  rotary_base_rescale_factor = 1.,
1203
1203
  weight_tie_layers = False,
1204
- custom_layers: Tuple[str] | None = None,
1205
- layers_execute_order: Tuple[int] | None = None,
1204
+ custom_layers: Tuple[str, ...] | None = None,
1205
+ layers_execute_order: Tuple[int, ...] | None = None,
1206
1206
  sandwich_coef = None,
1207
1207
  par_ratio = None,
1208
1208
  residual_attn = False,
@@ -1484,7 +1484,8 @@ class AttentionLayers(Module):
1484
1484
  cache_age = 1,
1485
1485
  return_hiddens = False,
1486
1486
  rotary_pos_emb = None,
1487
- condition = None
1487
+ condition = None,
1488
+ layers_execute_order: Tuple[int, ...] | None = None
1488
1489
  ):
1489
1490
  assert not (self.cross_attend ^ exists(context)), 'context must be passed in if cross_attend is set to True'
1490
1491
  assert not (exists(condition) ^ self.need_condition), 'condition needs to be passed in if using adaptive layernorm or vice versa'
@@ -1576,7 +1577,11 @@ class AttentionLayers(Module):
1576
1577
  self.layer_dropouts
1577
1578
  )
1578
1579
 
1579
- layer_variables = tuple(tuple(layer_variable[i] for i in self.layers_execute_order) for layer_variable in layer_variables)
1580
+ # able to override the layers execution order on forward, for trying to depth extrapolate
1581
+
1582
+ layers_execute_order = default(layers_execute_order, self.layers_execute_order)
1583
+
1584
+ layer_variables = tuple(tuple(layer_variable[i] for i in layers_execute_order) for layer_variable in layer_variables)
1580
1585
 
1581
1586
  # go through the attention and feedforward layers
1582
1587
 
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.1
2
2
  Name: x-transformers
3
- Version: 1.30.22
3
+ Version: 1.30.23
4
4
  Summary: X-Transformers - Pytorch
5
5
  Home-page: https://github.com/lucidrains/x-transformers
6
6
  Author: Phil Wang
@@ -4,11 +4,11 @@ x_transformers/autoregressive_wrapper.py,sha256=uX8Mb0zLsQrZECt_9UGt35g7tC05Rk3n
4
4
  x_transformers/continuous.py,sha256=WO52n9lFAXv5-SGadi2cApGF8dkouN8QSTEOuC7erj8,6180
5
5
  x_transformers/dpo.py,sha256=LjvWgCkqTl-UuehrzQ8nkX5guLr4whYwsmm7SKSwdls,3450
6
6
  x_transformers/nonautoregressive_wrapper.py,sha256=ys_p8obc7lTeeodCqvkRKxOXQ1C9T3j5Jwr-JbVgnXk,10432
7
- x_transformers/x_transformers.py,sha256=z1LzQMpJSNgalsnZJB-EsxqQk6aF0rQrv98VkiBTHnc,73959
7
+ x_transformers/x_transformers.py,sha256=tZixUvlsaEj3CpB49KLDOJ2BwYSPjdWotDUjB9Rbf7g,74213
8
8
  x_transformers/xl_autoregressive_wrapper.py,sha256=DCx4n0_c1tFai4nOqaWVnqx2p9eutsZsDMiMP1ckxNU,4117
9
9
  x_transformers/xval.py,sha256=QE1ltYZTR_eGgIHPP2BrMWVWVLqMW-OpDZh87BSmQEg,8563
10
- x_transformers-1.30.22.dist-info/LICENSE,sha256=As9u198X-U-vph5noInuUfqsAG2zX_oXPHDmdjwlPPY,1066
11
- x_transformers-1.30.22.dist-info/METADATA,sha256=aU33r8OyivazXN7QW5np0LdRaZGJY3x7pBSx55iFGdA,662
12
- x_transformers-1.30.22.dist-info/WHEEL,sha256=cpQTJ5IWu9CdaPViMhC9YzF8gZuS5-vlfoFihTBC86A,91
13
- x_transformers-1.30.22.dist-info/top_level.txt,sha256=hO6KGpFuGucRNEtRfme4A_rGcM53AKwGP7RVlRIxS5Q,15
14
- x_transformers-1.30.22.dist-info/RECORD,,
10
+ x_transformers-1.30.23.dist-info/LICENSE,sha256=As9u198X-U-vph5noInuUfqsAG2zX_oXPHDmdjwlPPY,1066
11
+ x_transformers-1.30.23.dist-info/METADATA,sha256=LM8Y0bkOF259zCn_FE2A-Uw5Yjr8YrqCKNYuW4DqtQY,662
12
+ x_transformers-1.30.23.dist-info/WHEEL,sha256=cpQTJ5IWu9CdaPViMhC9YzF8gZuS5-vlfoFihTBC86A,91
13
+ x_transformers-1.30.23.dist-info/top_level.txt,sha256=hO6KGpFuGucRNEtRfme4A_rGcM53AKwGP7RVlRIxS5Q,15
14
+ x_transformers-1.30.23.dist-info/RECORD,,