x-transformers 1.35.0__py3-none-any.whl → 1.35.1__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- x_transformers/autoregressive_wrapper.py +3 -2
- {x_transformers-1.35.0.dist-info → x_transformers-1.35.1.dist-info}/METADATA +1 -1
- {x_transformers-1.35.0.dist-info → x_transformers-1.35.1.dist-info}/RECORD +6 -6
- {x_transformers-1.35.0.dist-info → x_transformers-1.35.1.dist-info}/LICENSE +0 -0
- {x_transformers-1.35.0.dist-info → x_transformers-1.35.1.dist-info}/WHEEL +0 -0
- {x_transformers-1.35.0.dist-info → x_transformers-1.35.1.dist-info}/top_level.txt +0 -0
@@ -220,13 +220,14 @@ class AutoregressiveWrapper(Module):
|
|
220
220
|
if restrict_to_max_seq_len:
|
221
221
|
max_len_exceeded = out.shape[-1] > max_seq_len
|
222
222
|
|
223
|
-
assert not (cache_kv and max_len_exceeded and not self.net.can_cache_kv_outside_max_seq_len), 'the network cannot use cached key values when decoding outside the max sequence length. most likely because you are using absolute positional
|
223
|
+
assert not (cache_kv and max_len_exceeded and not self.net.can_cache_kv_outside_max_seq_len), 'the network cannot use cached key values when decoding outside the max sequence length. most likely because you are using absolute positional embedding. you can switch to rotary embeddings to resolve this issue'
|
224
224
|
|
225
225
|
x = out[:, -max_seq_len:]
|
226
226
|
|
227
227
|
if exists(cache):
|
228
228
|
for inter in cache.attn_intermediates:
|
229
|
-
inter.
|
229
|
+
if inter.layer_type == 'a':
|
230
|
+
inter.cached_kv = [t[..., -(max_seq_len - 1):, :] for t in inter.cached_kv]
|
230
231
|
|
231
232
|
logits, new_cache = self.net(
|
232
233
|
x,
|
@@ -1,6 +1,6 @@
|
|
1
1
|
x_transformers/__init__.py,sha256=-MkQrSc37cTVDX7AOykxunYnqVtFlQ7lb0Cse5dsGWU,793
|
2
2
|
x_transformers/attend.py,sha256=7q996VGYHGIsc0FQnN8WNiwHn3xny3i1biRwx7yW5vg,12090
|
3
|
-
x_transformers/autoregressive_wrapper.py,sha256=
|
3
|
+
x_transformers/autoregressive_wrapper.py,sha256=pDymmnPgWQoH7wwHKskI_gktsdQX-LysnQtIozodYGU,10422
|
4
4
|
x_transformers/continuous.py,sha256=cIVEdhfei258__ziV7kQBrJMxCel54bExBTDrO9rfCI,6450
|
5
5
|
x_transformers/dpo.py,sha256=LjvWgCkqTl-UuehrzQ8nkX5guLr4whYwsmm7SKSwdls,3450
|
6
6
|
x_transformers/multi_input.py,sha256=tCh-fTJDj2ib4SMGtsa-AM8MxKzJAQSwqAXOu3HU2mg,9252
|
@@ -8,8 +8,8 @@ x_transformers/nonautoregressive_wrapper.py,sha256=ys_p8obc7lTeeodCqvkRKxOXQ1C9T
|
|
8
8
|
x_transformers/x_transformers.py,sha256=2oQoQs7RMbFrVdMeOddy6yq1MhJxnficjORmMWBjjPo,80593
|
9
9
|
x_transformers/xl_autoregressive_wrapper.py,sha256=DCx4n0_c1tFai4nOqaWVnqx2p9eutsZsDMiMP1ckxNU,4117
|
10
10
|
x_transformers/xval.py,sha256=QE1ltYZTR_eGgIHPP2BrMWVWVLqMW-OpDZh87BSmQEg,8563
|
11
|
-
x_transformers-1.35.
|
12
|
-
x_transformers-1.35.
|
13
|
-
x_transformers-1.35.
|
14
|
-
x_transformers-1.35.
|
15
|
-
x_transformers-1.35.
|
11
|
+
x_transformers-1.35.1.dist-info/LICENSE,sha256=As9u198X-U-vph5noInuUfqsAG2zX_oXPHDmdjwlPPY,1066
|
12
|
+
x_transformers-1.35.1.dist-info/METADATA,sha256=4tuROkk-uVjndRdJpYZzuvTUNC4OHgECzUinyrW14Ns,661
|
13
|
+
x_transformers-1.35.1.dist-info/WHEEL,sha256=cVxcB9AmuTcXqmwrtPhNK88dr7IR_b6qagTj0UvIEbY,91
|
14
|
+
x_transformers-1.35.1.dist-info/top_level.txt,sha256=hO6KGpFuGucRNEtRfme4A_rGcM53AKwGP7RVlRIxS5Q,15
|
15
|
+
x_transformers-1.35.1.dist-info/RECORD,,
|
File without changes
|
File without changes
|
File without changes
|