x-transformers 1.43.1__tar.gz → 1.43.2__tar.gz
Sign up to get free protection for your applications and to get access to all the features.
- {x_transformers-1.43.1/x_transformers.egg-info → x_transformers-1.43.2}/PKG-INFO +1 -1
- {x_transformers-1.43.1 → x_transformers-1.43.2}/setup.py +1 -1
- {x_transformers-1.43.1 → x_transformers-1.43.2}/tests/test_x_transformers.py +4 -2
- {x_transformers-1.43.1 → x_transformers-1.43.2}/x_transformers/x_transformers.py +1 -1
- {x_transformers-1.43.1 → x_transformers-1.43.2/x_transformers.egg-info}/PKG-INFO +1 -1
- {x_transformers-1.43.1 → x_transformers-1.43.2}/LICENSE +0 -0
- {x_transformers-1.43.1 → x_transformers-1.43.2}/README.md +0 -0
- {x_transformers-1.43.1 → x_transformers-1.43.2}/setup.cfg +0 -0
- {x_transformers-1.43.1 → x_transformers-1.43.2}/x_transformers/__init__.py +0 -0
- {x_transformers-1.43.1 → x_transformers-1.43.2}/x_transformers/attend.py +0 -0
- {x_transformers-1.43.1 → x_transformers-1.43.2}/x_transformers/autoregressive_wrapper.py +0 -0
- {x_transformers-1.43.1 → x_transformers-1.43.2}/x_transformers/continuous.py +0 -0
- {x_transformers-1.43.1 → x_transformers-1.43.2}/x_transformers/dpo.py +0 -0
- {x_transformers-1.43.1 → x_transformers-1.43.2}/x_transformers/multi_input.py +0 -0
- {x_transformers-1.43.1 → x_transformers-1.43.2}/x_transformers/neo_mlp.py +0 -0
- {x_transformers-1.43.1 → x_transformers-1.43.2}/x_transformers/nonautoregressive_wrapper.py +0 -0
- {x_transformers-1.43.1 → x_transformers-1.43.2}/x_transformers/xl_autoregressive_wrapper.py +0 -0
- {x_transformers-1.43.1 → x_transformers-1.43.2}/x_transformers/xval.py +0 -0
- {x_transformers-1.43.1 → x_transformers-1.43.2}/x_transformers.egg-info/SOURCES.txt +0 -0
- {x_transformers-1.43.1 → x_transformers-1.43.2}/x_transformers.egg-info/dependency_links.txt +0 -0
- {x_transformers-1.43.1 → x_transformers-1.43.2}/x_transformers.egg-info/requires.txt +0 -0
- {x_transformers-1.43.1 → x_transformers-1.43.2}/x_transformers.egg-info/top_level.txt +0 -0
@@ -409,7 +409,8 @@ def test_custom_alibi(flash: bool):
|
|
409
409
|
|
410
410
|
logits = model(x, pos = pos)
|
411
411
|
|
412
|
-
|
412
|
+
@pytest.mark.parametrize('rotary_xpos', (True, False))
|
413
|
+
def test_custom_rotary_pos_emb(rotary_xpos):
|
413
414
|
from einops import repeat
|
414
415
|
|
415
416
|
model = TransformerWrapper(
|
@@ -419,7 +420,8 @@ def test_custom_rotary_pos_emb():
|
|
419
420
|
dim = 512,
|
420
421
|
depth = 2,
|
421
422
|
heads = 8,
|
422
|
-
rotary_pos_emb = True
|
423
|
+
rotary_pos_emb = True,
|
424
|
+
rotary_xpos = rotary_xpos
|
423
425
|
)
|
424
426
|
)
|
425
427
|
|
@@ -666,7 +666,7 @@ class RotaryEmbedding(Module):
|
|
666
666
|
return freqs, 1.
|
667
667
|
|
668
668
|
power = (t - (max_pos // 2)) / self.scale_base
|
669
|
-
scale = self.scale ** rearrange(power, 'n -> n 1')
|
669
|
+
scale = self.scale ** rearrange(power, '... n -> ... n 1')
|
670
670
|
scale = torch.stack((scale, scale), dim = -1)
|
671
671
|
scale = rearrange(scale, '... d r -> ... (d r)')
|
672
672
|
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
{x_transformers-1.43.1 → x_transformers-1.43.2}/x_transformers.egg-info/dependency_links.txt
RENAMED
File without changes
|
File without changes
|
File without changes
|