x-transformers 1.31.9__tar.gz → 1.31.11__tar.gz

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (20) hide show
  1. {x_transformers-1.31.9/x_transformers.egg-info → x_transformers-1.31.11}/PKG-INFO +1 -1
  2. {x_transformers-1.31.9 → x_transformers-1.31.11}/setup.py +1 -1
  3. {x_transformers-1.31.9 → x_transformers-1.31.11}/x_transformers/__init__.py +1 -0
  4. {x_transformers-1.31.9 → x_transformers-1.31.11}/x_transformers/attend.py +1 -1
  5. {x_transformers-1.31.9 → x_transformers-1.31.11}/x_transformers/x_transformers.py +1 -1
  6. {x_transformers-1.31.9 → x_transformers-1.31.11/x_transformers.egg-info}/PKG-INFO +1 -1
  7. {x_transformers-1.31.9 → x_transformers-1.31.11}/LICENSE +0 -0
  8. {x_transformers-1.31.9 → x_transformers-1.31.11}/README.md +0 -0
  9. {x_transformers-1.31.9 → x_transformers-1.31.11}/setup.cfg +0 -0
  10. {x_transformers-1.31.9 → x_transformers-1.31.11}/tests/test_x_transformers.py +0 -0
  11. {x_transformers-1.31.9 → x_transformers-1.31.11}/x_transformers/autoregressive_wrapper.py +0 -0
  12. {x_transformers-1.31.9 → x_transformers-1.31.11}/x_transformers/continuous.py +0 -0
  13. {x_transformers-1.31.9 → x_transformers-1.31.11}/x_transformers/dpo.py +0 -0
  14. {x_transformers-1.31.9 → x_transformers-1.31.11}/x_transformers/nonautoregressive_wrapper.py +0 -0
  15. {x_transformers-1.31.9 → x_transformers-1.31.11}/x_transformers/xl_autoregressive_wrapper.py +0 -0
  16. {x_transformers-1.31.9 → x_transformers-1.31.11}/x_transformers/xval.py +0 -0
  17. {x_transformers-1.31.9 → x_transformers-1.31.11}/x_transformers.egg-info/SOURCES.txt +0 -0
  18. {x_transformers-1.31.9 → x_transformers-1.31.11}/x_transformers.egg-info/dependency_links.txt +0 -0
  19. {x_transformers-1.31.9 → x_transformers-1.31.11}/x_transformers.egg-info/requires.txt +0 -0
  20. {x_transformers-1.31.9 → x_transformers-1.31.11}/x_transformers.egg-info/top_level.txt +0 -0
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.1
2
2
  Name: x-transformers
3
- Version: 1.31.9
3
+ Version: 1.31.11
4
4
  Summary: X-Transformers - Pytorch
5
5
  Home-page: https://github.com/lucidrains/x-transformers
6
6
  Author: Phil Wang
@@ -3,7 +3,7 @@ from setuptools import setup, find_packages
3
3
  setup(
4
4
  name = 'x-transformers',
5
5
  packages = find_packages(exclude=['examples']),
6
- version = '1.31.9',
6
+ version = '1.31.11',
7
7
  license='MIT',
8
8
  description = 'X-Transformers - Pytorch',
9
9
  author = 'Phil Wang',
@@ -7,6 +7,7 @@ from x_transformers.x_transformers import (
7
7
  Attention,
8
8
  FeedForward,
9
9
  RMSNorm,
10
+ AdaptiveRMSNorm,
10
11
  TransformerWrapper,
11
12
  ViTransformerWrapper
12
13
  )
@@ -82,7 +82,7 @@ class Attend(Module):
82
82
  qk_norm = False,
83
83
  flash = False,
84
84
  softclamp_logits = False,
85
- logit_softclamp_value = 30.,
85
+ logit_softclamp_value = 50.,
86
86
  add_zero_kv = False,
87
87
  cope = None,
88
88
  onnxable = False,
@@ -1272,7 +1272,7 @@ class AttentionLayers(Module):
1272
1272
  shift_tokens = 0,
1273
1273
  sandwich_norm = False,
1274
1274
  softclamp_output = False,
1275
- softclamp_output_value = 50.,
1275
+ softclamp_output_value = 30.,
1276
1276
  resi_dual = False,
1277
1277
  resi_dual_scale = 1.,
1278
1278
  zero_init_branch_output = False,
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.1
2
2
  Name: x-transformers
3
- Version: 1.31.9
3
+ Version: 1.31.11
4
4
  Summary: X-Transformers - Pytorch
5
5
  Home-page: https://github.com/lucidrains/x-transformers
6
6
  Author: Phil Wang