x-transformers 2.7.1__tar.gz → 2.7.2__tar.gz

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (63) hide show
  1. {x_transformers-2.7.1 → x_transformers-2.7.2}/PKG-INFO +1 -1
  2. {x_transformers-2.7.1 → x_transformers-2.7.2}/pyproject.toml +1 -1
  3. {x_transformers-2.7.1 → x_transformers-2.7.2}/x_transformers/x_transformers.py +5 -0
  4. {x_transformers-2.7.1 → x_transformers-2.7.2}/.github/FUNDING.yml +0 -0
  5. {x_transformers-2.7.1 → x_transformers-2.7.2}/.github/workflows/python-publish.yml +0 -0
  6. {x_transformers-2.7.1 → x_transformers-2.7.2}/.github/workflows/python-test.yaml +0 -0
  7. {x_transformers-2.7.1 → x_transformers-2.7.2}/.gitignore +0 -0
  8. {x_transformers-2.7.1 → x_transformers-2.7.2}/LICENSE +0 -0
  9. {x_transformers-2.7.1 → x_transformers-2.7.2}/README.md +0 -0
  10. {x_transformers-2.7.1 → x_transformers-2.7.2}/data/README.md +0 -0
  11. {x_transformers-2.7.1 → x_transformers-2.7.2}/data/enwik8.gz +0 -0
  12. {x_transformers-2.7.1 → x_transformers-2.7.2}/images/all-attention.png +0 -0
  13. {x_transformers-2.7.1 → x_transformers-2.7.2}/images/attention-on-attention.png +0 -0
  14. {x_transformers-2.7.1 → x_transformers-2.7.2}/images/cosine-sim-attention.png +0 -0
  15. {x_transformers-2.7.1 → x_transformers-2.7.2}/images/deepnorm.png +0 -0
  16. {x_transformers-2.7.1 → x_transformers-2.7.2}/images/dynamic-pos-bias-linear.png +0 -0
  17. {x_transformers-2.7.1 → x_transformers-2.7.2}/images/dynamic-pos-bias-log.png +0 -0
  18. {x_transformers-2.7.1 → x_transformers-2.7.2}/images/dynamic-pos-bias-sinusoidal.png +0 -0
  19. {x_transformers-2.7.1 → x_transformers-2.7.2}/images/dynamic-pos-bias.png +0 -0
  20. {x_transformers-2.7.1 → x_transformers-2.7.2}/images/enhanced-recurrence.png +0 -0
  21. {x_transformers-2.7.1 → x_transformers-2.7.2}/images/fcm.png +0 -0
  22. {x_transformers-2.7.1 → x_transformers-2.7.2}/images/ffglu.png +0 -0
  23. {x_transformers-2.7.1 → x_transformers-2.7.2}/images/flash-attention.png +0 -0
  24. {x_transformers-2.7.1 → x_transformers-2.7.2}/images/gate_values.png +0 -0
  25. {x_transformers-2.7.1 → x_transformers-2.7.2}/images/gating.png +0 -0
  26. {x_transformers-2.7.1 → x_transformers-2.7.2}/images/length-extrapolation-scale.png +0 -0
  27. {x_transformers-2.7.1 → x_transformers-2.7.2}/images/macaron-1.png +0 -0
  28. {x_transformers-2.7.1 → x_transformers-2.7.2}/images/macaron-2.png +0 -0
  29. {x_transformers-2.7.1 → x_transformers-2.7.2}/images/memory-transformer.png +0 -0
  30. {x_transformers-2.7.1 → x_transformers-2.7.2}/images/normformer.png +0 -0
  31. {x_transformers-2.7.1 → x_transformers-2.7.2}/images/pia.png +0 -0
  32. {x_transformers-2.7.1 → x_transformers-2.7.2}/images/qknorm-analysis.png +0 -0
  33. {x_transformers-2.7.1 → x_transformers-2.7.2}/images/resi_dual.png +0 -0
  34. {x_transformers-2.7.1 → x_transformers-2.7.2}/images/residual_attn.png +0 -0
  35. {x_transformers-2.7.1 → x_transformers-2.7.2}/images/rezero.png +0 -0
  36. {x_transformers-2.7.1 → x_transformers-2.7.2}/images/rotary.png +0 -0
  37. {x_transformers-2.7.1 → x_transformers-2.7.2}/images/sandwich-2.png +0 -0
  38. {x_transformers-2.7.1 → x_transformers-2.7.2}/images/sandwich.png +0 -0
  39. {x_transformers-2.7.1 → x_transformers-2.7.2}/images/sandwich_norm.png +0 -0
  40. {x_transformers-2.7.1 → x_transformers-2.7.2}/images/scalenorm.png +0 -0
  41. {x_transformers-2.7.1 → x_transformers-2.7.2}/images/talking-heads.png +0 -0
  42. {x_transformers-2.7.1 → x_transformers-2.7.2}/images/topk-attention.png +0 -0
  43. {x_transformers-2.7.1 → x_transformers-2.7.2}/images/xval.png +0 -0
  44. {x_transformers-2.7.1 → x_transformers-2.7.2}/tests/test_x_transformers.py +0 -0
  45. {x_transformers-2.7.1 → x_transformers-2.7.2}/train_belief_state.py +0 -0
  46. {x_transformers-2.7.1 → x_transformers-2.7.2}/train_copy.py +0 -0
  47. {x_transformers-2.7.1 → x_transformers-2.7.2}/train_entropy_tokenizer.py +0 -0
  48. {x_transformers-2.7.1 → x_transformers-2.7.2}/train_enwik8.py +0 -0
  49. {x_transformers-2.7.1 → x_transformers-2.7.2}/train_length_extrapolate.py +0 -0
  50. {x_transformers-2.7.1 → x_transformers-2.7.2}/train_parity.py +0 -0
  51. {x_transformers-2.7.1 → x_transformers-2.7.2}/x_transformers/__init__.py +0 -0
  52. {x_transformers-2.7.1 → x_transformers-2.7.2}/x_transformers/attend.py +0 -0
  53. {x_transformers-2.7.1 → x_transformers-2.7.2}/x_transformers/autoregressive_wrapper.py +0 -0
  54. {x_transformers-2.7.1 → x_transformers-2.7.2}/x_transformers/belief_state_wrapper.py +0 -0
  55. {x_transformers-2.7.1 → x_transformers-2.7.2}/x_transformers/continuous.py +0 -0
  56. {x_transformers-2.7.1 → x_transformers-2.7.2}/x_transformers/dpo.py +0 -0
  57. {x_transformers-2.7.1 → x_transformers-2.7.2}/x_transformers/entropy_based_tokenizer.py +0 -0
  58. {x_transformers-2.7.1 → x_transformers-2.7.2}/x_transformers/multi_input.py +0 -0
  59. {x_transformers-2.7.1 → x_transformers-2.7.2}/x_transformers/neo_mlp.py +0 -0
  60. {x_transformers-2.7.1 → x_transformers-2.7.2}/x_transformers/nonautoregressive_wrapper.py +0 -0
  61. {x_transformers-2.7.1 → x_transformers-2.7.2}/x_transformers/up_wrapper.py +0 -0
  62. {x_transformers-2.7.1 → x_transformers-2.7.2}/x_transformers/xl_autoregressive_wrapper.py +0 -0
  63. {x_transformers-2.7.1 → x_transformers-2.7.2}/x_transformers/xval.py +0 -0
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.4
2
2
  Name: x-transformers
3
- Version: 2.7.1
3
+ Version: 2.7.2
4
4
  Summary: X-Transformers
5
5
  Project-URL: Homepage, https://pypi.org/project/x-transformers/
6
6
  Project-URL: Repository, https://github.com/lucidrains/x-transformers
@@ -1,6 +1,6 @@
1
1
  [project]
2
2
  name = "x-transformers"
3
- version = "2.7.1"
3
+ version = "2.7.2"
4
4
  description = "X-Transformers"
5
5
  authors = [
6
6
  { name = "Phil Wang", email = "lucidrains@gmail.com" }
@@ -48,6 +48,7 @@ class LayerIntermediates:
48
48
  attn_z_loss: Tensor | None = None
49
49
  mems: Tensor | None = None
50
50
  last_layer_hiddens: Tensor | None = None
51
+ initial_embeds: Tensor | None = None
51
52
  attn_pooled_tokens: Tensor | None = None
52
53
  memory_tokens: Tensor | None = None
53
54
  logit_entropies: Tensor | None = None
@@ -3378,6 +3379,10 @@ class TransformerWrapper(Module):
3378
3379
 
3379
3380
  intermediates.last_layer_hiddens = x
3380
3381
 
3382
+ # store initial embed
3383
+
3384
+ intermediates.initial_embed = init_embed
3385
+
3381
3386
  # global average pool
3382
3387
 
3383
3388
  if self.average_pool_embed:
File without changes
File without changes