x-transformers 2.1.36__tar.gz → 2.1.37__tar.gz

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (60) hide show
  1. {x_transformers-2.1.36 → x_transformers-2.1.37}/PKG-INFO +1 -1
  2. {x_transformers-2.1.36 → x_transformers-2.1.37}/pyproject.toml +1 -1
  3. {x_transformers-2.1.36 → x_transformers-2.1.37}/x_transformers/x_transformers.py +5 -2
  4. {x_transformers-2.1.36 → x_transformers-2.1.37}/.github/FUNDING.yml +0 -0
  5. {x_transformers-2.1.36 → x_transformers-2.1.37}/.github/workflows/python-publish.yml +0 -0
  6. {x_transformers-2.1.36 → x_transformers-2.1.37}/.github/workflows/python-test.yaml +0 -0
  7. {x_transformers-2.1.36 → x_transformers-2.1.37}/.gitignore +0 -0
  8. {x_transformers-2.1.36 → x_transformers-2.1.37}/LICENSE +0 -0
  9. {x_transformers-2.1.36 → x_transformers-2.1.37}/README.md +0 -0
  10. {x_transformers-2.1.36 → x_transformers-2.1.37}/data/README.md +0 -0
  11. {x_transformers-2.1.36 → x_transformers-2.1.37}/data/enwik8.gz +0 -0
  12. {x_transformers-2.1.36 → x_transformers-2.1.37}/images/all-attention.png +0 -0
  13. {x_transformers-2.1.36 → x_transformers-2.1.37}/images/attention-on-attention.png +0 -0
  14. {x_transformers-2.1.36 → x_transformers-2.1.37}/images/cosine-sim-attention.png +0 -0
  15. {x_transformers-2.1.36 → x_transformers-2.1.37}/images/deepnorm.png +0 -0
  16. {x_transformers-2.1.36 → x_transformers-2.1.37}/images/dynamic-pos-bias-linear.png +0 -0
  17. {x_transformers-2.1.36 → x_transformers-2.1.37}/images/dynamic-pos-bias-log.png +0 -0
  18. {x_transformers-2.1.36 → x_transformers-2.1.37}/images/dynamic-pos-bias-sinusoidal.png +0 -0
  19. {x_transformers-2.1.36 → x_transformers-2.1.37}/images/dynamic-pos-bias.png +0 -0
  20. {x_transformers-2.1.36 → x_transformers-2.1.37}/images/enhanced-recurrence.png +0 -0
  21. {x_transformers-2.1.36 → x_transformers-2.1.37}/images/fcm.png +0 -0
  22. {x_transformers-2.1.36 → x_transformers-2.1.37}/images/ffglu.png +0 -0
  23. {x_transformers-2.1.36 → x_transformers-2.1.37}/images/flash-attention.png +0 -0
  24. {x_transformers-2.1.36 → x_transformers-2.1.37}/images/gate_values.png +0 -0
  25. {x_transformers-2.1.36 → x_transformers-2.1.37}/images/gating.png +0 -0
  26. {x_transformers-2.1.36 → x_transformers-2.1.37}/images/length-extrapolation-scale.png +0 -0
  27. {x_transformers-2.1.36 → x_transformers-2.1.37}/images/macaron-1.png +0 -0
  28. {x_transformers-2.1.36 → x_transformers-2.1.37}/images/macaron-2.png +0 -0
  29. {x_transformers-2.1.36 → x_transformers-2.1.37}/images/memory-transformer.png +0 -0
  30. {x_transformers-2.1.36 → x_transformers-2.1.37}/images/normformer.png +0 -0
  31. {x_transformers-2.1.36 → x_transformers-2.1.37}/images/pia.png +0 -0
  32. {x_transformers-2.1.36 → x_transformers-2.1.37}/images/qknorm-analysis.png +0 -0
  33. {x_transformers-2.1.36 → x_transformers-2.1.37}/images/resi_dual.png +0 -0
  34. {x_transformers-2.1.36 → x_transformers-2.1.37}/images/residual_attn.png +0 -0
  35. {x_transformers-2.1.36 → x_transformers-2.1.37}/images/rezero.png +0 -0
  36. {x_transformers-2.1.36 → x_transformers-2.1.37}/images/rotary.png +0 -0
  37. {x_transformers-2.1.36 → x_transformers-2.1.37}/images/sandwich-2.png +0 -0
  38. {x_transformers-2.1.36 → x_transformers-2.1.37}/images/sandwich.png +0 -0
  39. {x_transformers-2.1.36 → x_transformers-2.1.37}/images/sandwich_norm.png +0 -0
  40. {x_transformers-2.1.36 → x_transformers-2.1.37}/images/scalenorm.png +0 -0
  41. {x_transformers-2.1.36 → x_transformers-2.1.37}/images/talking-heads.png +0 -0
  42. {x_transformers-2.1.36 → x_transformers-2.1.37}/images/topk-attention.png +0 -0
  43. {x_transformers-2.1.36 → x_transformers-2.1.37}/images/xval.png +0 -0
  44. {x_transformers-2.1.36 → x_transformers-2.1.37}/tests/test_x_transformers.py +0 -0
  45. {x_transformers-2.1.36 → x_transformers-2.1.37}/train_belief_state.py +0 -0
  46. {x_transformers-2.1.36 → x_transformers-2.1.37}/train_copy.py +0 -0
  47. {x_transformers-2.1.36 → x_transformers-2.1.37}/train_enwik8.py +0 -0
  48. {x_transformers-2.1.36 → x_transformers-2.1.37}/train_length_extrapolate.py +0 -0
  49. {x_transformers-2.1.36 → x_transformers-2.1.37}/train_parity.py +0 -0
  50. {x_transformers-2.1.36 → x_transformers-2.1.37}/x_transformers/__init__.py +0 -0
  51. {x_transformers-2.1.36 → x_transformers-2.1.37}/x_transformers/attend.py +0 -0
  52. {x_transformers-2.1.36 → x_transformers-2.1.37}/x_transformers/autoregressive_wrapper.py +0 -0
  53. {x_transformers-2.1.36 → x_transformers-2.1.37}/x_transformers/belief_state_wrapper.py +0 -0
  54. {x_transformers-2.1.36 → x_transformers-2.1.37}/x_transformers/continuous.py +0 -0
  55. {x_transformers-2.1.36 → x_transformers-2.1.37}/x_transformers/dpo.py +0 -0
  56. {x_transformers-2.1.36 → x_transformers-2.1.37}/x_transformers/multi_input.py +0 -0
  57. {x_transformers-2.1.36 → x_transformers-2.1.37}/x_transformers/neo_mlp.py +0 -0
  58. {x_transformers-2.1.36 → x_transformers-2.1.37}/x_transformers/nonautoregressive_wrapper.py +0 -0
  59. {x_transformers-2.1.36 → x_transformers-2.1.37}/x_transformers/xl_autoregressive_wrapper.py +0 -0
  60. {x_transformers-2.1.36 → x_transformers-2.1.37}/x_transformers/xval.py +0 -0
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.4
2
2
  Name: x-transformers
3
- Version: 2.1.36
3
+ Version: 2.1.37
4
4
  Summary: X-Transformers
5
5
  Project-URL: Homepage, https://pypi.org/project/x-transformers/
6
6
  Project-URL: Repository, https://github.com/lucidrains/x-transformers
@@ -1,6 +1,6 @@
1
1
  [project]
2
2
  name = "x-transformers"
3
- version = "2.1.36"
3
+ version = "2.1.37"
4
4
  description = "X-Transformers"
5
5
  authors = [
6
6
  { name = "Phil Wang", email = "lucidrains@gmail.com" }
@@ -2909,6 +2909,7 @@ class TransformerWrapper(Module):
2909
2909
  return_embeddings = False,
2910
2910
  return_logits_and_embeddings = False,
2911
2911
  return_intermediates = False,
2912
+ return_embeddings_and_intermediates = False,
2912
2913
  return_logit_entropies = False,
2913
2914
  mask = None,
2914
2915
  return_mems = False,
@@ -2940,8 +2941,8 @@ class TransformerWrapper(Module):
2940
2941
 
2941
2942
  b, n, device, num_mems, has_memory_tokens, emb_frac_gradient, orig_mask = x.shape[0], x.shape[1], x.device, self.num_memory_tokens, self.num_memory_tokens > 0, self.emb_frac_gradient, mask
2942
2943
 
2943
- return_hiddens = return_mems | return_attn | return_intermediates | return_attn_z_loss
2944
- return_embeddings = return_embeddings | (not exists(self.to_logits))
2944
+ return_hiddens = return_mems | return_attn | return_intermediates | return_attn_z_loss | return_embeddings_and_intermediates
2945
+ return_embeddings = return_embeddings | (not exists(self.to_logits)) | return_embeddings_and_intermediates
2945
2946
 
2946
2947
  # absolute positional embedding
2947
2948
 
@@ -3131,6 +3132,8 @@ class TransformerWrapper(Module):
3131
3132
 
3132
3133
  if return_logits_and_embeddings:
3133
3134
  out = (logits, x)
3135
+ elif return_embeddings_and_intermediates:
3136
+ out = (x, intermediates)
3134
3137
  elif return_embeddings:
3135
3138
  out = x
3136
3139
  else:
File without changes