titans-pytorch 0.0.37__tar.gz → 0.0.38__tar.gz

Sign up to get free protection for your applications and to get access to all the features.

Potentially problematic release.


This version of titans-pytorch might be problematic. Click here for more details.

Files changed (20) hide show
  1. {titans_pytorch-0.0.37 → titans_pytorch-0.0.38}/PKG-INFO +1 -1
  2. {titans_pytorch-0.0.37 → titans_pytorch-0.0.38}/pyproject.toml +1 -1
  3. {titans_pytorch-0.0.37 → titans_pytorch-0.0.38}/titans_pytorch/mac_transformer.py +3 -2
  4. {titans_pytorch-0.0.37 → titans_pytorch-0.0.38}/titans_pytorch/titans.py +8 -5
  5. {titans_pytorch-0.0.37 → titans_pytorch-0.0.38}/.github/workflows/python-publish.yml +0 -0
  6. {titans_pytorch-0.0.37 → titans_pytorch-0.0.38}/.github/workflows/test.yaml +0 -0
  7. {titans_pytorch-0.0.37 → titans_pytorch-0.0.38}/.gitignore +0 -0
  8. {titans_pytorch-0.0.37 → titans_pytorch-0.0.38}/LICENSE +0 -0
  9. {titans_pytorch-0.0.37 → titans_pytorch-0.0.38}/README.md +0 -0
  10. {titans_pytorch-0.0.37 → titans_pytorch-0.0.38}/data/README.md +0 -0
  11. {titans_pytorch-0.0.37 → titans_pytorch-0.0.38}/data/enwik8.gz +0 -0
  12. {titans_pytorch-0.0.37 → titans_pytorch-0.0.38}/fig1.png +0 -0
  13. {titans_pytorch-0.0.37 → titans_pytorch-0.0.38}/fig2.png +0 -0
  14. {titans_pytorch-0.0.37 → titans_pytorch-0.0.38}/requirements.txt +0 -0
  15. {titans_pytorch-0.0.37 → titans_pytorch-0.0.38}/tests/test_titans.py +0 -0
  16. {titans_pytorch-0.0.37 → titans_pytorch-0.0.38}/titans_pytorch/__init__.py +0 -0
  17. {titans_pytorch-0.0.37 → titans_pytorch-0.0.38}/titans_pytorch/associative_scan.py +0 -0
  18. {titans_pytorch-0.0.37 → titans_pytorch-0.0.38}/titans_pytorch/titans_attn_memory.py +0 -0
  19. {titans_pytorch-0.0.37 → titans_pytorch-0.0.38}/train.py +0 -0
  20. {titans_pytorch-0.0.37 → titans_pytorch-0.0.38}/train_mac.py +0 -0
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.4
2
2
  Name: titans-pytorch
3
- Version: 0.0.37
3
+ Version: 0.0.38
4
4
  Summary: Titans
5
5
  Project-URL: Homepage, https://pypi.org/project/titans-pytorch/
6
6
  Project-URL: Repository, https://github.com/lucidrains/titans-pytorch
@@ -1,6 +1,6 @@
1
1
  [project]
2
2
  name = "titans-pytorch"
3
- version = "0.0.37"
3
+ version = "0.0.38"
4
4
  description = "Titans"
5
5
  authors = [
6
6
  { name = "Phil Wang", email = "lucidrains@gmail.com" }
@@ -288,7 +288,8 @@ class MemoryAsContextTransformer(Module):
288
288
  for (attn, ff), maybe_neural_mem in zip(self.layers, self.neural_mem_layers):
289
289
 
290
290
  if exists(maybe_neural_mem):
291
- mems = maybe_neural_mem(mems)
291
+ x = maybe_neural_mem(x)
292
+
292
293
 
293
294
  x = attn(x)
294
295
 
@@ -300,7 +301,7 @@ class MemoryAsContextTransformer(Module):
300
301
 
301
302
  x, inverse_segment = pad_and_segment_with_inverse(x, total_segment_len)
302
303
 
303
- x, mem = unpack(x, mem_ps, 'b * d')
304
+ x, _ = unpack(x, mem_ps, 'b * d')
304
305
 
305
306
  x = inverse_segment(x)
306
307
 
@@ -27,9 +27,7 @@ n - sequence
27
27
  d - feature dimension
28
28
  c - intra-chunk
29
29
  """
30
-
31
- # constants
32
-
30
+ 7
33
31
  LinearNoBias = partial(Linear, bias = False)
34
32
 
35
33
  # functions
@@ -390,7 +388,10 @@ class NeuralMemory(Module):
390
388
 
391
389
  padding = next_seq_len - curtailed_seq_len
392
390
 
393
- seq = pad_at_dim(seq, (0, padding), dim = 1)
391
+ needs_pad = padding > 0
392
+
393
+ if needs_pad:
394
+ seq = pad_at_dim(seq, (0, padding), dim = 1)
394
395
 
395
396
  # the parameters of the memory model stores the memories of the key / values
396
397
  # when the MLP has only 1 weight matrix, it is equivalent to `kv` fast weight memories from linear attention literature (recall fetching of memories is q @ (kv)) / schmidhuber's paper
@@ -442,7 +443,9 @@ class NeuralMemory(Module):
442
443
  empty_memory_embeds = self.init_empty_memory_embed(values.shape[0], chunk_size - 1)
443
444
  values = torch.cat((empty_memory_embeds, values), dim = -2)
444
445
 
445
- values = values[:, :-padding]
446
+ if needs_pad:
447
+ values = values[:, :-padding]
448
+
446
449
  return values
447
450
 
448
451
  def forward(
File without changes