rxnn 0.1.65__tar.gz → 0.1.67__tar.gz

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (31) hide show
  1. {rxnn-0.1.65 → rxnn-0.1.67}/PKG-INFO +1 -1
  2. {rxnn-0.1.65 → rxnn-0.1.67}/pyproject.toml +1 -1
  3. {rxnn-0.1.65 → rxnn-0.1.67}/src/rxnn/training/dataset.py +10 -12
  4. {rxnn-0.1.65 → rxnn-0.1.67}/LICENSE +0 -0
  5. {rxnn-0.1.65 → rxnn-0.1.67}/README.md +0 -0
  6. {rxnn-0.1.65 → rxnn-0.1.67}/src/rxnn/__init__.py +0 -0
  7. {rxnn-0.1.65 → rxnn-0.1.67}/src/rxnn/experimental/__init__.py +0 -0
  8. {rxnn-0.1.65 → rxnn-0.1.67}/src/rxnn/experimental/attention.py +0 -0
  9. {rxnn-0.1.65 → rxnn-0.1.67}/src/rxnn/experimental/models.py +0 -0
  10. {rxnn-0.1.65 → rxnn-0.1.67}/src/rxnn/experimental/moe.py +0 -0
  11. {rxnn-0.1.65 → rxnn-0.1.67}/src/rxnn/memory/__init__.py +0 -0
  12. {rxnn-0.1.65 → rxnn-0.1.67}/src/rxnn/memory/norm.py +0 -0
  13. {rxnn-0.1.65 → rxnn-0.1.67}/src/rxnn/memory/stm.py +0 -0
  14. {rxnn-0.1.65 → rxnn-0.1.67}/src/rxnn/rxt/__init__.py +0 -0
  15. {rxnn-0.1.65 → rxnn-0.1.67}/src/rxnn/rxt/models.py +0 -0
  16. {rxnn-0.1.65 → rxnn-0.1.67}/src/rxnn/training/__init__.py +0 -0
  17. {rxnn-0.1.65 → rxnn-0.1.67}/src/rxnn/training/base.py +0 -0
  18. {rxnn-0.1.65 → rxnn-0.1.67}/src/rxnn/training/bml.py +0 -0
  19. {rxnn-0.1.65 → rxnn-0.1.67}/src/rxnn/training/callbacks.py +0 -0
  20. {rxnn-0.1.65 → rxnn-0.1.67}/src/rxnn/training/scheduler.py +0 -0
  21. {rxnn-0.1.65 → rxnn-0.1.67}/src/rxnn/training/tokenizer.py +0 -0
  22. {rxnn-0.1.65 → rxnn-0.1.67}/src/rxnn/transformers/__init__.py +0 -0
  23. {rxnn-0.1.65 → rxnn-0.1.67}/src/rxnn/transformers/attention.py +0 -0
  24. {rxnn-0.1.65 → rxnn-0.1.67}/src/rxnn/transformers/ff.py +0 -0
  25. {rxnn-0.1.65 → rxnn-0.1.67}/src/rxnn/transformers/layers.py +0 -0
  26. {rxnn-0.1.65 → rxnn-0.1.67}/src/rxnn/transformers/mask.py +0 -0
  27. {rxnn-0.1.65 → rxnn-0.1.67}/src/rxnn/transformers/models.py +0 -0
  28. {rxnn-0.1.65 → rxnn-0.1.67}/src/rxnn/transformers/moe.py +0 -0
  29. {rxnn-0.1.65 → rxnn-0.1.67}/src/rxnn/transformers/positional.py +0 -0
  30. {rxnn-0.1.65 → rxnn-0.1.67}/src/rxnn/transformers/sampler.py +0 -0
  31. {rxnn-0.1.65 → rxnn-0.1.67}/src/rxnn/utils.py +0 -0
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.3
2
2
  Name: rxnn
3
- Version: 0.1.65
3
+ Version: 0.1.67
4
4
  Summary: RxNN: Reactive Neural Networks Platform
5
5
  License: Apache-2.0
6
6
  Keywords: deep-learning,ai,machine-learning
@@ -4,7 +4,7 @@ build-backend = "poetry.core.masonry.api"
4
4
 
5
5
  [tool.poetry]
6
6
  name = "rxnn"
7
- version = "0.1.65"
7
+ version = "0.1.67"
8
8
  description = "RxNN: Reactive Neural Networks Platform"
9
9
 
10
10
  license = "Apache-2.0"
@@ -15,7 +15,7 @@ class BaseDataset(Dataset):
15
15
  max_seq_len: int = 1024,
16
16
  hf_field: str = 'text',
17
17
  cache_tokenized: bool = False,
18
- cache_remove_text: bool = False,
18
+ cache_remove_text: bool = True,
19
19
  *args,
20
20
  **kwargs
21
21
  ):
@@ -29,6 +29,9 @@ class BaseDataset(Dataset):
29
29
  self.cache_remove_text = cache_remove_text
30
30
  self.inputs = [] if self.cache_tokenized else None
31
31
 
32
+ def __len__(self):
33
+ return len(self.texts if not self.is_pre_tokenized else self.inputs)
34
+
32
35
  def get_tokenized_text(self, idx: int):
33
36
  if self.is_pre_tokenized:
34
37
  return self.inputs[idx]
@@ -63,8 +66,12 @@ class BaseDataset(Dataset):
63
66
 
64
67
  def get_subset(self, size: float, from_start: bool = False, use_hf_select: bool = False, **kwargs) -> "BaseDataset":
65
68
  split_point = int(len(self.texts) * ((1 - size) if not from_start else size))
66
- subset = self.texts.select(range(split_point, len(self.texts))) if use_hf_select and not isinstance(self.texts, list) else self.texts[:split_point]
67
- self.texts = self.texts.select(range(split_point)) if use_hf_select and not isinstance(self.texts, list) else self.texts[split_point:]
69
+ if use_hf_select and not isinstance(self.texts, list):
70
+ subset = self.texts.select(range(split_point, len(self.texts)) if not from_start else range(split_point))
71
+ self.texts = self.texts.select(range(split_point) if not from_start else range(split_point, len(self.texts)))
72
+ else:
73
+ subset = self.texts[split_point:-1] if not from_start else self.texts[0:split_point]
74
+ self.texts = self.texts[0:split_point] if not from_start else self.texts[split_point:-1]
68
75
  return self.__class__(subset, self.tokenizer, self.max_seq_len, self.hf_field, **kwargs)
69
76
 
70
77
  def pre_tokenize(self, remove_texts: bool = True):
@@ -209,9 +216,6 @@ class JointLMDataset(BaseDataset):
209
216
  'attention_mask': attention_mask,
210
217
  }
211
218
 
212
- def __len__(self):
213
- return len(self.texts)
214
-
215
219
 
216
220
  class MaskedLMDataset(BaseDataset):
217
221
  def __init__(
@@ -249,9 +253,6 @@ class MaskedLMDataset(BaseDataset):
249
253
  'labels': labels
250
254
  }
251
255
 
252
- def __len__(self):
253
- return len(self.texts)
254
-
255
256
 
256
257
  class AutoregressiveLMDataset(BaseDataset):
257
258
  def __init__(
@@ -277,6 +278,3 @@ class AutoregressiveLMDataset(BaseDataset):
277
278
  'attention_mask': attention_mask,
278
279
  'targets': targets
279
280
  }
280
-
281
- def __len__(self):
282
- return len(self.texts)
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes