rxnn 0.1.1__py3-none-any.whl → 0.1.2__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.3
2
2
  Name: rxnn
3
- Version: 0.1.1
3
+ Version: 0.1.2
4
4
  Summary: RxNN: Reactive Neural Networks Platform
5
5
  License: Apache-2.0
6
6
  Keywords: deep-learning,ai,machine-learning
@@ -0,0 +1,29 @@
1
+ src/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
2
+ src/experimental/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
3
+ src/experimental/attention.py,sha256=BxNFOqQcz3v_wzn5n07UuZQIRLkVxNJEqZedvv8K_WQ,5633
4
+ src/memory/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
5
+ src/memory/norm.py,sha256=Ofl8Q5NYEF9GQeO0bhM43tkTW91J0y6TSvTAOYMgloM,6278
6
+ src/memory/stm.py,sha256=EsD8slSP4_9dLuq6aFPDmuFe8PWilxh90so5Z3nm-ig,2057
7
+ src/rxt/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
8
+ src/rxt/models.py,sha256=jmxdQ0FHjqYKVrwoEURez1WV-0quv-5q2eBFGhXyJM0,6932
9
+ src/training/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
10
+ src/training/base.py,sha256=DHankUIta0wWuczN1O0pK7MCis9dYUvc5dAFkap9t5Q,11030
11
+ src/training/bml.py,sha256=Al-qcQYQ-GsdXfa1DB1YF2vGJKgfIfMBy0GlZabUz1Y,14579
12
+ src/training/callbacks.py,sha256=VAjwaAoSgYgU5WV9FybINznLhBRX9WMWNdgS41FtNfY,21093
13
+ src/training/dataset.py,sha256=vQ5mDF3bA0HXya474n4D4iL8Mn3AEpJukgzFNVkxjGU,5106
14
+ src/training/scheduler.py,sha256=ow6oALzWjWQmHSpcJEjv6tg4g4CDMvr73TypxfcefMc,712
15
+ src/training/tokenizer.py,sha256=4Y41f07uo2KPA_7bp3FCcwGKbXoS2hsckOoXUsXfQxY,8052
16
+ src/transformers/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
17
+ src/transformers/attention.py,sha256=fqNziF9gZOLM-RnrkaCOkel8w-EkLEAOiv7WO5G6IMw,14021
18
+ src/transformers/ff.py,sha256=jJnuBDsnnX5uYC_WZH8cXAYrMnz0P-iX7MwcPivjRtI,2533
19
+ src/transformers/layers.py,sha256=PBlnQzS_cyaAjEAQEgdsttSXhtwWsx4PKxRAtPMhTbY,5361
20
+ src/transformers/mask.py,sha256=J0cfLVLt3SzS2ra3KcY4khrkhI975Dw4CjpUi3Sn25s,419
21
+ src/transformers/models.py,sha256=aVY3YTKGKvwzRraduhIgO8phPDRkJytdTglDQm_U2_c,6716
22
+ src/transformers/moe.py,sha256=JQ5QSX4FS7S-fqB7-s1ZmJbPpOeD_Injn8o4vo7wGQE,4936
23
+ src/transformers/positional.py,sha256=2l38RS0Dini3f6Z3LUHr3XwWzg1UK7fO2C6wazWDAYU,4292
24
+ src/transformers/sampler.py,sha256=wSz_1wNloqtuiix5w2Mcsj5NhaO9QlY0j__TVG7wJnM,3938
25
+ src/utils.py,sha256=d5U8i5ukovgDyqiycc2AoxObTz_eF_bgo2MKvdtJ98s,467
26
+ rxnn-0.1.2.dist-info/LICENSE,sha256=C8coDFIUYuOcke4JLPwTqahQUCyXyGq6WOaigOkx8tY,11275
27
+ rxnn-0.1.2.dist-info/METADATA,sha256=N856s1leESK8WagTniy8g-o11VFlkT-PPn9RtcLo42o,14486
28
+ rxnn-0.1.2.dist-info/WHEEL,sha256=fGIA9gx4Qxk2KDKeNJCbOEwSrmLtjWCwzBz351GyrPQ,88
29
+ rxnn-0.1.2.dist-info/RECORD,,
src/__init__.py ADDED
File without changes
File without changes
@@ -1,6 +1,6 @@
1
1
  import torch
2
2
  from torch import nn
3
- from src.rxnn.transformers.attention import MultiHeadAttention
3
+ from src.transformers.attention import MultiHeadAttention
4
4
 
5
5
  class FlexAttention(MultiHeadAttention):
6
6
  def __init__(
src/memory/__init__.py ADDED
File without changes
src/rxt/__init__.py ADDED
File without changes
@@ -2,13 +2,13 @@ import torch
2
2
  from torch import nn
3
3
  from typing import TypedDict, Union
4
4
  from huggingface_hub import PyTorchModelHubMixin
5
- from src.rxnn.transformers.positional import RotaryPositionalEmbedding
6
- from src.rxnn.transformers.attention import init_attention
7
- from src.rxnn.transformers.layers import ReactiveTransformerLayer
8
- from src.rxnn.transformers.models import ReactiveTransformerBase, ReactiveTransformerEncoder, ReactiveTransformerDecoder
9
- from src.rxnn.transformers.ff import get_activation_layer
10
- from src.rxnn.memory.stm import ShortTermMemory
11
- from src.rxnn.utils import get_model_size
5
+ from src.transformers.positional import RotaryPositionalEmbedding
6
+ from src.transformers.attention import init_attention
7
+ from src.transformers.layers import ReactiveTransformerLayer
8
+ from src.transformers.models import ReactiveTransformerBase, ReactiveTransformerEncoder, ReactiveTransformerDecoder
9
+ from src.transformers.ff import get_activation_layer
10
+ from src.memory.stm import ShortTermMemory
11
+ from src.utils import get_model_size
12
12
 
13
13
 
14
14
  class RxTAlphaComponentConfig(TypedDict):
File without changes
@@ -5,8 +5,8 @@ import math
5
5
  from huggingface_hub import PyTorchModelHubMixin
6
6
  from typing import Union
7
7
  import torch.distributed as dist
8
- from src.rxnn.transformers.models import ReactiveTransformerEncoder, ReactiveTransformerDecoder
9
- from src.rxnn.training.base import BaseTrainer
8
+ from src.transformers.models import ReactiveTransformerEncoder, ReactiveTransformerDecoder
9
+ from src.training.base import BaseTrainer
10
10
 
11
11
  class MLMHead(nn.Module, PyTorchModelHubMixin, license="apache-2.0"):
12
12
  def __init__(self, embed_dim: int, vocab_size: int, *args, **kwargs):
@@ -3,7 +3,7 @@ import numpy as np
3
3
  import torch
4
4
  import torch.nn as nn
5
5
  from typing import Union
6
- from src.rxnn.utils import human_format
6
+ from src.utils import human_format
7
7
  from torch.nn.parallel import DistributedDataParallel
8
8
  from huggingface_hub import PyTorchModelHubMixin
9
9
 
File without changes
@@ -2,7 +2,7 @@ import torch
2
2
  import torch.nn as nn
3
3
  from positional import AbsolutePositionalEmbedding
4
4
  from mask import create_causal_mask
5
- from src.rxnn.memory.stm import ShortTermMemory
5
+ from src.memory.stm import ShortTermMemory
6
6
 
7
7
 
8
8
  class ReactiveTransformerBase(nn.Module):
@@ -1,23 +0,0 @@
1
- rxnn/experimental/attention.py,sha256=z7RkLvgY815JHEfCLq1gtW9yKdumjEJ3yGCrA7jCyRU,5638
2
- rxnn/memory/norm.py,sha256=Ofl8Q5NYEF9GQeO0bhM43tkTW91J0y6TSvTAOYMgloM,6278
3
- rxnn/memory/stm.py,sha256=EsD8slSP4_9dLuq6aFPDmuFe8PWilxh90so5Z3nm-ig,2057
4
- rxnn/rxt/models.py,sha256=M63lz7qN3f3cvxgc4bV_WLgUQCnVRnDU8NH54wE_eks,6967
5
- rxnn/training/base.py,sha256=DHankUIta0wWuczN1O0pK7MCis9dYUvc5dAFkap9t5Q,11030
6
- rxnn/training/bml.py,sha256=zX7bvlKyLIPZKYAvt7JVLNZBMz8eSTd3kuucr92S-dc,14589
7
- rxnn/training/callbacks.py,sha256=8RjFuMCtHwmlXJMp6X68wWTJdZ8xyFykckdmb7FjHNU,21098
8
- rxnn/training/dataset.py,sha256=vQ5mDF3bA0HXya474n4D4iL8Mn3AEpJukgzFNVkxjGU,5106
9
- rxnn/training/scheduler.py,sha256=ow6oALzWjWQmHSpcJEjv6tg4g4CDMvr73TypxfcefMc,712
10
- rxnn/training/tokenizer.py,sha256=4Y41f07uo2KPA_7bp3FCcwGKbXoS2hsckOoXUsXfQxY,8052
11
- rxnn/transformers/attention.py,sha256=fqNziF9gZOLM-RnrkaCOkel8w-EkLEAOiv7WO5G6IMw,14021
12
- rxnn/transformers/ff.py,sha256=jJnuBDsnnX5uYC_WZH8cXAYrMnz0P-iX7MwcPivjRtI,2533
13
- rxnn/transformers/layers.py,sha256=PBlnQzS_cyaAjEAQEgdsttSXhtwWsx4PKxRAtPMhTbY,5361
14
- rxnn/transformers/mask.py,sha256=J0cfLVLt3SzS2ra3KcY4khrkhI975Dw4CjpUi3Sn25s,419
15
- rxnn/transformers/models.py,sha256=6QVuUq7VYNniIjGor1YKUy2xK0LWMOFuJzymGVY6Tcc,6721
16
- rxnn/transformers/moe.py,sha256=JQ5QSX4FS7S-fqB7-s1ZmJbPpOeD_Injn8o4vo7wGQE,4936
17
- rxnn/transformers/positional.py,sha256=2l38RS0Dini3f6Z3LUHr3XwWzg1UK7fO2C6wazWDAYU,4292
18
- rxnn/transformers/sampler.py,sha256=wSz_1wNloqtuiix5w2Mcsj5NhaO9QlY0j__TVG7wJnM,3938
19
- rxnn/utils.py,sha256=d5U8i5ukovgDyqiycc2AoxObTz_eF_bgo2MKvdtJ98s,467
20
- rxnn-0.1.1.dist-info/LICENSE,sha256=C8coDFIUYuOcke4JLPwTqahQUCyXyGq6WOaigOkx8tY,11275
21
- rxnn-0.1.1.dist-info/METADATA,sha256=WC7ZkMApI01e4hTndWCk40J00mHx12f19oZmas95jYE,14486
22
- rxnn-0.1.1.dist-info/WHEEL,sha256=fGIA9gx4Qxk2KDKeNJCbOEwSrmLtjWCwzBz351GyrPQ,88
23
- rxnn-0.1.1.dist-info/RECORD,,
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
{rxnn → src}/utils.py RENAMED
File without changes