rxnn 0.1.3__py3-none-any.whl → 0.1.5__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- rxnn/experimental/attention.py +1 -1
- rxnn/rxt/models.py +7 -7
- rxnn/training/bml.py +2 -2
- rxnn/training/callbacks.py +1 -1
- rxnn/transformers/models.py +1 -1
- {rxnn-0.1.3.dist-info → rxnn-0.1.5.dist-info}/METADATA +1 -1
- {rxnn-0.1.3.dist-info → rxnn-0.1.5.dist-info}/RECORD +9 -9
- {rxnn-0.1.3.dist-info → rxnn-0.1.5.dist-info}/LICENSE +0 -0
- {rxnn-0.1.3.dist-info → rxnn-0.1.5.dist-info}/WHEEL +0 -0
rxnn/experimental/attention.py
CHANGED
rxnn/rxt/models.py
CHANGED
@@ -2,13 +2,13 @@ import torch
|
|
2
2
|
from torch import nn
|
3
3
|
from typing import TypedDict, Union
|
4
4
|
from huggingface_hub import PyTorchModelHubMixin
|
5
|
-
from
|
6
|
-
from
|
7
|
-
from
|
8
|
-
from
|
9
|
-
from
|
10
|
-
from
|
11
|
-
from
|
5
|
+
from rxnn.transformers.positional import RotaryPositionalEmbedding
|
6
|
+
from rxnn.transformers.attention import init_attention
|
7
|
+
from rxnn.transformers.layers import ReactiveTransformerLayer
|
8
|
+
from rxnn.transformers.models import ReactiveTransformerBase, ReactiveTransformerEncoder, ReactiveTransformerDecoder
|
9
|
+
from rxnn.transformers.ff import get_activation_layer
|
10
|
+
from rxnn.memory.stm import ShortTermMemory
|
11
|
+
from rxnn.utils import get_model_size
|
12
12
|
|
13
13
|
|
14
14
|
class RxTAlphaComponentConfig(TypedDict):
|
rxnn/training/bml.py
CHANGED
@@ -5,8 +5,8 @@ import math
|
|
5
5
|
from huggingface_hub import PyTorchModelHubMixin
|
6
6
|
from typing import Union
|
7
7
|
import torch.distributed as dist
|
8
|
-
from
|
9
|
-
from
|
8
|
+
from rxnn.transformers.models import ReactiveTransformerEncoder, ReactiveTransformerDecoder
|
9
|
+
from rxnn.training.base import BaseTrainer
|
10
10
|
|
11
11
|
class MLMHead(nn.Module, PyTorchModelHubMixin, license="apache-2.0"):
|
12
12
|
def __init__(self, embed_dim: int, vocab_size: int, *args, **kwargs):
|
rxnn/training/callbacks.py
CHANGED
@@ -3,7 +3,7 @@ import numpy as np
|
|
3
3
|
import torch
|
4
4
|
import torch.nn as nn
|
5
5
|
from typing import Union
|
6
|
-
from
|
6
|
+
from rxnn.utils import human_format
|
7
7
|
from torch.nn.parallel import DistributedDataParallel
|
8
8
|
from huggingface_hub import PyTorchModelHubMixin
|
9
9
|
|
rxnn/transformers/models.py
CHANGED
@@ -2,7 +2,7 @@ import torch
|
|
2
2
|
import torch.nn as nn
|
3
3
|
from positional import AbsolutePositionalEmbedding
|
4
4
|
from mask import create_causal_mask
|
5
|
-
from
|
5
|
+
from rxnn.memory.stm import ShortTermMemory
|
6
6
|
|
7
7
|
|
8
8
|
class ReactiveTransformerBase(nn.Module):
|
@@ -1,15 +1,15 @@
|
|
1
1
|
rxnn/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
2
2
|
rxnn/experimental/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
3
|
-
rxnn/experimental/attention.py,sha256=
|
3
|
+
rxnn/experimental/attention.py,sha256=M85p_GFU0fbUjfUhXdcwIGW-amrdzwKpU8qSABr7brQ,5634
|
4
4
|
rxnn/memory/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
5
5
|
rxnn/memory/norm.py,sha256=Ofl8Q5NYEF9GQeO0bhM43tkTW91J0y6TSvTAOYMgloM,6278
|
6
6
|
rxnn/memory/stm.py,sha256=EsD8slSP4_9dLuq6aFPDmuFe8PWilxh90so5Z3nm-ig,2057
|
7
7
|
rxnn/rxt/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
8
|
-
rxnn/rxt/models.py,sha256=
|
8
|
+
rxnn/rxt/models.py,sha256=NOXEAT262-UVh6q8LjLmdiqOpyze2BkcWTEHoYKAggk,6939
|
9
9
|
rxnn/training/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
10
10
|
rxnn/training/base.py,sha256=DHankUIta0wWuczN1O0pK7MCis9dYUvc5dAFkap9t5Q,11030
|
11
|
-
rxnn/training/bml.py,sha256=
|
12
|
-
rxnn/training/callbacks.py,sha256=
|
11
|
+
rxnn/training/bml.py,sha256=FlMh4_iH67Hsj6kOioeQB5DSqO4qKAlgYsn5OfVmPyc,14581
|
12
|
+
rxnn/training/callbacks.py,sha256=BATxW0NrJ3KEpwSquctlJRICKSI0dFYI1MLrji5FdbI,21094
|
13
13
|
rxnn/training/dataset.py,sha256=vQ5mDF3bA0HXya474n4D4iL8Mn3AEpJukgzFNVkxjGU,5106
|
14
14
|
rxnn/training/scheduler.py,sha256=ow6oALzWjWQmHSpcJEjv6tg4g4CDMvr73TypxfcefMc,712
|
15
15
|
rxnn/training/tokenizer.py,sha256=4Y41f07uo2KPA_7bp3FCcwGKbXoS2hsckOoXUsXfQxY,8052
|
@@ -18,12 +18,12 @@ rxnn/transformers/attention.py,sha256=fqNziF9gZOLM-RnrkaCOkel8w-EkLEAOiv7WO5G6IM
|
|
18
18
|
rxnn/transformers/ff.py,sha256=jJnuBDsnnX5uYC_WZH8cXAYrMnz0P-iX7MwcPivjRtI,2533
|
19
19
|
rxnn/transformers/layers.py,sha256=PBlnQzS_cyaAjEAQEgdsttSXhtwWsx4PKxRAtPMhTbY,5361
|
20
20
|
rxnn/transformers/mask.py,sha256=J0cfLVLt3SzS2ra3KcY4khrkhI975Dw4CjpUi3Sn25s,419
|
21
|
-
rxnn/transformers/models.py,sha256=
|
21
|
+
rxnn/transformers/models.py,sha256=WP7QoHf3FHbDkc0M5Y69HCLgk2Uwxykt4ScDAAk8EMQ,6717
|
22
22
|
rxnn/transformers/moe.py,sha256=JQ5QSX4FS7S-fqB7-s1ZmJbPpOeD_Injn8o4vo7wGQE,4936
|
23
23
|
rxnn/transformers/positional.py,sha256=2l38RS0Dini3f6Z3LUHr3XwWzg1UK7fO2C6wazWDAYU,4292
|
24
24
|
rxnn/transformers/sampler.py,sha256=wSz_1wNloqtuiix5w2Mcsj5NhaO9QlY0j__TVG7wJnM,3938
|
25
25
|
rxnn/utils.py,sha256=d5U8i5ukovgDyqiycc2AoxObTz_eF_bgo2MKvdtJ98s,467
|
26
|
-
rxnn-0.1.
|
27
|
-
rxnn-0.1.
|
28
|
-
rxnn-0.1.
|
29
|
-
rxnn-0.1.
|
26
|
+
rxnn-0.1.5.dist-info/LICENSE,sha256=C8coDFIUYuOcke4JLPwTqahQUCyXyGq6WOaigOkx8tY,11275
|
27
|
+
rxnn-0.1.5.dist-info/METADATA,sha256=9cW2wFwTrdUqpY7J33ap2y8NPKsx7Tnny0fzidhGTGI,14486
|
28
|
+
rxnn-0.1.5.dist-info/WHEEL,sha256=fGIA9gx4Qxk2KDKeNJCbOEwSrmLtjWCwzBz351GyrPQ,88
|
29
|
+
rxnn-0.1.5.dist-info/RECORD,,
|
File without changes
|
File without changes
|