orbit-torch 0.0.4a1__py3-none-any.whl → 0.1.0b1__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- orbit/__init__.py +3 -1
- orbit/callback.py +4 -3
- orbit/dataset/__init__.py +1 -0
- orbit/dataset/cogn.py +138 -0
- orbit/dataset/data/cogn_en.jsonl +45 -0
- orbit/dataset/data/cogn_zh.jsonl +113 -0
- orbit/engine.py +210 -146
- orbit/kit/__init__.py +2 -0
- orbit/kit/interface.py +154 -0
- orbit/kit/wrapper.py +157 -0
- orbit/model/__init__.py +5 -0
- orbit/model/base.py +125 -0
- orbit/model/block/__init__.py +34 -0
- orbit/model/block/attention.py +265 -0
- orbit/model/block/bio.py +537 -0
- orbit/model/block/codebook.py +122 -0
- orbit/model/block/conv.py +505 -0
- orbit/model/block/embedding.py +252 -0
- orbit/model/block/film.py +176 -0
- orbit/model/block/fusion.py +335 -0
- orbit/model/block/gate.py +334 -0
- orbit/model/block/lora.py +776 -0
- orbit/model/block/mlp.py +68 -0
- orbit/model/block/moe.py +94 -0
- orbit/model/block/tcn.py +99 -0
- orbit/model/config.py +62 -0
- orbit/model/kit/__init__.py +6 -0
- orbit/model/kit/discriminator.py +46 -0
- orbit/model/kit/losses.py +193 -0
- orbit/model/motif/__init__.py +0 -0
- orbit/model/motif/vision/__init__.py +0 -0
- orbit/model/motif/vision/v1.py +645 -0
- orbit/model/registry.py +53 -0
- orbit/optim/__init__.py +2 -2
- orbit/optim/sam.py +10 -3
- orbit/plugin/__init__.py +12 -8
- orbit/plugin/board.py +1 -2
- orbit/plugin/checkpoint.py +137 -62
- orbit/plugin/classification.py +2 -2
- orbit/plugin/display_model.py +1 -2
- orbit/plugin/early_stopping.py +1 -2
- orbit/plugin/ema.py +1 -2
- orbit/plugin/gradient_accumulation.py +1 -2
- orbit/plugin/lora.py +346 -0
- orbit/plugin/memory_estimator.py +1 -2
- orbit/plugin/warmup.py +1 -2
- orbit/utils/__init__.py +24 -1
- orbit/utils/cuda.py +10 -0
- orbit/utils/freeze.py +61 -17
- orbit/utils/image.py +164 -0
- orbit/utils/initialization.py +184 -94
- orbit/utils/layer_io.py +66 -7
- orbit/utils/lora.py +480 -0
- orbit/utils/moe.py +55 -0
- orbit/utils/seed.py +3 -19
- orbit/utils/sft.py +93 -0
- orbit_torch-0.1.0b1.dist-info/METADATA +208 -0
- orbit_torch-0.1.0b1.dist-info/RECORD +65 -0
- orbit_torch-0.0.4a1.dist-info/METADATA +0 -25
- orbit_torch-0.0.4a1.dist-info/RECORD +0 -29
- {orbit_torch-0.0.4a1.dist-info → orbit_torch-0.1.0b1.dist-info}/WHEEL +0 -0
- {orbit_torch-0.0.4a1.dist-info → orbit_torch-0.1.0b1.dist-info}/top_level.txt +0 -0
|
@@ -0,0 +1,208 @@
|
|
|
1
|
+
Metadata-Version: 2.4
|
|
2
|
+
Name: orbit-torch
|
|
3
|
+
Version: 0.1.0b1
|
|
4
|
+
Summary: A PyTorch training engine with plugin system and advanced model components
|
|
5
|
+
Home-page: https://github.com/A03HCY/Orbit
|
|
6
|
+
Author: Aiden Hopkins
|
|
7
|
+
Author-email: acdphc@qq.com
|
|
8
|
+
Classifier: Programming Language :: Python :: 3
|
|
9
|
+
Classifier: License :: OSI Approved :: MIT License
|
|
10
|
+
Classifier: Operating System :: OS Independent
|
|
11
|
+
Requires-Python: >=3.8
|
|
12
|
+
Description-Content-Type: text/markdown
|
|
13
|
+
Requires-Dist: torch>=2.0.0
|
|
14
|
+
Requires-Dist: rich
|
|
15
|
+
Requires-Dist: tensorboard
|
|
16
|
+
Requires-Dist: matplotlib
|
|
17
|
+
Requires-Dist: seaborn
|
|
18
|
+
Requires-Dist: numpy
|
|
19
|
+
Requires-Dist: scikit-learn
|
|
20
|
+
Requires-Dist: einops
|
|
21
|
+
Requires-Dist: tokenizers
|
|
22
|
+
Requires-Dist: transformers
|
|
23
|
+
Requires-Dist: safetensors
|
|
24
|
+
Requires-Dist: accelerate
|
|
25
|
+
Requires-Dist: lpips
|
|
26
|
+
Dynamic: author
|
|
27
|
+
Dynamic: author-email
|
|
28
|
+
Dynamic: classifier
|
|
29
|
+
Dynamic: description
|
|
30
|
+
Dynamic: description-content-type
|
|
31
|
+
Dynamic: home-page
|
|
32
|
+
Dynamic: requires-dist
|
|
33
|
+
Dynamic: requires-python
|
|
34
|
+
Dynamic: summary
|
|
35
|
+
|
|
36
|
+
# Orbit
|
|
37
|
+
|
|
38
|
+
Orbit is a flexible, plugin-based PyTorch training engine designed to simplify the training loop while providing powerful components for modern deep learning models, including LLMs.
|
|
39
|
+
|
|
40
|
+
It features a modular design with a rich set of plugins, advanced model building blocks (like MoE, RoPE, GQA), comprehensive LoRA/DoRA support, and cutting-edge optimizers.
|
|
41
|
+
|
|
42
|
+
## Features
|
|
43
|
+
|
|
44
|
+
### 🚀 Core Engine
|
|
45
|
+
- **Plugin System**: Decoupled training logic using plugins for callbacks, logging, and training strategies.
|
|
46
|
+
- **Simplified Loop**: Clean `train` and `eval` interfaces.
|
|
47
|
+
- **Flexible Updates**:
|
|
48
|
+
- `auto_update()`: Automatically handles forward pass, loss calculation, backward pass, optimizer step, and zero grad.
|
|
49
|
+
- `update(loss)`: Allows manual control over the update step if you need custom forward/loss logic.
|
|
50
|
+
|
|
51
|
+
### 🧩 Model Components (`orbit.model`)
|
|
52
|
+
Orbit provides a collection of high-performance, reusable layers:
|
|
53
|
+
- **Attention**: `MultiHeadAttention` with support for **GQA** (Grouped Query Attention), **RoPE** (Rotary Positional Embeddings), and FlashAttention.
|
|
54
|
+
- **LoRA & DoRA**: Full support for Low-Rank Adaptation and **Weight-Decomposed Low-Rank Adaptation (DoRA)** across `Linear`, `Conv2d`, `Conv1d`, and `Embedding` layers. Also supports **Gated LoRA**.
|
|
55
|
+
- **MoE**: Mixture of Experts block with `TopKGate` routing.
|
|
56
|
+
- **Gates**: A variety of gating mechanisms including `SigmoidGate`, `TanhGate`, `SoftmaxGate`, `GLUGate`, `ContextGate`, and `TopKGate`.
|
|
57
|
+
- **Others**: `FiLM` (Feature-wise Linear Modulation), `MLP` (with Gated support), `RotaryPositionalEmbedding`.
|
|
58
|
+
|
|
59
|
+
### 🛠️ Utilities & Kit (`orbit.utils`)
|
|
60
|
+
|
|
61
|
+
Orbit provides a comprehensive toolkit to speed up development:
|
|
62
|
+
|
|
63
|
+
#### 🔧 LoRA Utilities
|
|
64
|
+
Manual control over LoRA injection and management (alternative to the Plugin approach).
|
|
65
|
+
- **Injection**:
|
|
66
|
+
- `inject_lora(model, r=8, ...)`: Manually inject LoRA/DoRA/Gated LoRA into specific layers.
|
|
67
|
+
- `inject_lora_file(model, path)`: Automatically inject and load LoRA configuration/weights from a file.
|
|
68
|
+
- **Management**:
|
|
69
|
+
- `merge_lora(model)` / `unmerge_lora(model)`: Merge weights for faster inference or unmerge to resume training.
|
|
70
|
+
- `save_lora(model, path)` / `load_lora(model, path)`: Efficiently save/load only LoRA parameters.
|
|
71
|
+
- `freeze_backbone_only(model)`: Helper to freeze the base model while keeping LoRA and specified heads trainable.
|
|
72
|
+
- **Diagnosis**:
|
|
73
|
+
- `LoRADiagnoser`: Check for rank collapse and monitor gradient norms during training.
|
|
74
|
+
|
|
75
|
+
#### ❄️ Model Freezing
|
|
76
|
+
- `freeze_layers(model, targets=['encoder'])`: Freeze layers matching the target names (supports wildcards).
|
|
77
|
+
- `unfreeze_layers(model, targets)`: Unfreeze specific layers.
|
|
78
|
+
- `get_trainable_params(model)`: Get parameters for the optimizer.
|
|
79
|
+
|
|
80
|
+
#### 🎭 Masking
|
|
81
|
+
- `make_causal_mask`: Create causal masks for autoregressive models.
|
|
82
|
+
- `make_padding_mask`, `make_lookahead_mask`, `make_sliding_window_mask`.
|
|
83
|
+
|
|
84
|
+
#### 💾 Layer I/O
|
|
85
|
+
- `save_layer(model, layer_name, path)`: Save weights of a specific sub-module (e.g., just the backbone).
|
|
86
|
+
- `load_layer(model, layer_name, path)`: Load weights into a specific sub-module.
|
|
87
|
+
- `get_model_by_name(model, name)`: Access sub-modules using dot notation strings (e.g., "backbone.layer1").
|
|
88
|
+
|
|
89
|
+
#### 📝 SFT Helpers
|
|
90
|
+
- `build_sft`: Prepares data for Supervised Fine-Tuning (handles chat templates, tokenization, and label masking).
|
|
91
|
+
- `train_sft(engine)`: A specialized training step for SFT that handles the forward pass and loss calculation automatically.
|
|
92
|
+
|
|
93
|
+
#### ⚙️ Optimization (`orbit.optim`)
|
|
94
|
+
- **Muon**: MomentUm Orthogonalized by Newton-schulz optimizer.
|
|
95
|
+
- **SAM**: Sharpness-Aware Minimization wrapper.
|
|
96
|
+
|
|
97
|
+
#### 🌱 Initialization & Seeding
|
|
98
|
+
- `auto_initialize(model)`: Automatically initializes weights based on layer type (Linear, Conv, Embedding, etc.).
|
|
99
|
+
- `seed_everything(seed)`: Sets seeds for Python, NumPy, PyTorch, and CUDA for reproducibility.
|
|
100
|
+
|
|
101
|
+
#### 🖥️ CUDA
|
|
102
|
+
- `cuda_alloc(size)`: Optimizes PyTorch CUDA memory allocation configuration (e.g., `max_split_size_mb`).
|
|
103
|
+
|
|
104
|
+
### 🔌 Plugins (`orbit.plugin`)
|
|
105
|
+
- `EarlyStopping`: Stop training when a metric stops improving.
|
|
106
|
+
- `GradientAccumulation`: Simulate larger batch sizes.
|
|
107
|
+
- `Warmup`: Learning rate warmup.
|
|
108
|
+
- `Mentor`: Training assistant/logger.
|
|
109
|
+
- `MemoryEstimator`: Monitor CUDA memory usage.
|
|
110
|
+
- `LoRA`: Easy injection of LoRA layers via plugin.
|
|
111
|
+
- `Board`: TensorBoard integration.
|
|
112
|
+
|
|
113
|
+
## Installation
|
|
114
|
+
|
|
115
|
+
```bash
|
|
116
|
+
pip install orbit-torch
|
|
117
|
+
```
|
|
118
|
+
|
|
119
|
+
**Requirements**:
|
|
120
|
+
- Python >= 3.8
|
|
121
|
+
- PyTorch >= 2.0.0 (Required for FlashAttention backend)
|
|
122
|
+
|
|
123
|
+
## Quick Start
|
|
124
|
+
|
|
125
|
+
### 1. Basic Training (CIFAR-10)
|
|
126
|
+
|
|
127
|
+
```python
|
|
128
|
+
import torch
|
|
129
|
+
import torch.nn as nn
|
|
130
|
+
from orbit.engine import Engine
|
|
131
|
+
from orbit.plugin import EarlyStopping, GradientAccumulation, Mentor
|
|
132
|
+
from orbit.utils import auto_initialize
|
|
133
|
+
|
|
134
|
+
# Define your model
|
|
135
|
+
model = MyConvNet()
|
|
136
|
+
auto_initialize(model)
|
|
137
|
+
|
|
138
|
+
# Setup Engine
|
|
139
|
+
trainer = Engine(
|
|
140
|
+
model=model,
|
|
141
|
+
criterion=nn.CrossEntropyLoss(),
|
|
142
|
+
optimizer=torch.optim.Adam(model.parameters(), lr=1e-3),
|
|
143
|
+
plugins=[
|
|
144
|
+
Mentor(),
|
|
145
|
+
EarlyStopping(monitor='val_acc', patience=3),
|
|
146
|
+
GradientAccumulation(steps=2)
|
|
147
|
+
]
|
|
148
|
+
)
|
|
149
|
+
|
|
150
|
+
# Train
|
|
151
|
+
for _ in trainer.train(train_loader, num_epochs=10):
|
|
152
|
+
trainer.auto_update() # Handles forward, backward, step, zero_grad
|
|
153
|
+
|
|
154
|
+
# Handle Epoch End (e.g., Validation)
|
|
155
|
+
if not trainer.is_epoch_end: continue
|
|
156
|
+
|
|
157
|
+
for _ in trainer.eval(test_loader):
|
|
158
|
+
trainer.auto_update()
|
|
159
|
+
```
|
|
160
|
+
|
|
161
|
+
### 2. LLM SFT with LoRA/DoRA
|
|
162
|
+
|
|
163
|
+
Orbit makes it easy to fine-tune LLMs using LoRA or DoRA.
|
|
164
|
+
|
|
165
|
+
```python
|
|
166
|
+
from transformers import AutoModelForCausalLM, AutoTokenizer
|
|
167
|
+
from orbit.engine import Engine
|
|
168
|
+
from orbit.plugin import LoRA, GradientAccumulation
|
|
169
|
+
from orbit.utils import train_sft, seed_everything
|
|
170
|
+
|
|
171
|
+
seed_everything(42)
|
|
172
|
+
|
|
173
|
+
# Load Model
|
|
174
|
+
model = AutoModelForCausalLM.from_pretrained("Qwen/Qwen2.5-3B", device_map="auto")
|
|
175
|
+
tokenizer = AutoTokenizer.from_pretrained("Qwen/Qwen2.5-3B")
|
|
176
|
+
|
|
177
|
+
# Setup Engine with LoRA Plugin
|
|
178
|
+
trainer = Engine(
|
|
179
|
+
model=model,
|
|
180
|
+
optimizer=torch.optim.AdamW(model.parameters(), lr=1e-4),
|
|
181
|
+
plugins=[
|
|
182
|
+
# Inject DoRA into MLP layers
|
|
183
|
+
LoRA(target_names=['mlp'], dora=True, r=16, alpha=32),
|
|
184
|
+
GradientAccumulation(steps=8)
|
|
185
|
+
]
|
|
186
|
+
)
|
|
187
|
+
|
|
188
|
+
# Train Loop
|
|
189
|
+
# Assuming `dataloader` yields SFT batches (input_ids, attention_mask, labels)
|
|
190
|
+
for _ in trainer.train(dataloader, num_epochs=3):
|
|
191
|
+
# train_sft handles the forward pass and loss calculation for CausalLM
|
|
192
|
+
train_sft(trainer)
|
|
193
|
+
```
|
|
194
|
+
|
|
195
|
+
### 3. Chat Interface
|
|
196
|
+
|
|
197
|
+
Interact with your trained model in the terminal:
|
|
198
|
+
|
|
199
|
+
```python
|
|
200
|
+
from orbit.kit import ChatInterface
|
|
201
|
+
|
|
202
|
+
chat = ChatInterface(model_id="path/to/model", device="cuda")
|
|
203
|
+
chat.interact()
|
|
204
|
+
```
|
|
205
|
+
|
|
206
|
+
## License
|
|
207
|
+
|
|
208
|
+
MIT License
|
|
@@ -0,0 +1,65 @@
|
|
|
1
|
+
orbit/__init__.py,sha256=Pnt9BItQOlPZR4StcWQ3C-XmYHvqmLSyJiF0honrEew,78
|
|
2
|
+
orbit/callback.py,sha256=qI4iaHwY5AXpzgSHNdiAEgK5XJxv8eEMTg9YxpzVeFk,1669
|
|
3
|
+
orbit/engine.py,sha256=-S0kNNu_rgX17xrurKDgigqjzgiYN3AKoD4QDXqi80g,35894
|
|
4
|
+
orbit/dataset/__init__.py,sha256=OAhQmm3GYwOfsznpGRNEhPr0Lanw_iIrqJc7FV2gbBc,46
|
|
5
|
+
orbit/dataset/cogn.py,sha256=kN_CPkV_VnsS7pVN0tdXsah0dYr780K7gfo3ROIGWyY,4749
|
|
6
|
+
orbit/dataset/data/cogn_en.jsonl,sha256=Ba_KQKW3BJs3jWD7B_z0_Lwq34Zujgu0M2xe7RWO0dA,14862
|
|
7
|
+
orbit/dataset/data/cogn_zh.jsonl,sha256=d__Dr_5Z3w33KcX8CikK34VNt9DVAFUnDvnhCJikumk,39055
|
|
8
|
+
orbit/kit/__init__.py,sha256=40DZDZwGKgOeJ0vUsqhCBKHt-H1Tj0a79dYrjfKKOgs,82
|
|
9
|
+
orbit/kit/interface.py,sha256=_eB6PguJk-EjpLjnCuGDTHJ2woEhHxqW36sasCYC9Ws,5658
|
|
10
|
+
orbit/kit/wrapper.py,sha256=PSBDEx4fsnuznZ3-9EYcUTr4hgQsRPDT9EiI-rxvcaY,5941
|
|
11
|
+
orbit/model/__init__.py,sha256=DY9le8idTuOgCu-K2Os4j83Fa6X55G3VT1DpGP6GvTs,151
|
|
12
|
+
orbit/model/base.py,sha256=NekTUd1P_0e8yHsRUoIJWJbNYqf0-XdDTgFk9WeNBOQ,4264
|
|
13
|
+
orbit/model/config.py,sha256=2IMe6bxES_qf1hXNhroMuAosN9BCZfJJTMJryrz0Q2c,1888
|
|
14
|
+
orbit/model/registry.py,sha256=r1nMCYTpBT24Fu-mOL-S37zOQA-ngpd4uHvonnkN6Lo,1606
|
|
15
|
+
orbit/model/block/__init__.py,sha256=NvNs7F1MfUxOmpHQgrTE0XcJ20OOreTBCGKTYxtggtc,938
|
|
16
|
+
orbit/model/block/attention.py,sha256=7eSE0GM5wW2gsjcTdqUoHEyd8VqCJ1bNbfF-HY7mAUU,10131
|
|
17
|
+
orbit/model/block/bio.py,sha256=SNCG1urEfqDoX-B7-hNpHWJ_vHrU8EBJP9-sjdeh2Nw,21291
|
|
18
|
+
orbit/model/block/codebook.py,sha256=IhxOp0_--euNvrNCU95_h8e38XA6HBuZef8dDC7Ex-g,4195
|
|
19
|
+
orbit/model/block/conv.py,sha256=NswgOt5daz60ROOqmorrftcxI5I0aDQ14rnorYfUcpM,18616
|
|
20
|
+
orbit/model/block/embedding.py,sha256=hMFsx2gYhm_qNoF22Xg2rkOKf5IP16fflUttV8q1OA8,9101
|
|
21
|
+
orbit/model/block/film.py,sha256=oBWmm2mclSjE4yyVf-Qy8n5AeoGkj4xBGERylqNFkMs,6542
|
|
22
|
+
orbit/model/block/fusion.py,sha256=YbyFiIOgbYD0ywrqIPPr0T662yEmc4VcJBT0OSR-yEw,12361
|
|
23
|
+
orbit/model/block/gate.py,sha256=ODE-bKRVaqkTzhAI3z_JCNINfCq1fosQHSrWHqeTmBk,10648
|
|
24
|
+
orbit/model/block/lora.py,sha256=ggmbj9_qQu0FzkqE1A1XRnuIv_bOCVDIPergN-kZPQc,31208
|
|
25
|
+
orbit/model/block/mlp.py,sha256=EcJWbLREPoBVCgQDHW6nJ-2zuxZ0J-biFhZNXd0BFRQ,2226
|
|
26
|
+
orbit/model/block/moe.py,sha256=JW-1esQ986T6_wDQwF7065RjkTSqPx9vEcNppSPG0tE,3151
|
|
27
|
+
orbit/model/block/tcn.py,sha256=ClkYw2cMFcXFm6mMbrUZnraZ5YtokGotU1_sJBGANAU,3836
|
|
28
|
+
orbit/model/kit/__init__.py,sha256=-Q8mGi39sDmUZP9P3nZ1GoyvEOz-rfa0_JEXSzlfplg,156
|
|
29
|
+
orbit/model/kit/discriminator.py,sha256=JA30tHCjmZ1L1oPf-7o5w1nM-_9h1gB9PiPkkNz_JSE,1484
|
|
30
|
+
orbit/model/kit/losses.py,sha256=CSR3aH6SeUIA72sopi2XAQDDHyFelKhZAZcNG3tLKmM,7697
|
|
31
|
+
orbit/model/motif/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
32
|
+
orbit/model/motif/vision/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
33
|
+
orbit/model/motif/vision/v1.py,sha256=a1guncpocl1x57pGpVa60FR_7rHVBV3v_JOVLgLee4o,22981
|
|
34
|
+
orbit/optim/__init__.py,sha256=tO24DdBQ7hfBbk2tKIwh5Y0WXqEYjxowLuwXCuScFMI,45
|
|
35
|
+
orbit/optim/muon.py,sha256=IT0l1b0mAtLpB-SUuPyo9uaYHO2x_wkqg1TLcDJ6Bvk,7342
|
|
36
|
+
orbit/optim/sam.py,sha256=aeqWF7aCfZuhuzlSXFEGP8fUTbEDp4Cm75Q-Xr8I6Ho,3949
|
|
37
|
+
orbit/plugin/__init__.py,sha256=VJbpVtYaGB8roeSrv1PCaGsKmlqDTHpJ91LkeEh1xHg,466
|
|
38
|
+
orbit/plugin/board.py,sha256=ofuxEJf1CC2I7Y5fR_3_Ln04HPQqBKTVp2-8CU7lvI8,2266
|
|
39
|
+
orbit/plugin/checkpoint.py,sha256=R-R-180KWcF264DPvWgjGAoWtX-Vkn0nS0MQjOBfDkU,15264
|
|
40
|
+
orbit/plugin/classification.py,sha256=h2I_a19NzC8RLrO5qD1TbaZ6QBh3Wig8m0-kBiXNdn4,7020
|
|
41
|
+
orbit/plugin/display_model.py,sha256=DuvefqcHCHUCyS8bybBzxaXr6Y8gxlpql-Sa9hePBuw,2906
|
|
42
|
+
orbit/plugin/early_stopping.py,sha256=E8l4vSiZ1oECK9SeXFcrhMuDEC72PGHTwB4p_oZghIc,4060
|
|
43
|
+
orbit/plugin/ema.py,sha256=4HnVOfFDRwdjUeOPkkOVj-A6fA8GPEBetsFPlI5wVuk,3887
|
|
44
|
+
orbit/plugin/gradient_accumulation.py,sha256=Hr2TyPDUx8SOBJ6IpBfP5wPBvKj-JgYBNlF28LIs9-U,1061
|
|
45
|
+
orbit/plugin/lora.py,sha256=LmpBc1Fo54eANsAVf36W93O8tLiKcv9z1Ofv8PgtIJ0,17032
|
|
46
|
+
orbit/plugin/memory_estimator.py,sha256=7YrlAJb2yOh6nTaNK29NAu_aLWHpboIN8mnqYIL4LZo,9738
|
|
47
|
+
orbit/plugin/mentor.py,sha256=sx3gLHnxYJX1tJhqnqHPTW2KuTeNM7h0PqOIquxMaxQ,12474
|
|
48
|
+
orbit/plugin/overfit.py,sha256=lpfxX4igi4uNdY4OR2_fpayw52iEfkCQnQ0HgzodZd4,1194
|
|
49
|
+
orbit/plugin/warmup.py,sha256=XvU47kNSk43OSVPkS1riyhndySOK-Vo5M_SReUjxSHo,4991
|
|
50
|
+
orbit/plugin/data/mentor_i18n.json,sha256=KHnEKRKHeLga_nqEUZVgiTh4jb4j2f_yD0MJT-7sdis,8005
|
|
51
|
+
orbit/utils/__init__.py,sha256=2rxWZ27U4oZH0XxqicYOV8oeWQh2NTv-igakgdRPbMI,1093
|
|
52
|
+
orbit/utils/cuda.py,sha256=5JUxXVDJ_IzQfEI9oguTKSz2uSLT2GTFNXbxSZVWhFM,267
|
|
53
|
+
orbit/utils/freeze.py,sha256=ee6MxqoLaOZ-kf5DZyzUAsS5JUCFQPU_2LxkLZQIxJo,4265
|
|
54
|
+
orbit/utils/image.py,sha256=wklEHDcH_TGvT3aDBPKSJvJWHGgHu4yxssuVfZ8u4No,6775
|
|
55
|
+
orbit/utils/initialization.py,sha256=tIRQMlyNfcVDvrtLwLKthpyCjJ_Cy7AN5ZkggFt9o0s,23407
|
|
56
|
+
orbit/utils/layer_io.py,sha256=_qGmt2GDOxVb0RacTK4ZMALek36oQLYVqCeuwcXzaZE,3920
|
|
57
|
+
orbit/utils/lora.py,sha256=LXqSuh9QGTRn_WJsOnJPmiSy0uzpFnWpLxAmLWq_twc,18458
|
|
58
|
+
orbit/utils/mask.py,sha256=DayMxmKlWMYMYPceYhYlf9lIx4FD9s4Tp1WvzfVpsu0,3072
|
|
59
|
+
orbit/utils/moe.py,sha256=miLesFrUHIe9gWr0STs4d4NgnUbV7nskq5N4ywRtEj0,2132
|
|
60
|
+
orbit/utils/seed.py,sha256=PaApRMEO2codOXDcmelybRJ2ZgAk-Sh4K5SHlZ4KpHE,1789
|
|
61
|
+
orbit/utils/sft.py,sha256=Y6hDGW4tAgtzmEeSqolkYcBKYrOJvuW1vxgJMWvzdRU,3127
|
|
62
|
+
orbit_torch-0.1.0b1.dist-info/METADATA,sha256=IzDxu8AeHbXbrub3QisQybVGdYoYzjx3c1t12MRIzZE,7831
|
|
63
|
+
orbit_torch-0.1.0b1.dist-info/WHEEL,sha256=_zCd3N1l69ArxyTb8rzEoP9TpbYXkqRFSNOD5OuxnTs,91
|
|
64
|
+
orbit_torch-0.1.0b1.dist-info/top_level.txt,sha256=emrF0of931NzTSL4R5yBKpGoewFCB-cAwYNcUF5cqBs,6
|
|
65
|
+
orbit_torch-0.1.0b1.dist-info/RECORD,,
|
|
@@ -1,25 +0,0 @@
|
|
|
1
|
-
Metadata-Version: 2.4
|
|
2
|
-
Name: orbit-torch
|
|
3
|
-
Version: 0.0.4a1
|
|
4
|
-
Summary: A PyTorch training engine with plugin system
|
|
5
|
-
Author: Aiden Hopkins
|
|
6
|
-
Author-email: acdphc@qq.com
|
|
7
|
-
Classifier: Programming Language :: Python :: 3
|
|
8
|
-
Classifier: License :: OSI Approved :: MIT License
|
|
9
|
-
Classifier: Operating System :: OS Independent
|
|
10
|
-
Requires-Python: >=3.8
|
|
11
|
-
Requires-Dist: torch>=1.10.0
|
|
12
|
-
Requires-Dist: rich
|
|
13
|
-
Requires-Dist: tensorboard
|
|
14
|
-
Requires-Dist: matplotlib
|
|
15
|
-
Requires-Dist: seaborn
|
|
16
|
-
Requires-Dist: numpy
|
|
17
|
-
Requires-Dist: scikit-learn
|
|
18
|
-
Requires-Dist: einops
|
|
19
|
-
Requires-Dist: tokenizers
|
|
20
|
-
Dynamic: author
|
|
21
|
-
Dynamic: author-email
|
|
22
|
-
Dynamic: classifier
|
|
23
|
-
Dynamic: requires-dist
|
|
24
|
-
Dynamic: requires-python
|
|
25
|
-
Dynamic: summary
|
|
@@ -1,29 +0,0 @@
|
|
|
1
|
-
orbit/__init__.py,sha256=xikJCuBzK0VSwJ8gvPt0cZijw5TzVI2t3a62_aK-uEo,51
|
|
2
|
-
orbit/callback.py,sha256=FXq-bOVfoYW0S5S95ry55yCn1QgYkecpHonO35BNKwE,1738
|
|
3
|
-
orbit/engine.py,sha256=4Vskd39cfVa7jUWucikva8oDc68om4DtGH6OZLUDhKE,34230
|
|
4
|
-
orbit/optim/__init__.py,sha256=jp3TZFLM5LRcHeeYjR6qNwzwo1dfeR65A6DSGCEPnFg,67
|
|
5
|
-
orbit/optim/muon.py,sha256=IT0l1b0mAtLpB-SUuPyo9uaYHO2x_wkqg1TLcDJ6Bvk,7342
|
|
6
|
-
orbit/optim/sam.py,sha256=lDfou1jQYM_qtlWgYzfUKp2Zdi-uBSKZS7iYxB-yGkU,3702
|
|
7
|
-
orbit/plugin/__init__.py,sha256=9EcMoS__kzldM4MBNUwv5xUs21rliqrBALVmzIk_ADw,435
|
|
8
|
-
orbit/plugin/board.py,sha256=dK3aaTmznq9vSYvpSPgfE9kt30tieR93oK0oDaC8yJw,2267
|
|
9
|
-
orbit/plugin/checkpoint.py,sha256=w9R5O2-jzbRPSGPiuGuUycq21EebAdqigEFK0T0waOE,10898
|
|
10
|
-
orbit/plugin/classification.py,sha256=iVOFNUaL4SUMNEtnmbnioprb2nH1sOV4IsGH7UPhU8U,7019
|
|
11
|
-
orbit/plugin/display_model.py,sha256=Uch563Jq0R78lGHYiDPotXB6dmWfqdClSOiuOIKw7i8,2907
|
|
12
|
-
orbit/plugin/early_stopping.py,sha256=eDTjIzxDZvnCXANR2RXn8vrMh5wX8F2rQjl0BNp-6dU,4065
|
|
13
|
-
orbit/plugin/ema.py,sha256=NtQVB3Rz5YreQwWicXCNy_CfD6d7pu0crGGrz6-IJI8,3892
|
|
14
|
-
orbit/plugin/gradient_accumulation.py,sha256=JaSl-U9Zo_jeEaJE-KAXEPhQdjFA_He729VaCw2W8n4,1066
|
|
15
|
-
orbit/plugin/memory_estimator.py,sha256=YxSXbjAwl6jJh6yFEod5Tu_pgZh1Qadb0JsgmyfF7Mc,9743
|
|
16
|
-
orbit/plugin/mentor.py,sha256=sx3gLHnxYJX1tJhqnqHPTW2KuTeNM7h0PqOIquxMaxQ,12474
|
|
17
|
-
orbit/plugin/overfit.py,sha256=lpfxX4igi4uNdY4OR2_fpayw52iEfkCQnQ0HgzodZd4,1194
|
|
18
|
-
orbit/plugin/warmup.py,sha256=jS3Q52zuhTdqN9Bw85zcoH1YXZoqyRvnD7JHSSLRUwQ,4996
|
|
19
|
-
orbit/plugin/data/mentor_i18n.json,sha256=KHnEKRKHeLga_nqEUZVgiTh4jb4j2f_yD0MJT-7sdis,8005
|
|
20
|
-
orbit/utils/__init__.py,sha256=gXNfTz1a_kL5UZXT2o5BhoItc4Dgrk8w4K2hsfkbiYk,589
|
|
21
|
-
orbit/utils/freeze.py,sha256=ujR5wl2GHuqIQi0rm2i_yz6W8dj6X4YqgyZ3tBbv4yU,2405
|
|
22
|
-
orbit/utils/initialization.py,sha256=X0-OP0FPWX60_SRuB-x_Od3KdDh4EAD9UbOCZFj6NtI,20560
|
|
23
|
-
orbit/utils/layer_io.py,sha256=E-YnYw7av2ZdczDzhs08hUQtbZrcHPvqDdzEvWgs7Gs,1872
|
|
24
|
-
orbit/utils/mask.py,sha256=DayMxmKlWMYMYPceYhYlf9lIx4FD9s4Tp1WvzfVpsu0,3072
|
|
25
|
-
orbit/utils/seed.py,sha256=tbjF2jIPQfb0M_-wXOMVH3_XzWVsDkuhzD5lctGqalA,2578
|
|
26
|
-
orbit_torch-0.0.4a1.dist-info/METADATA,sha256=g2NSH_PVm9afSyVAfvct5xvk2hFColq85tKdbGQfNFc,700
|
|
27
|
-
orbit_torch-0.0.4a1.dist-info/WHEEL,sha256=_zCd3N1l69ArxyTb8rzEoP9TpbYXkqRFSNOD5OuxnTs,91
|
|
28
|
-
orbit_torch-0.0.4a1.dist-info/top_level.txt,sha256=emrF0of931NzTSL4R5yBKpGoewFCB-cAwYNcUF5cqBs,6
|
|
29
|
-
orbit_torch-0.0.4a1.dist-info/RECORD,,
|
|
File without changes
|
|
File without changes
|