orbit-torch 0.0.4a1__py3-none-any.whl → 0.1.0b1__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (62) hide show
  1. orbit/__init__.py +3 -1
  2. orbit/callback.py +4 -3
  3. orbit/dataset/__init__.py +1 -0
  4. orbit/dataset/cogn.py +138 -0
  5. orbit/dataset/data/cogn_en.jsonl +45 -0
  6. orbit/dataset/data/cogn_zh.jsonl +113 -0
  7. orbit/engine.py +210 -146
  8. orbit/kit/__init__.py +2 -0
  9. orbit/kit/interface.py +154 -0
  10. orbit/kit/wrapper.py +157 -0
  11. orbit/model/__init__.py +5 -0
  12. orbit/model/base.py +125 -0
  13. orbit/model/block/__init__.py +34 -0
  14. orbit/model/block/attention.py +265 -0
  15. orbit/model/block/bio.py +537 -0
  16. orbit/model/block/codebook.py +122 -0
  17. orbit/model/block/conv.py +505 -0
  18. orbit/model/block/embedding.py +252 -0
  19. orbit/model/block/film.py +176 -0
  20. orbit/model/block/fusion.py +335 -0
  21. orbit/model/block/gate.py +334 -0
  22. orbit/model/block/lora.py +776 -0
  23. orbit/model/block/mlp.py +68 -0
  24. orbit/model/block/moe.py +94 -0
  25. orbit/model/block/tcn.py +99 -0
  26. orbit/model/config.py +62 -0
  27. orbit/model/kit/__init__.py +6 -0
  28. orbit/model/kit/discriminator.py +46 -0
  29. orbit/model/kit/losses.py +193 -0
  30. orbit/model/motif/__init__.py +0 -0
  31. orbit/model/motif/vision/__init__.py +0 -0
  32. orbit/model/motif/vision/v1.py +645 -0
  33. orbit/model/registry.py +53 -0
  34. orbit/optim/__init__.py +2 -2
  35. orbit/optim/sam.py +10 -3
  36. orbit/plugin/__init__.py +12 -8
  37. orbit/plugin/board.py +1 -2
  38. orbit/plugin/checkpoint.py +137 -62
  39. orbit/plugin/classification.py +2 -2
  40. orbit/plugin/display_model.py +1 -2
  41. orbit/plugin/early_stopping.py +1 -2
  42. orbit/plugin/ema.py +1 -2
  43. orbit/plugin/gradient_accumulation.py +1 -2
  44. orbit/plugin/lora.py +346 -0
  45. orbit/plugin/memory_estimator.py +1 -2
  46. orbit/plugin/warmup.py +1 -2
  47. orbit/utils/__init__.py +24 -1
  48. orbit/utils/cuda.py +10 -0
  49. orbit/utils/freeze.py +61 -17
  50. orbit/utils/image.py +164 -0
  51. orbit/utils/initialization.py +184 -94
  52. orbit/utils/layer_io.py +66 -7
  53. orbit/utils/lora.py +480 -0
  54. orbit/utils/moe.py +55 -0
  55. orbit/utils/seed.py +3 -19
  56. orbit/utils/sft.py +93 -0
  57. orbit_torch-0.1.0b1.dist-info/METADATA +208 -0
  58. orbit_torch-0.1.0b1.dist-info/RECORD +65 -0
  59. orbit_torch-0.0.4a1.dist-info/METADATA +0 -25
  60. orbit_torch-0.0.4a1.dist-info/RECORD +0 -29
  61. {orbit_torch-0.0.4a1.dist-info → orbit_torch-0.1.0b1.dist-info}/WHEEL +0 -0
  62. {orbit_torch-0.0.4a1.dist-info → orbit_torch-0.1.0b1.dist-info}/top_level.txt +0 -0
@@ -0,0 +1,208 @@
1
+ Metadata-Version: 2.4
2
+ Name: orbit-torch
3
+ Version: 0.1.0b1
4
+ Summary: A PyTorch training engine with plugin system and advanced model components
5
+ Home-page: https://github.com/A03HCY/Orbit
6
+ Author: Aiden Hopkins
7
+ Author-email: acdphc@qq.com
8
+ Classifier: Programming Language :: Python :: 3
9
+ Classifier: License :: OSI Approved :: MIT License
10
+ Classifier: Operating System :: OS Independent
11
+ Requires-Python: >=3.8
12
+ Description-Content-Type: text/markdown
13
+ Requires-Dist: torch>=2.0.0
14
+ Requires-Dist: rich
15
+ Requires-Dist: tensorboard
16
+ Requires-Dist: matplotlib
17
+ Requires-Dist: seaborn
18
+ Requires-Dist: numpy
19
+ Requires-Dist: scikit-learn
20
+ Requires-Dist: einops
21
+ Requires-Dist: tokenizers
22
+ Requires-Dist: transformers
23
+ Requires-Dist: safetensors
24
+ Requires-Dist: accelerate
25
+ Requires-Dist: lpips
26
+ Dynamic: author
27
+ Dynamic: author-email
28
+ Dynamic: classifier
29
+ Dynamic: description
30
+ Dynamic: description-content-type
31
+ Dynamic: home-page
32
+ Dynamic: requires-dist
33
+ Dynamic: requires-python
34
+ Dynamic: summary
35
+
36
+ # Orbit
37
+
38
+ Orbit is a flexible, plugin-based PyTorch training engine designed to simplify the training loop while providing powerful components for modern deep learning models, including LLMs.
39
+
40
+ It features a modular design with a rich set of plugins, advanced model building blocks (like MoE, RoPE, GQA), comprehensive LoRA/DoRA support, and cutting-edge optimizers.
41
+
42
+ ## Features
43
+
44
+ ### 🚀 Core Engine
45
+ - **Plugin System**: Decoupled training logic using plugins for callbacks, logging, and training strategies.
46
+ - **Simplified Loop**: Clean `train` and `eval` interfaces.
47
+ - **Flexible Updates**:
48
+ - `auto_update()`: Automatically handles forward pass, loss calculation, backward pass, optimizer step, and zero grad.
49
+ - `update(loss)`: Allows manual control over the update step if you need custom forward/loss logic.
50
+
51
+ ### 🧩 Model Components (`orbit.model`)
52
+ Orbit provides a collection of high-performance, reusable layers:
53
+ - **Attention**: `MultiHeadAttention` with support for **GQA** (Grouped Query Attention), **RoPE** (Rotary Positional Embeddings), and FlashAttention.
54
+ - **LoRA & DoRA**: Full support for Low-Rank Adaptation and **Weight-Decomposed Low-Rank Adaptation (DoRA)** across `Linear`, `Conv2d`, `Conv1d`, and `Embedding` layers. Also supports **Gated LoRA**.
55
+ - **MoE**: Mixture of Experts block with `TopKGate` routing.
56
+ - **Gates**: A variety of gating mechanisms including `SigmoidGate`, `TanhGate`, `SoftmaxGate`, `GLUGate`, `ContextGate`, and `TopKGate`.
57
+ - **Others**: `FiLM` (Feature-wise Linear Modulation), `MLP` (with Gated support), `RotaryPositionalEmbedding`.
58
+
59
+ ### 🛠️ Utilities & Kit (`orbit.utils`)
60
+
61
+ Orbit provides a comprehensive toolkit to speed up development:
62
+
63
+ #### 🔧 LoRA Utilities
64
+ Manual control over LoRA injection and management (alternative to the Plugin approach).
65
+ - **Injection**:
66
+ - `inject_lora(model, r=8, ...)`: Manually inject LoRA/DoRA/Gated LoRA into specific layers.
67
+ - `inject_lora_file(model, path)`: Automatically inject and load LoRA configuration/weights from a file.
68
+ - **Management**:
69
+ - `merge_lora(model)` / `unmerge_lora(model)`: Merge weights for faster inference or unmerge to resume training.
70
+ - `save_lora(model, path)` / `load_lora(model, path)`: Efficiently save/load only LoRA parameters.
71
+ - `freeze_backbone_only(model)`: Helper to freeze the base model while keeping LoRA and specified heads trainable.
72
+ - **Diagnosis**:
73
+ - `LoRADiagnoser`: Check for rank collapse and monitor gradient norms during training.
74
+
75
+ #### ❄️ Model Freezing
76
+ - `freeze_layers(model, targets=['encoder'])`: Freeze layers matching the target names (supports wildcards).
77
+ - `unfreeze_layers(model, targets)`: Unfreeze specific layers.
78
+ - `get_trainable_params(model)`: Get parameters for the optimizer.
79
+
80
+ #### 🎭 Masking
81
+ - `make_causal_mask`: Create causal masks for autoregressive models.
82
+ - `make_padding_mask`, `make_lookahead_mask`, `make_sliding_window_mask`.
83
+
84
+ #### 💾 Layer I/O
85
+ - `save_layer(model, layer_name, path)`: Save weights of a specific sub-module (e.g., just the backbone).
86
+ - `load_layer(model, layer_name, path)`: Load weights into a specific sub-module.
87
+ - `get_model_by_name(model, name)`: Access sub-modules using dot notation strings (e.g., "backbone.layer1").
88
+
89
+ #### 📝 SFT Helpers
90
+ - `build_sft`: Prepares data for Supervised Fine-Tuning (handles chat templates, tokenization, and label masking).
91
+ - `train_sft(engine)`: A specialized training step for SFT that handles the forward pass and loss calculation automatically.
92
+
93
+ #### ⚙️ Optimization (`orbit.optim`)
94
+ - **Muon**: MomentUm Orthogonalized by Newton-schulz optimizer.
95
+ - **SAM**: Sharpness-Aware Minimization wrapper.
96
+
97
+ #### 🌱 Initialization & Seeding
98
+ - `auto_initialize(model)`: Automatically initializes weights based on layer type (Linear, Conv, Embedding, etc.).
99
+ - `seed_everything(seed)`: Sets seeds for Python, NumPy, PyTorch, and CUDA for reproducibility.
100
+
101
+ #### 🖥️ CUDA
102
+ - `cuda_alloc(size)`: Optimizes PyTorch CUDA memory allocation configuration (e.g., `max_split_size_mb`).
103
+
104
+ ### 🔌 Plugins (`orbit.plugin`)
105
+ - `EarlyStopping`: Stop training when a metric stops improving.
106
+ - `GradientAccumulation`: Simulate larger batch sizes.
107
+ - `Warmup`: Learning rate warmup.
108
+ - `Mentor`: Training assistant/logger.
109
+ - `MemoryEstimator`: Monitor CUDA memory usage.
110
+ - `LoRA`: Easy injection of LoRA layers via plugin.
111
+ - `Board`: TensorBoard integration.
112
+
113
+ ## Installation
114
+
115
+ ```bash
116
+ pip install orbit-torch
117
+ ```
118
+
119
+ **Requirements**:
120
+ - Python >= 3.8
121
+ - PyTorch >= 2.0.0 (Required for FlashAttention backend)
122
+
123
+ ## Quick Start
124
+
125
+ ### 1. Basic Training (CIFAR-10)
126
+
127
+ ```python
128
+ import torch
129
+ import torch.nn as nn
130
+ from orbit.engine import Engine
131
+ from orbit.plugin import EarlyStopping, GradientAccumulation, Mentor
132
+ from orbit.utils import auto_initialize
133
+
134
+ # Define your model
135
+ model = MyConvNet()
136
+ auto_initialize(model)
137
+
138
+ # Setup Engine
139
+ trainer = Engine(
140
+ model=model,
141
+ criterion=nn.CrossEntropyLoss(),
142
+ optimizer=torch.optim.Adam(model.parameters(), lr=1e-3),
143
+ plugins=[
144
+ Mentor(),
145
+ EarlyStopping(monitor='val_acc', patience=3),
146
+ GradientAccumulation(steps=2)
147
+ ]
148
+ )
149
+
150
+ # Train
151
+ for _ in trainer.train(train_loader, num_epochs=10):
152
+ trainer.auto_update() # Handles forward, backward, step, zero_grad
153
+
154
+ # Handle Epoch End (e.g., Validation)
155
+ if not trainer.is_epoch_end: continue
156
+
157
+ for _ in trainer.eval(test_loader):
158
+ trainer.auto_update()
159
+ ```
160
+
161
+ ### 2. LLM SFT with LoRA/DoRA
162
+
163
+ Orbit makes it easy to fine-tune LLMs using LoRA or DoRA.
164
+
165
+ ```python
166
+ from transformers import AutoModelForCausalLM, AutoTokenizer
167
+ from orbit.engine import Engine
168
+ from orbit.plugin import LoRA, GradientAccumulation
169
+ from orbit.utils import train_sft, seed_everything
170
+
171
+ seed_everything(42)
172
+
173
+ # Load Model
174
+ model = AutoModelForCausalLM.from_pretrained("Qwen/Qwen2.5-3B", device_map="auto")
175
+ tokenizer = AutoTokenizer.from_pretrained("Qwen/Qwen2.5-3B")
176
+
177
+ # Setup Engine with LoRA Plugin
178
+ trainer = Engine(
179
+ model=model,
180
+ optimizer=torch.optim.AdamW(model.parameters(), lr=1e-4),
181
+ plugins=[
182
+ # Inject DoRA into MLP layers
183
+ LoRA(target_names=['mlp'], dora=True, r=16, alpha=32),
184
+ GradientAccumulation(steps=8)
185
+ ]
186
+ )
187
+
188
+ # Train Loop
189
+ # Assuming `dataloader` yields SFT batches (input_ids, attention_mask, labels)
190
+ for _ in trainer.train(dataloader, num_epochs=3):
191
+ # train_sft handles the forward pass and loss calculation for CausalLM
192
+ train_sft(trainer)
193
+ ```
194
+
195
+ ### 3. Chat Interface
196
+
197
+ Interact with your trained model in the terminal:
198
+
199
+ ```python
200
+ from orbit.kit import ChatInterface
201
+
202
+ chat = ChatInterface(model_id="path/to/model", device="cuda")
203
+ chat.interact()
204
+ ```
205
+
206
+ ## License
207
+
208
+ MIT License
@@ -0,0 +1,65 @@
1
+ orbit/__init__.py,sha256=Pnt9BItQOlPZR4StcWQ3C-XmYHvqmLSyJiF0honrEew,78
2
+ orbit/callback.py,sha256=qI4iaHwY5AXpzgSHNdiAEgK5XJxv8eEMTg9YxpzVeFk,1669
3
+ orbit/engine.py,sha256=-S0kNNu_rgX17xrurKDgigqjzgiYN3AKoD4QDXqi80g,35894
4
+ orbit/dataset/__init__.py,sha256=OAhQmm3GYwOfsznpGRNEhPr0Lanw_iIrqJc7FV2gbBc,46
5
+ orbit/dataset/cogn.py,sha256=kN_CPkV_VnsS7pVN0tdXsah0dYr780K7gfo3ROIGWyY,4749
6
+ orbit/dataset/data/cogn_en.jsonl,sha256=Ba_KQKW3BJs3jWD7B_z0_Lwq34Zujgu0M2xe7RWO0dA,14862
7
+ orbit/dataset/data/cogn_zh.jsonl,sha256=d__Dr_5Z3w33KcX8CikK34VNt9DVAFUnDvnhCJikumk,39055
8
+ orbit/kit/__init__.py,sha256=40DZDZwGKgOeJ0vUsqhCBKHt-H1Tj0a79dYrjfKKOgs,82
9
+ orbit/kit/interface.py,sha256=_eB6PguJk-EjpLjnCuGDTHJ2woEhHxqW36sasCYC9Ws,5658
10
+ orbit/kit/wrapper.py,sha256=PSBDEx4fsnuznZ3-9EYcUTr4hgQsRPDT9EiI-rxvcaY,5941
11
+ orbit/model/__init__.py,sha256=DY9le8idTuOgCu-K2Os4j83Fa6X55G3VT1DpGP6GvTs,151
12
+ orbit/model/base.py,sha256=NekTUd1P_0e8yHsRUoIJWJbNYqf0-XdDTgFk9WeNBOQ,4264
13
+ orbit/model/config.py,sha256=2IMe6bxES_qf1hXNhroMuAosN9BCZfJJTMJryrz0Q2c,1888
14
+ orbit/model/registry.py,sha256=r1nMCYTpBT24Fu-mOL-S37zOQA-ngpd4uHvonnkN6Lo,1606
15
+ orbit/model/block/__init__.py,sha256=NvNs7F1MfUxOmpHQgrTE0XcJ20OOreTBCGKTYxtggtc,938
16
+ orbit/model/block/attention.py,sha256=7eSE0GM5wW2gsjcTdqUoHEyd8VqCJ1bNbfF-HY7mAUU,10131
17
+ orbit/model/block/bio.py,sha256=SNCG1urEfqDoX-B7-hNpHWJ_vHrU8EBJP9-sjdeh2Nw,21291
18
+ orbit/model/block/codebook.py,sha256=IhxOp0_--euNvrNCU95_h8e38XA6HBuZef8dDC7Ex-g,4195
19
+ orbit/model/block/conv.py,sha256=NswgOt5daz60ROOqmorrftcxI5I0aDQ14rnorYfUcpM,18616
20
+ orbit/model/block/embedding.py,sha256=hMFsx2gYhm_qNoF22Xg2rkOKf5IP16fflUttV8q1OA8,9101
21
+ orbit/model/block/film.py,sha256=oBWmm2mclSjE4yyVf-Qy8n5AeoGkj4xBGERylqNFkMs,6542
22
+ orbit/model/block/fusion.py,sha256=YbyFiIOgbYD0ywrqIPPr0T662yEmc4VcJBT0OSR-yEw,12361
23
+ orbit/model/block/gate.py,sha256=ODE-bKRVaqkTzhAI3z_JCNINfCq1fosQHSrWHqeTmBk,10648
24
+ orbit/model/block/lora.py,sha256=ggmbj9_qQu0FzkqE1A1XRnuIv_bOCVDIPergN-kZPQc,31208
25
+ orbit/model/block/mlp.py,sha256=EcJWbLREPoBVCgQDHW6nJ-2zuxZ0J-biFhZNXd0BFRQ,2226
26
+ orbit/model/block/moe.py,sha256=JW-1esQ986T6_wDQwF7065RjkTSqPx9vEcNppSPG0tE,3151
27
+ orbit/model/block/tcn.py,sha256=ClkYw2cMFcXFm6mMbrUZnraZ5YtokGotU1_sJBGANAU,3836
28
+ orbit/model/kit/__init__.py,sha256=-Q8mGi39sDmUZP9P3nZ1GoyvEOz-rfa0_JEXSzlfplg,156
29
+ orbit/model/kit/discriminator.py,sha256=JA30tHCjmZ1L1oPf-7o5w1nM-_9h1gB9PiPkkNz_JSE,1484
30
+ orbit/model/kit/losses.py,sha256=CSR3aH6SeUIA72sopi2XAQDDHyFelKhZAZcNG3tLKmM,7697
31
+ orbit/model/motif/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
32
+ orbit/model/motif/vision/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
33
+ orbit/model/motif/vision/v1.py,sha256=a1guncpocl1x57pGpVa60FR_7rHVBV3v_JOVLgLee4o,22981
34
+ orbit/optim/__init__.py,sha256=tO24DdBQ7hfBbk2tKIwh5Y0WXqEYjxowLuwXCuScFMI,45
35
+ orbit/optim/muon.py,sha256=IT0l1b0mAtLpB-SUuPyo9uaYHO2x_wkqg1TLcDJ6Bvk,7342
36
+ orbit/optim/sam.py,sha256=aeqWF7aCfZuhuzlSXFEGP8fUTbEDp4Cm75Q-Xr8I6Ho,3949
37
+ orbit/plugin/__init__.py,sha256=VJbpVtYaGB8roeSrv1PCaGsKmlqDTHpJ91LkeEh1xHg,466
38
+ orbit/plugin/board.py,sha256=ofuxEJf1CC2I7Y5fR_3_Ln04HPQqBKTVp2-8CU7lvI8,2266
39
+ orbit/plugin/checkpoint.py,sha256=R-R-180KWcF264DPvWgjGAoWtX-Vkn0nS0MQjOBfDkU,15264
40
+ orbit/plugin/classification.py,sha256=h2I_a19NzC8RLrO5qD1TbaZ6QBh3Wig8m0-kBiXNdn4,7020
41
+ orbit/plugin/display_model.py,sha256=DuvefqcHCHUCyS8bybBzxaXr6Y8gxlpql-Sa9hePBuw,2906
42
+ orbit/plugin/early_stopping.py,sha256=E8l4vSiZ1oECK9SeXFcrhMuDEC72PGHTwB4p_oZghIc,4060
43
+ orbit/plugin/ema.py,sha256=4HnVOfFDRwdjUeOPkkOVj-A6fA8GPEBetsFPlI5wVuk,3887
44
+ orbit/plugin/gradient_accumulation.py,sha256=Hr2TyPDUx8SOBJ6IpBfP5wPBvKj-JgYBNlF28LIs9-U,1061
45
+ orbit/plugin/lora.py,sha256=LmpBc1Fo54eANsAVf36W93O8tLiKcv9z1Ofv8PgtIJ0,17032
46
+ orbit/plugin/memory_estimator.py,sha256=7YrlAJb2yOh6nTaNK29NAu_aLWHpboIN8mnqYIL4LZo,9738
47
+ orbit/plugin/mentor.py,sha256=sx3gLHnxYJX1tJhqnqHPTW2KuTeNM7h0PqOIquxMaxQ,12474
48
+ orbit/plugin/overfit.py,sha256=lpfxX4igi4uNdY4OR2_fpayw52iEfkCQnQ0HgzodZd4,1194
49
+ orbit/plugin/warmup.py,sha256=XvU47kNSk43OSVPkS1riyhndySOK-Vo5M_SReUjxSHo,4991
50
+ orbit/plugin/data/mentor_i18n.json,sha256=KHnEKRKHeLga_nqEUZVgiTh4jb4j2f_yD0MJT-7sdis,8005
51
+ orbit/utils/__init__.py,sha256=2rxWZ27U4oZH0XxqicYOV8oeWQh2NTv-igakgdRPbMI,1093
52
+ orbit/utils/cuda.py,sha256=5JUxXVDJ_IzQfEI9oguTKSz2uSLT2GTFNXbxSZVWhFM,267
53
+ orbit/utils/freeze.py,sha256=ee6MxqoLaOZ-kf5DZyzUAsS5JUCFQPU_2LxkLZQIxJo,4265
54
+ orbit/utils/image.py,sha256=wklEHDcH_TGvT3aDBPKSJvJWHGgHu4yxssuVfZ8u4No,6775
55
+ orbit/utils/initialization.py,sha256=tIRQMlyNfcVDvrtLwLKthpyCjJ_Cy7AN5ZkggFt9o0s,23407
56
+ orbit/utils/layer_io.py,sha256=_qGmt2GDOxVb0RacTK4ZMALek36oQLYVqCeuwcXzaZE,3920
57
+ orbit/utils/lora.py,sha256=LXqSuh9QGTRn_WJsOnJPmiSy0uzpFnWpLxAmLWq_twc,18458
58
+ orbit/utils/mask.py,sha256=DayMxmKlWMYMYPceYhYlf9lIx4FD9s4Tp1WvzfVpsu0,3072
59
+ orbit/utils/moe.py,sha256=miLesFrUHIe9gWr0STs4d4NgnUbV7nskq5N4ywRtEj0,2132
60
+ orbit/utils/seed.py,sha256=PaApRMEO2codOXDcmelybRJ2ZgAk-Sh4K5SHlZ4KpHE,1789
61
+ orbit/utils/sft.py,sha256=Y6hDGW4tAgtzmEeSqolkYcBKYrOJvuW1vxgJMWvzdRU,3127
62
+ orbit_torch-0.1.0b1.dist-info/METADATA,sha256=IzDxu8AeHbXbrub3QisQybVGdYoYzjx3c1t12MRIzZE,7831
63
+ orbit_torch-0.1.0b1.dist-info/WHEEL,sha256=_zCd3N1l69ArxyTb8rzEoP9TpbYXkqRFSNOD5OuxnTs,91
64
+ orbit_torch-0.1.0b1.dist-info/top_level.txt,sha256=emrF0of931NzTSL4R5yBKpGoewFCB-cAwYNcUF5cqBs,6
65
+ orbit_torch-0.1.0b1.dist-info/RECORD,,
@@ -1,25 +0,0 @@
1
- Metadata-Version: 2.4
2
- Name: orbit-torch
3
- Version: 0.0.4a1
4
- Summary: A PyTorch training engine with plugin system
5
- Author: Aiden Hopkins
6
- Author-email: acdphc@qq.com
7
- Classifier: Programming Language :: Python :: 3
8
- Classifier: License :: OSI Approved :: MIT License
9
- Classifier: Operating System :: OS Independent
10
- Requires-Python: >=3.8
11
- Requires-Dist: torch>=1.10.0
12
- Requires-Dist: rich
13
- Requires-Dist: tensorboard
14
- Requires-Dist: matplotlib
15
- Requires-Dist: seaborn
16
- Requires-Dist: numpy
17
- Requires-Dist: scikit-learn
18
- Requires-Dist: einops
19
- Requires-Dist: tokenizers
20
- Dynamic: author
21
- Dynamic: author-email
22
- Dynamic: classifier
23
- Dynamic: requires-dist
24
- Dynamic: requires-python
25
- Dynamic: summary
@@ -1,29 +0,0 @@
1
- orbit/__init__.py,sha256=xikJCuBzK0VSwJ8gvPt0cZijw5TzVI2t3a62_aK-uEo,51
2
- orbit/callback.py,sha256=FXq-bOVfoYW0S5S95ry55yCn1QgYkecpHonO35BNKwE,1738
3
- orbit/engine.py,sha256=4Vskd39cfVa7jUWucikva8oDc68om4DtGH6OZLUDhKE,34230
4
- orbit/optim/__init__.py,sha256=jp3TZFLM5LRcHeeYjR6qNwzwo1dfeR65A6DSGCEPnFg,67
5
- orbit/optim/muon.py,sha256=IT0l1b0mAtLpB-SUuPyo9uaYHO2x_wkqg1TLcDJ6Bvk,7342
6
- orbit/optim/sam.py,sha256=lDfou1jQYM_qtlWgYzfUKp2Zdi-uBSKZS7iYxB-yGkU,3702
7
- orbit/plugin/__init__.py,sha256=9EcMoS__kzldM4MBNUwv5xUs21rliqrBALVmzIk_ADw,435
8
- orbit/plugin/board.py,sha256=dK3aaTmznq9vSYvpSPgfE9kt30tieR93oK0oDaC8yJw,2267
9
- orbit/plugin/checkpoint.py,sha256=w9R5O2-jzbRPSGPiuGuUycq21EebAdqigEFK0T0waOE,10898
10
- orbit/plugin/classification.py,sha256=iVOFNUaL4SUMNEtnmbnioprb2nH1sOV4IsGH7UPhU8U,7019
11
- orbit/plugin/display_model.py,sha256=Uch563Jq0R78lGHYiDPotXB6dmWfqdClSOiuOIKw7i8,2907
12
- orbit/plugin/early_stopping.py,sha256=eDTjIzxDZvnCXANR2RXn8vrMh5wX8F2rQjl0BNp-6dU,4065
13
- orbit/plugin/ema.py,sha256=NtQVB3Rz5YreQwWicXCNy_CfD6d7pu0crGGrz6-IJI8,3892
14
- orbit/plugin/gradient_accumulation.py,sha256=JaSl-U9Zo_jeEaJE-KAXEPhQdjFA_He729VaCw2W8n4,1066
15
- orbit/plugin/memory_estimator.py,sha256=YxSXbjAwl6jJh6yFEod5Tu_pgZh1Qadb0JsgmyfF7Mc,9743
16
- orbit/plugin/mentor.py,sha256=sx3gLHnxYJX1tJhqnqHPTW2KuTeNM7h0PqOIquxMaxQ,12474
17
- orbit/plugin/overfit.py,sha256=lpfxX4igi4uNdY4OR2_fpayw52iEfkCQnQ0HgzodZd4,1194
18
- orbit/plugin/warmup.py,sha256=jS3Q52zuhTdqN9Bw85zcoH1YXZoqyRvnD7JHSSLRUwQ,4996
19
- orbit/plugin/data/mentor_i18n.json,sha256=KHnEKRKHeLga_nqEUZVgiTh4jb4j2f_yD0MJT-7sdis,8005
20
- orbit/utils/__init__.py,sha256=gXNfTz1a_kL5UZXT2o5BhoItc4Dgrk8w4K2hsfkbiYk,589
21
- orbit/utils/freeze.py,sha256=ujR5wl2GHuqIQi0rm2i_yz6W8dj6X4YqgyZ3tBbv4yU,2405
22
- orbit/utils/initialization.py,sha256=X0-OP0FPWX60_SRuB-x_Od3KdDh4EAD9UbOCZFj6NtI,20560
23
- orbit/utils/layer_io.py,sha256=E-YnYw7av2ZdczDzhs08hUQtbZrcHPvqDdzEvWgs7Gs,1872
24
- orbit/utils/mask.py,sha256=DayMxmKlWMYMYPceYhYlf9lIx4FD9s4Tp1WvzfVpsu0,3072
25
- orbit/utils/seed.py,sha256=tbjF2jIPQfb0M_-wXOMVH3_XzWVsDkuhzD5lctGqalA,2578
26
- orbit_torch-0.0.4a1.dist-info/METADATA,sha256=g2NSH_PVm9afSyVAfvct5xvk2hFColq85tKdbGQfNFc,700
27
- orbit_torch-0.0.4a1.dist-info/WHEEL,sha256=_zCd3N1l69ArxyTb8rzEoP9TpbYXkqRFSNOD5OuxnTs,91
28
- orbit_torch-0.0.4a1.dist-info/top_level.txt,sha256=emrF0of931NzTSL4R5yBKpGoewFCB-cAwYNcUF5cqBs,6
29
- orbit_torch-0.0.4a1.dist-info/RECORD,,