adv-optm 1.2.dev15__tar.gz → 1.2.dev17__tar.gz
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of adv-optm might be problematic. Click here for more details.
- {adv_optm-1.2.dev15 → adv_optm-1.2.dev17}/PKG-INFO +1 -1
- {adv_optm-1.2.dev15 → adv_optm-1.2.dev17}/adv_optm/__init__.py +1 -1
- {adv_optm-1.2.dev15 → adv_optm-1.2.dev17}/adv_optm/optim/Muon_adv.py +3 -0
- {adv_optm-1.2.dev15 → adv_optm-1.2.dev17}/adv_optm/util/Kourkoutas.py +2 -1
- {adv_optm-1.2.dev15 → adv_optm-1.2.dev17}/adv_optm.egg-info/PKG-INFO +1 -1
- {adv_optm-1.2.dev15 → adv_optm-1.2.dev17}/setup.py +1 -1
- {adv_optm-1.2.dev15 → adv_optm-1.2.dev17}/LICENSE +0 -0
- {adv_optm-1.2.dev15 → adv_optm-1.2.dev17}/README.md +0 -0
- {adv_optm-1.2.dev15 → adv_optm-1.2.dev17}/adv_optm/optim/AdaMuon_adv.py +0 -0
- {adv_optm-1.2.dev15 → adv_optm-1.2.dev17}/adv_optm/optim/AdamW_adv.py +0 -0
- {adv_optm-1.2.dev15 → adv_optm-1.2.dev17}/adv_optm/optim/Adopt_adv.py +0 -0
- {adv_optm-1.2.dev15 → adv_optm-1.2.dev17}/adv_optm/optim/Lion_Prodigy_adv.py +0 -0
- {adv_optm-1.2.dev15 → adv_optm-1.2.dev17}/adv_optm/optim/Lion_adv.py +0 -0
- {adv_optm-1.2.dev15 → adv_optm-1.2.dev17}/adv_optm/optim/Prodigy_adv.py +0 -0
- {adv_optm-1.2.dev15 → adv_optm-1.2.dev17}/adv_optm/optim/Simplified_AdEMAMix.py +0 -0
- {adv_optm-1.2.dev15 → adv_optm-1.2.dev17}/adv_optm/optim/__init__.py +0 -0
- {adv_optm-1.2.dev15 → adv_optm-1.2.dev17}/adv_optm/util/BF16_Stochastic_Rounding.py +0 -0
- {adv_optm-1.2.dev15 → adv_optm-1.2.dev17}/adv_optm/util/Effective_Shape.py +0 -0
- {adv_optm-1.2.dev15 → adv_optm-1.2.dev17}/adv_optm/util/NNMF.py +0 -0
- {adv_optm-1.2.dev15 → adv_optm-1.2.dev17}/adv_optm/util/Newton_Schulz.py +0 -0
- {adv_optm-1.2.dev15 → adv_optm-1.2.dev17}/adv_optm/util/One_Bit_Boolean.py +0 -0
- {adv_optm-1.2.dev15 → adv_optm-1.2.dev17}/adv_optm/util/OrthoGrad.py +0 -0
- {adv_optm-1.2.dev15 → adv_optm-1.2.dev17}/adv_optm/util/__init__.py +0 -0
- {adv_optm-1.2.dev15 → adv_optm-1.2.dev17}/adv_optm.egg-info/SOURCES.txt +0 -0
- {adv_optm-1.2.dev15 → adv_optm-1.2.dev17}/adv_optm.egg-info/dependency_links.txt +0 -0
- {adv_optm-1.2.dev15 → adv_optm-1.2.dev17}/adv_optm.egg-info/requires.txt +0 -0
- {adv_optm-1.2.dev15 → adv_optm-1.2.dev17}/adv_optm.egg-info/top_level.txt +0 -0
- {adv_optm-1.2.dev15 → adv_optm-1.2.dev17}/setup.cfg +0 -0
|
@@ -76,6 +76,7 @@ class Muon_adv(torch.optim.Optimizer):
|
|
|
76
76
|
adam_beta3_ema (float): Beta3 for AdEMAMix.
|
|
77
77
|
adam_alpha (float): Alpha for AdEMAMix.
|
|
78
78
|
adam_kourkoutas_beta (bool): Kourkoutas-β for AdamW.
|
|
79
|
+
adam_nnmf_factor (bool): 1-bit factored for AdamW.
|
|
79
80
|
"""
|
|
80
81
|
|
|
81
82
|
def __init__(
|
|
@@ -125,6 +126,7 @@ class Muon_adv(torch.optim.Optimizer):
|
|
|
125
126
|
adam_ema_alpha: float = 0.95,
|
|
126
127
|
adam_tiny_spike: float = 1e-9,
|
|
127
128
|
adam_k_warmup_steps: int = 0,
|
|
129
|
+
adam_nnmf_factor: bool = False,
|
|
128
130
|
):
|
|
129
131
|
if not (lr >= 0.0):
|
|
130
132
|
raise ValueError(f"Learning-rate should be >= 0.0. Got {lr}")
|
|
@@ -165,6 +167,7 @@ class Muon_adv(torch.optim.Optimizer):
|
|
|
165
167
|
"adam_kourkoutas_beta": adam_kourkoutas_beta, "adam_beta2_min": adam_beta2_min,
|
|
166
168
|
"adam_ema_alpha": adam_ema_alpha, "adam_tiny_spike": adam_tiny_spike,
|
|
167
169
|
"adam_k_warmup_steps": adam_k_warmup_steps,
|
|
170
|
+
"adam_nnmf_factor":adam_nnmf_factor,
|
|
168
171
|
}
|
|
169
172
|
self.stochastic_rounding = stochastic_rounding
|
|
170
173
|
self.compiled_optimizer = compiled_optimizer
|
|
@@ -155,4 +155,5 @@ class KourkoutasHelper:
|
|
|
155
155
|
"""
|
|
156
156
|
layer_key = self.optimizer.layer_key_fn(p)
|
|
157
157
|
# The default is the max value, which is correct for unmapped params or edge cases
|
|
158
|
-
|
|
158
|
+
beta2_default = group.get('betas', group.get('adam_betas'))[1] if group.get('betas', group.get('adam_betas')) else 0.999
|
|
159
|
+
return self.layer_state.get(layer_key, {}).get('dynamic_beta2', beta2_default)
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|