gradboard 2.3.3__tar.gz → 2.5.0__tar.gz

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of gradboard might be problematic. Click here for more details.

@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.3
2
2
  Name: gradboard
3
- Version: 2.3.3
3
+ Version: 2.5.0
4
4
  Summary: Easily snowboard down gnarly loss gradients
5
5
  License: MIT
6
6
  Author: Nicholas Bailey
@@ -138,15 +138,7 @@ def get_optimiser(model, optimiser=AdamW, lr=1e-3, weight_decay=1e-2):
138
138
  """
139
139
  # TODO: print a warning when params are excluded from weight decay IFF wd is set
140
140
  weight_decay_exclude = []
141
- for keyword in [
142
- "nondecay",
143
- "bias",
144
- "norm",
145
- "embedding",
146
- "beta",
147
- "input_query",
148
- "reentrant_query",
149
- ]:
141
+ for keyword in ["nondecay", "bias", "norm", "embedding", "beta"]:
150
142
  weight_decay_exclude += [
151
143
  p for name, p in model.named_parameters() if keyword in name.lower()
152
144
  ]
@@ -165,13 +165,11 @@ class PASS:
165
165
  self.max_lr = None
166
166
  self.cool_point = None
167
167
  for p in sorted(points_left_of_min, key=lambda x: x[0]):
168
- if (self.cool_point is None) and (p[1] < minimum[1] + 0.66 * difference):
169
- self.cool_point = p[0]
170
- elif (self.max_lr is None) and (p[1] < minimum[1] + 0.33 * difference):
168
+ if (self.max_lr is None) and (p[1] < minimum[1] + 0.2 * difference):
171
169
  self.max_lr = p[0]
172
170
  else:
173
171
  continue
174
- self.cool_point = min(self.cool_point, self.max_lr / 4)
172
+ self.cool_point = self.max_lr / 60
175
173
  print("High LR", self.max_lr)
176
174
  print("Cool point", self.cool_point)
177
175
 
@@ -1,6 +1,6 @@
1
1
  [project]
2
2
  name = "gradboard"
3
- version = "2.3.3"
3
+ version = "2.5.0"
4
4
  description = "Easily snowboard down gnarly loss gradients"
5
5
  authors = [
6
6
  {name = "Nicholas Bailey"}
File without changes
File without changes
File without changes