evolutionary-policy-optimization 0.1.18__tar.gz → 0.1.19__tar.gz

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (16) hide show
  1. {evolutionary_policy_optimization-0.1.18 → evolutionary_policy_optimization-0.1.19}/PKG-INFO +1 -1
  2. {evolutionary_policy_optimization-0.1.18 → evolutionary_policy_optimization-0.1.19}/evolutionary_policy_optimization/epo.py +19 -0
  3. {evolutionary_policy_optimization-0.1.18 → evolutionary_policy_optimization-0.1.19}/pyproject.toml +1 -1
  4. {evolutionary_policy_optimization-0.1.18 → evolutionary_policy_optimization-0.1.19}/.github/workflows/python-publish.yml +0 -0
  5. {evolutionary_policy_optimization-0.1.18 → evolutionary_policy_optimization-0.1.19}/.github/workflows/test.yml +0 -0
  6. {evolutionary_policy_optimization-0.1.18 → evolutionary_policy_optimization-0.1.19}/.gitignore +0 -0
  7. {evolutionary_policy_optimization-0.1.18 → evolutionary_policy_optimization-0.1.19}/LICENSE +0 -0
  8. {evolutionary_policy_optimization-0.1.18 → evolutionary_policy_optimization-0.1.19}/README.md +0 -0
  9. {evolutionary_policy_optimization-0.1.18 → evolutionary_policy_optimization-0.1.19}/evolutionary_policy_optimization/__init__.py +0 -0
  10. {evolutionary_policy_optimization-0.1.18 → evolutionary_policy_optimization-0.1.19}/evolutionary_policy_optimization/distributed.py +0 -0
  11. {evolutionary_policy_optimization-0.1.18 → evolutionary_policy_optimization-0.1.19}/evolutionary_policy_optimization/env_wrappers.py +0 -0
  12. {evolutionary_policy_optimization-0.1.18 → evolutionary_policy_optimization-0.1.19}/evolutionary_policy_optimization/experimental.py +0 -0
  13. {evolutionary_policy_optimization-0.1.18 → evolutionary_policy_optimization-0.1.19}/evolutionary_policy_optimization/mock_env.py +0 -0
  14. {evolutionary_policy_optimization-0.1.18 → evolutionary_policy_optimization-0.1.19}/requirements.txt +0 -0
  15. {evolutionary_policy_optimization-0.1.18 → evolutionary_policy_optimization-0.1.19}/tests/test_epo.py +0 -0
  16. {evolutionary_policy_optimization-0.1.18 → evolutionary_policy_optimization-0.1.19}/train_gym.py +0 -0
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.4
2
2
  Name: evolutionary-policy-optimization
3
- Version: 0.1.18
3
+ Version: 0.1.19
4
4
  Summary: EPO - Pytorch
5
5
  Project-URL: Homepage, https://pypi.org/project/evolutionary-policy-optimization/
6
6
  Project-URL: Repository, https://github.com/lucidrains/evolutionary-policy-optimization
@@ -1068,6 +1068,12 @@ class Agent(Module):
1068
1068
  def unwrapped_latent_gene_pool(self):
1069
1069
  return self.unwrap_model(self.latent_gene_pool)
1070
1070
 
1071
+ def log(self, **data_kwargs):
1072
+ if not self.wrap_with_accelerate:
1073
+ return
1074
+
1075
+ self.accelerate.log(data_kwargs, step = self.step)
1076
+
1071
1077
  def save(self, path, overwrite = False):
1072
1078
  path = Path(path)
1073
1079
 
@@ -1283,6 +1289,14 @@ class Agent(Module):
1283
1289
  self.critic_optim.step()
1284
1290
  self.critic_optim.zero_grad()
1285
1291
 
1292
+ # log actor critic loss
1293
+
1294
+ self.log(
1295
+ actor_loss = actor_loss.item(),
1296
+ critic_loss = critic_loss.item(),
1297
+ fitness_scores = fitness_scores
1298
+ )
1299
+
1286
1300
  # maybe ema update critic
1287
1301
 
1288
1302
  if self.use_critic_ema:
@@ -1307,6 +1321,11 @@ class Agent(Module):
1307
1321
  self.latent_optim.step()
1308
1322
  self.latent_optim.zero_grad()
1309
1323
 
1324
+ if self.has_diversity_loss:
1325
+ self.log(
1326
+ diversity_loss = diversity_loss.item()
1327
+ )
1328
+
1310
1329
  # apply evolution
1311
1330
 
1312
1331
  if self.has_latent_genes:
@@ -1,6 +1,6 @@
1
1
  [project]
2
2
  name = "evolutionary-policy-optimization"
3
- version = "0.1.18"
3
+ version = "0.1.19"
4
4
  description = "EPO - Pytorch"
5
5
  authors = [
6
6
  { name = "Phil Wang", email = "lucidrains@gmail.com" }