evolutionary-policy-optimization 0.0.16__py3-none-any.whl → 0.0.18__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -15,6 +15,8 @@ from assoc_scan import AssocScan
15
15
 
16
16
  from adam_atan2_pytorch import AdoptAtan2
17
17
 
18
+ from hl_gauss_pytorch import HLGaussLayer
19
+
18
20
  # helpers
19
21
 
20
22
  def exists(v):
@@ -204,6 +206,8 @@ class Actor(Module):
204
206
  assert len(dim_hiddens) >= 2
205
207
  dim_first, *_, dim_last = dim_hiddens
206
208
 
209
+ self.dim_latent = dim_latent
210
+
207
211
  self.init_layer = nn.Sequential(
208
212
  nn.Linear(dim_state, dim_first),
209
213
  nn.SiLU()
@@ -240,6 +244,8 @@ class Critic(Module):
240
244
  assert len(dim_hiddens) >= 2
241
245
  dim_first, *_, dim_last = dim_hiddens
242
246
 
247
+ self.dim_latent = dim_latent
248
+
243
249
  self.init_layer = nn.Sequential(
244
250
  nn.Linear(dim_state, dim_first),
245
251
  nn.SiLU()
@@ -488,6 +494,8 @@ class Agent(Module):
488
494
 
489
495
  self.latent_gene_pool = latent_gene_pool
490
496
 
497
+ assert actor.dim_latent == critic.dim_latent == latent_gene_pool.dim_latent
498
+
491
499
  # optimizers
492
500
 
493
501
  self.actor_optim = optim_klass(actor.parameters(), lr = actor_lr, **actor_optim_kwargs)
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.4
2
2
  Name: evolutionary-policy-optimization
3
- Version: 0.0.16
3
+ Version: 0.0.18
4
4
  Summary: EPO - Pytorch
5
5
  Project-URL: Homepage, https://pypi.org/project/evolutionary-policy-optimization/
6
6
  Project-URL: Repository, https://github.com/lucidrains/evolutionary-policy-optimization
@@ -33,20 +33,17 @@ Classifier: Intended Audience :: Developers
33
33
  Classifier: License :: OSI Approved :: MIT License
34
34
  Classifier: Programming Language :: Python :: 3.8
35
35
  Classifier: Topic :: Scientific/Engineering :: Artificial Intelligence
36
- Requires-Python: >=3.8
36
+ Requires-Python: >=3.9
37
37
  Requires-Dist: adam-atan2-pytorch
38
38
  Requires-Dist: assoc-scan
39
39
  Requires-Dist: einops>=0.8.0
40
+ Requires-Dist: hl-gauss-pytorch>=0.1.19
40
41
  Requires-Dist: torch>=2.2
41
42
  Requires-Dist: tqdm
42
43
  Provides-Extra: examples
43
44
  Requires-Dist: numpy; extra == 'examples'
44
45
  Requires-Dist: pufferlib>=2.0.6; extra == 'examples'
45
46
  Requires-Dist: tqdm; extra == 'examples'
46
- Provides-Extra: examples-gym
47
- Requires-Dist: box2d-py; extra == 'examples-gym'
48
- Requires-Dist: gymnasium[box2d]>=1.0.0; extra == 'examples-gym'
49
- Requires-Dist: tqdm; extra == 'examples-gym'
50
47
  Provides-Extra: test
51
48
  Requires-Dist: pytest; extra == 'test'
52
49
  Description-Content-Type: text/markdown
@@ -110,4 +107,15 @@ latent_pool.genetic_algorithm_step(fitness) # update latent genes with genetic a
110
107
  }
111
108
  ```
112
109
 
110
+ ```bibtex
111
+ @article{Farebrother2024StopRT,
112
+ title = {Stop Regressing: Training Value Functions via Classification for Scalable Deep RL},
113
+ author = {Jesse Farebrother and Jordi Orbay and Quan Ho Vuong and Adrien Ali Taiga and Yevgen Chebotar and Ted Xiao and Alex Irpan and Sergey Levine and Pablo Samuel Castro and Aleksandra Faust and Aviral Kumar and Rishabh Agarwal},
114
+ journal = {ArXiv},
115
+ year = {2024},
116
+ volume = {abs/2403.03950},
117
+ url = {https://api.semanticscholar.org/CorpusID:268253088}
118
+ }
119
+ ```
120
+
113
121
  *Evolution is cleverer than you are.* - Leslie Orgel
@@ -0,0 +1,7 @@
1
+ evolutionary_policy_optimization/__init__.py,sha256=Qavcia0n13jjaWIS_LPW7QrxSLT_BBeKujCjF9kQjbA,133
2
+ evolutionary_policy_optimization/epo.py,sha256=3pvYPwAEZdrxwwV95Ea1qG4CQjLnyaxAr40opk07LDw,17747
3
+ evolutionary_policy_optimization/experimental.py,sha256=ktBKxRF27Qsj7WIgBpYlWXqMVxO9zOx2oD1JuDYRAwM,548
4
+ evolutionary_policy_optimization-0.0.18.dist-info/METADATA,sha256=BIyCXw2IbMs-x2hDbFs9NR5s2dYEbfbeK_LadUeUc8Q,4860
5
+ evolutionary_policy_optimization-0.0.18.dist-info/WHEEL,sha256=qtCwoSJWgHk21S1Kb4ihdzI2rlJ1ZKaIurTj_ngOhyQ,87
6
+ evolutionary_policy_optimization-0.0.18.dist-info/licenses/LICENSE,sha256=1yCiA9b5nhslTavxPjsQAO-wpOnwJR9-l8LTVi7GJuk,1066
7
+ evolutionary_policy_optimization-0.0.18.dist-info/RECORD,,
@@ -1,7 +0,0 @@
1
- evolutionary_policy_optimization/__init__.py,sha256=Qavcia0n13jjaWIS_LPW7QrxSLT_BBeKujCjF9kQjbA,133
2
- evolutionary_policy_optimization/epo.py,sha256=f_e-TkJRFF1VHG3psJDgLGNIzlEvDSjX0nOsbLaOBrw,17543
3
- evolutionary_policy_optimization/experimental.py,sha256=ktBKxRF27Qsj7WIgBpYlWXqMVxO9zOx2oD1JuDYRAwM,548
4
- evolutionary_policy_optimization-0.0.16.dist-info/METADATA,sha256=Go0IYw3PbRLeHIu_Volhn31YX8WKLiI_YdllA3mj-bE,4480
5
- evolutionary_policy_optimization-0.0.16.dist-info/WHEEL,sha256=qtCwoSJWgHk21S1Kb4ihdzI2rlJ1ZKaIurTj_ngOhyQ,87
6
- evolutionary_policy_optimization-0.0.16.dist-info/licenses/LICENSE,sha256=1yCiA9b5nhslTavxPjsQAO-wpOnwJR9-l8LTVi7GJuk,1066
7
- evolutionary_policy_optimization-0.0.16.dist-info/RECORD,,