evolutionary-policy-optimization 0.0.16__py3-none-any.whl → 0.0.18__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- evolutionary_policy_optimization/epo.py +8 -0
- {evolutionary_policy_optimization-0.0.16.dist-info → evolutionary_policy_optimization-0.0.18.dist-info}/METADATA +14 -6
- evolutionary_policy_optimization-0.0.18.dist-info/RECORD +7 -0
- evolutionary_policy_optimization-0.0.16.dist-info/RECORD +0 -7
- {evolutionary_policy_optimization-0.0.16.dist-info → evolutionary_policy_optimization-0.0.18.dist-info}/WHEEL +0 -0
- {evolutionary_policy_optimization-0.0.16.dist-info → evolutionary_policy_optimization-0.0.18.dist-info}/licenses/LICENSE +0 -0
@@ -15,6 +15,8 @@ from assoc_scan import AssocScan
|
|
15
15
|
|
16
16
|
from adam_atan2_pytorch import AdoptAtan2
|
17
17
|
|
18
|
+
from hl_gauss_pytorch import HLGaussLayer
|
19
|
+
|
18
20
|
# helpers
|
19
21
|
|
20
22
|
def exists(v):
|
@@ -204,6 +206,8 @@ class Actor(Module):
|
|
204
206
|
assert len(dim_hiddens) >= 2
|
205
207
|
dim_first, *_, dim_last = dim_hiddens
|
206
208
|
|
209
|
+
self.dim_latent = dim_latent
|
210
|
+
|
207
211
|
self.init_layer = nn.Sequential(
|
208
212
|
nn.Linear(dim_state, dim_first),
|
209
213
|
nn.SiLU()
|
@@ -240,6 +244,8 @@ class Critic(Module):
|
|
240
244
|
assert len(dim_hiddens) >= 2
|
241
245
|
dim_first, *_, dim_last = dim_hiddens
|
242
246
|
|
247
|
+
self.dim_latent = dim_latent
|
248
|
+
|
243
249
|
self.init_layer = nn.Sequential(
|
244
250
|
nn.Linear(dim_state, dim_first),
|
245
251
|
nn.SiLU()
|
@@ -488,6 +494,8 @@ class Agent(Module):
|
|
488
494
|
|
489
495
|
self.latent_gene_pool = latent_gene_pool
|
490
496
|
|
497
|
+
assert actor.dim_latent == critic.dim_latent == latent_gene_pool.dim_latent
|
498
|
+
|
491
499
|
# optimizers
|
492
500
|
|
493
501
|
self.actor_optim = optim_klass(actor.parameters(), lr = actor_lr, **actor_optim_kwargs)
|
@@ -1,6 +1,6 @@
|
|
1
1
|
Metadata-Version: 2.4
|
2
2
|
Name: evolutionary-policy-optimization
|
3
|
-
Version: 0.0.
|
3
|
+
Version: 0.0.18
|
4
4
|
Summary: EPO - Pytorch
|
5
5
|
Project-URL: Homepage, https://pypi.org/project/evolutionary-policy-optimization/
|
6
6
|
Project-URL: Repository, https://github.com/lucidrains/evolutionary-policy-optimization
|
@@ -33,20 +33,17 @@ Classifier: Intended Audience :: Developers
|
|
33
33
|
Classifier: License :: OSI Approved :: MIT License
|
34
34
|
Classifier: Programming Language :: Python :: 3.8
|
35
35
|
Classifier: Topic :: Scientific/Engineering :: Artificial Intelligence
|
36
|
-
Requires-Python: >=3.
|
36
|
+
Requires-Python: >=3.9
|
37
37
|
Requires-Dist: adam-atan2-pytorch
|
38
38
|
Requires-Dist: assoc-scan
|
39
39
|
Requires-Dist: einops>=0.8.0
|
40
|
+
Requires-Dist: hl-gauss-pytorch>=0.1.19
|
40
41
|
Requires-Dist: torch>=2.2
|
41
42
|
Requires-Dist: tqdm
|
42
43
|
Provides-Extra: examples
|
43
44
|
Requires-Dist: numpy; extra == 'examples'
|
44
45
|
Requires-Dist: pufferlib>=2.0.6; extra == 'examples'
|
45
46
|
Requires-Dist: tqdm; extra == 'examples'
|
46
|
-
Provides-Extra: examples-gym
|
47
|
-
Requires-Dist: box2d-py; extra == 'examples-gym'
|
48
|
-
Requires-Dist: gymnasium[box2d]>=1.0.0; extra == 'examples-gym'
|
49
|
-
Requires-Dist: tqdm; extra == 'examples-gym'
|
50
47
|
Provides-Extra: test
|
51
48
|
Requires-Dist: pytest; extra == 'test'
|
52
49
|
Description-Content-Type: text/markdown
|
@@ -110,4 +107,15 @@ latent_pool.genetic_algorithm_step(fitness) # update latent genes with genetic a
|
|
110
107
|
}
|
111
108
|
```
|
112
109
|
|
110
|
+
```bibtex
|
111
|
+
@article{Farebrother2024StopRT,
|
112
|
+
title = {Stop Regressing: Training Value Functions via Classification for Scalable Deep RL},
|
113
|
+
author = {Jesse Farebrother and Jordi Orbay and Quan Ho Vuong and Adrien Ali Taiga and Yevgen Chebotar and Ted Xiao and Alex Irpan and Sergey Levine and Pablo Samuel Castro and Aleksandra Faust and Aviral Kumar and Rishabh Agarwal},
|
114
|
+
journal = {ArXiv},
|
115
|
+
year = {2024},
|
116
|
+
volume = {abs/2403.03950},
|
117
|
+
url = {https://api.semanticscholar.org/CorpusID:268253088}
|
118
|
+
}
|
119
|
+
```
|
120
|
+
|
113
121
|
*Evolution is cleverer than you are.* - Leslie Orgel
|
@@ -0,0 +1,7 @@
|
|
1
|
+
evolutionary_policy_optimization/__init__.py,sha256=Qavcia0n13jjaWIS_LPW7QrxSLT_BBeKujCjF9kQjbA,133
|
2
|
+
evolutionary_policy_optimization/epo.py,sha256=3pvYPwAEZdrxwwV95Ea1qG4CQjLnyaxAr40opk07LDw,17747
|
3
|
+
evolutionary_policy_optimization/experimental.py,sha256=ktBKxRF27Qsj7WIgBpYlWXqMVxO9zOx2oD1JuDYRAwM,548
|
4
|
+
evolutionary_policy_optimization-0.0.18.dist-info/METADATA,sha256=BIyCXw2IbMs-x2hDbFs9NR5s2dYEbfbeK_LadUeUc8Q,4860
|
5
|
+
evolutionary_policy_optimization-0.0.18.dist-info/WHEEL,sha256=qtCwoSJWgHk21S1Kb4ihdzI2rlJ1ZKaIurTj_ngOhyQ,87
|
6
|
+
evolutionary_policy_optimization-0.0.18.dist-info/licenses/LICENSE,sha256=1yCiA9b5nhslTavxPjsQAO-wpOnwJR9-l8LTVi7GJuk,1066
|
7
|
+
evolutionary_policy_optimization-0.0.18.dist-info/RECORD,,
|
@@ -1,7 +0,0 @@
|
|
1
|
-
evolutionary_policy_optimization/__init__.py,sha256=Qavcia0n13jjaWIS_LPW7QrxSLT_BBeKujCjF9kQjbA,133
|
2
|
-
evolutionary_policy_optimization/epo.py,sha256=f_e-TkJRFF1VHG3psJDgLGNIzlEvDSjX0nOsbLaOBrw,17543
|
3
|
-
evolutionary_policy_optimization/experimental.py,sha256=ktBKxRF27Qsj7WIgBpYlWXqMVxO9zOx2oD1JuDYRAwM,548
|
4
|
-
evolutionary_policy_optimization-0.0.16.dist-info/METADATA,sha256=Go0IYw3PbRLeHIu_Volhn31YX8WKLiI_YdllA3mj-bE,4480
|
5
|
-
evolutionary_policy_optimization-0.0.16.dist-info/WHEEL,sha256=qtCwoSJWgHk21S1Kb4ihdzI2rlJ1ZKaIurTj_ngOhyQ,87
|
6
|
-
evolutionary_policy_optimization-0.0.16.dist-info/licenses/LICENSE,sha256=1yCiA9b5nhslTavxPjsQAO-wpOnwJR9-l8LTVi7GJuk,1066
|
7
|
-
evolutionary_policy_optimization-0.0.16.dist-info/RECORD,,
|
File without changes
|