evolutionary-policy-optimization 0.0.41__tar.gz → 0.0.43__tar.gz
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- {evolutionary_policy_optimization-0.0.41 → evolutionary_policy_optimization-0.0.43}/PKG-INFO +1 -1
- {evolutionary_policy_optimization-0.0.41 → evolutionary_policy_optimization-0.0.43}/evolutionary_policy_optimization/epo.py +11 -4
- {evolutionary_policy_optimization-0.0.41 → evolutionary_policy_optimization-0.0.43}/evolutionary_policy_optimization/mock_env.py +1 -1
- {evolutionary_policy_optimization-0.0.41 → evolutionary_policy_optimization-0.0.43}/pyproject.toml +1 -1
- {evolutionary_policy_optimization-0.0.41 → evolutionary_policy_optimization-0.0.43}/tests/test_epo.py +8 -2
- {evolutionary_policy_optimization-0.0.41 → evolutionary_policy_optimization-0.0.43}/.github/workflows/python-publish.yml +0 -0
- {evolutionary_policy_optimization-0.0.41 → evolutionary_policy_optimization-0.0.43}/.github/workflows/test.yml +0 -0
- {evolutionary_policy_optimization-0.0.41 → evolutionary_policy_optimization-0.0.43}/.gitignore +0 -0
- {evolutionary_policy_optimization-0.0.41 → evolutionary_policy_optimization-0.0.43}/LICENSE +0 -0
- {evolutionary_policy_optimization-0.0.41 → evolutionary_policy_optimization-0.0.43}/README.md +0 -0
- {evolutionary_policy_optimization-0.0.41 → evolutionary_policy_optimization-0.0.43}/evolutionary_policy_optimization/__init__.py +0 -0
- {evolutionary_policy_optimization-0.0.41 → evolutionary_policy_optimization-0.0.43}/evolutionary_policy_optimization/experimental.py +0 -0
- {evolutionary_policy_optimization-0.0.41 → evolutionary_policy_optimization-0.0.43}/requirements.txt +0 -0
{evolutionary_policy_optimization-0.0.41 → evolutionary_policy_optimization-0.0.43}/PKG-INFO
RENAMED
@@ -1,6 +1,6 @@
|
|
1
1
|
Metadata-Version: 2.4
|
2
2
|
Name: evolutionary-policy-optimization
|
3
|
-
Version: 0.0.
|
3
|
+
Version: 0.0.43
|
4
4
|
Summary: EPO - Pytorch
|
5
5
|
Project-URL: Homepage, https://pypi.org/project/evolutionary-policy-optimization/
|
6
6
|
Project-URL: Repository, https://github.com/lucidrains/evolutionary-policy-optimization
|
@@ -875,25 +875,32 @@ def create_agent(
|
|
875
875
|
actor_num_actions,
|
876
876
|
actor_dim_hiddens: int | tuple[int, ...],
|
877
877
|
critic_dim_hiddens: int | tuple[int, ...],
|
878
|
+
latent_gene_pool_kwargs: dict = dict(),
|
879
|
+
actor_kwargs: dict = dict(),
|
880
|
+
critic_kwargs: dict = dict(),
|
878
881
|
) -> Agent:
|
879
882
|
|
880
883
|
latent_gene_pool = LatentGenePool(
|
881
884
|
num_latents = num_latents,
|
882
|
-
dim_latent = dim_latent
|
885
|
+
dim_latent = dim_latent,
|
886
|
+
**latent_gene_pool_kwargs
|
883
887
|
)
|
884
888
|
|
885
889
|
actor = Actor(
|
886
890
|
num_actions = actor_num_actions,
|
887
891
|
dim_state = dim_state,
|
888
892
|
dim_latent = dim_latent,
|
889
|
-
dim_hiddens = actor_dim_hiddens
|
893
|
+
dim_hiddens = actor_dim_hiddens,
|
894
|
+
**actor_kwargs
|
890
895
|
)
|
891
896
|
|
892
897
|
critic = Critic(
|
893
898
|
dim_state = dim_state,
|
894
899
|
dim_latent = dim_latent,
|
895
|
-
dim_hiddens = critic_dim_hiddens
|
896
|
-
|
900
|
+
dim_hiddens = critic_dim_hiddens,
|
901
|
+
**critic_kwargs
|
902
|
+
)
|
903
|
+
|
897
904
|
return Agent(actor = actor, critic = critic, latent_gene_pool = latent_gene_pool)
|
898
905
|
|
899
906
|
# EPO - which is just PPO with natural selection of a population of latent variables conditioning the agent
|
@@ -73,7 +73,10 @@ def test_create_agent(
|
|
73
73
|
agent.save('./agent.pt', overwrite = True)
|
74
74
|
agent.load('./agent.pt')
|
75
75
|
|
76
|
-
|
76
|
+
@pytest.mark.parametrize('frozen_latents', (False, True))
|
77
|
+
def test_e2e_with_mock_env(
|
78
|
+
frozen_latents
|
79
|
+
):
|
77
80
|
from evolutionary_policy_optimization import create_agent, EPO, Env
|
78
81
|
|
79
82
|
agent = create_agent(
|
@@ -82,7 +85,10 @@ def test_e2e_with_mock_env():
|
|
82
85
|
dim_latent = 32,
|
83
86
|
actor_num_actions = 5,
|
84
87
|
actor_dim_hiddens = (256, 128),
|
85
|
-
critic_dim_hiddens = (256, 128, 64)
|
88
|
+
critic_dim_hiddens = (256, 128, 64),
|
89
|
+
latent_gene_pool_kwargs = dict(
|
90
|
+
frozen_latents = frozen_latents
|
91
|
+
)
|
86
92
|
)
|
87
93
|
|
88
94
|
epo = EPO(
|
File without changes
|
File without changes
|
{evolutionary_policy_optimization-0.0.41 → evolutionary_policy_optimization-0.0.43}/.gitignore
RENAMED
File without changes
|
File without changes
|
{evolutionary_policy_optimization-0.0.41 → evolutionary_policy_optimization-0.0.43}/README.md
RENAMED
File without changes
|
File without changes
|
File without changes
|
{evolutionary_policy_optimization-0.0.41 → evolutionary_policy_optimization-0.0.43}/requirements.txt
RENAMED
File without changes
|