evolutionary-policy-optimization 0.0.41__py3-none-any.whl → 0.0.43__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -875,25 +875,32 @@ def create_agent(
875
875
  actor_num_actions,
876
876
  actor_dim_hiddens: int | tuple[int, ...],
877
877
  critic_dim_hiddens: int | tuple[int, ...],
878
+ latent_gene_pool_kwargs: dict = dict(),
879
+ actor_kwargs: dict = dict(),
880
+ critic_kwargs: dict = dict(),
878
881
  ) -> Agent:
879
882
 
880
883
  latent_gene_pool = LatentGenePool(
881
884
  num_latents = num_latents,
882
- dim_latent = dim_latent
885
+ dim_latent = dim_latent,
886
+ **latent_gene_pool_kwargs
883
887
  )
884
888
 
885
889
  actor = Actor(
886
890
  num_actions = actor_num_actions,
887
891
  dim_state = dim_state,
888
892
  dim_latent = dim_latent,
889
- dim_hiddens = actor_dim_hiddens
893
+ dim_hiddens = actor_dim_hiddens,
894
+ **actor_kwargs
890
895
  )
891
896
 
892
897
  critic = Critic(
893
898
  dim_state = dim_state,
894
899
  dim_latent = dim_latent,
895
- dim_hiddens = critic_dim_hiddens
896
- )
900
+ dim_hiddens = critic_dim_hiddens,
901
+ **critic_kwargs
902
+ )
903
+
897
904
  return Agent(actor = actor, critic = critic, latent_gene_pool = latent_gene_pool)
898
905
 
899
906
  # EPO - which is just PPO with natural selection of a population of latent variables conditioning the agent
@@ -7,7 +7,7 @@ from torch.nn import Module
7
7
  # functions
8
8
 
9
9
  def cast_tuple(v):
10
- return v if isinstance(v, tuple) else v\
10
+ return v if isinstance(v, tuple) else (v,)
11
11
 
12
12
  # mock env
13
13
 
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.4
2
2
  Name: evolutionary-policy-optimization
3
- Version: 0.0.41
3
+ Version: 0.0.43
4
4
  Summary: EPO - Pytorch
5
5
  Project-URL: Homepage, https://pypi.org/project/evolutionary-policy-optimization/
6
6
  Project-URL: Repository, https://github.com/lucidrains/evolutionary-policy-optimization
@@ -0,0 +1,8 @@
1
+ evolutionary_policy_optimization/__init__.py,sha256=0q0aBuFgWi06MLMD8FiHzBYQ3_W4LYWrwmCtF3u5H2A,201
2
+ evolutionary_policy_optimization/epo.py,sha256=Yf-iw1gqmAUEVzg6_PwYy-q4005eroZKUYGxNgwCsKk,30440
3
+ evolutionary_policy_optimization/experimental.py,sha256=9FrJGviLESlYysHI3i83efT9g2ZB9ha4u3K9HXN98_w,1100
4
+ evolutionary_policy_optimization/mock_env.py,sha256=6AIc4mwL_C6JkAxwESJgCLxXHMzCAu2FcffVg3HkSm0,920
5
+ evolutionary_policy_optimization-0.0.43.dist-info/METADATA,sha256=pMVLppijepjmI1A9wVVhdX2IXo4BNPsOozpMAAsS6Lo,6213
6
+ evolutionary_policy_optimization-0.0.43.dist-info/WHEEL,sha256=qtCwoSJWgHk21S1Kb4ihdzI2rlJ1ZKaIurTj_ngOhyQ,87
7
+ evolutionary_policy_optimization-0.0.43.dist-info/licenses/LICENSE,sha256=1yCiA9b5nhslTavxPjsQAO-wpOnwJR9-l8LTVi7GJuk,1066
8
+ evolutionary_policy_optimization-0.0.43.dist-info/RECORD,,
@@ -1,8 +0,0 @@
1
- evolutionary_policy_optimization/__init__.py,sha256=0q0aBuFgWi06MLMD8FiHzBYQ3_W4LYWrwmCtF3u5H2A,201
2
- evolutionary_policy_optimization/epo.py,sha256=GL3nH5crOj4y_Amu2BY0s95MJL7F2t-X085y40SgUK0,30260
3
- evolutionary_policy_optimization/experimental.py,sha256=9FrJGviLESlYysHI3i83efT9g2ZB9ha4u3K9HXN98_w,1100
4
- evolutionary_policy_optimization/mock_env.py,sha256=QqVPZVJtrvQmSDcnYDTob_A5sDwiUzGj6_tmo6BII5c,918
5
- evolutionary_policy_optimization-0.0.41.dist-info/METADATA,sha256=TFKI2B2PeyU6pHwqmCu130k-U2Li_QmUkvVB39-4uDw,6213
6
- evolutionary_policy_optimization-0.0.41.dist-info/WHEEL,sha256=qtCwoSJWgHk21S1Kb4ihdzI2rlJ1ZKaIurTj_ngOhyQ,87
7
- evolutionary_policy_optimization-0.0.41.dist-info/licenses/LICENSE,sha256=1yCiA9b5nhslTavxPjsQAO-wpOnwJR9-l8LTVi7GJuk,1066
8
- evolutionary_policy_optimization-0.0.41.dist-info/RECORD,,