gym-examples 3.0.253__py3-none-any.whl → 3.0.255__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
gym_examples/__init__.py CHANGED
@@ -5,4 +5,4 @@ register(
5
5
  entry_point="gym_examples.envs:WSNRoutingEnv",
6
6
  )
7
7
 
8
- __version__ = "3.0.253"
8
+ __version__ = "3.0.255"
@@ -88,15 +88,15 @@ class WSNRoutingEnv(gym.Env):
88
88
 
89
89
  def reset(self):
90
90
 
91
- if self.episode_count > 1 and os.getenv('PRINT_STATS') == 'True': # Statistics for the PPO algorithm during the training phase
92
- self.episode_returns.append(self.episode_return)
93
- self.episode_std_remaining_energy.append(np.std(self.remaining_energy))
94
- self.episode_mean_remaining_energy.append(np.mean(self.remaining_energy))
95
- self.episode_total_consumption_energy.append(np.sum(initial_energy - self.remaining_energy))
96
- self.episode_network_throughput.append(self.network_throughput)
97
- self.episode_packet_delivery_ratio.append(self.packet_delivery_ratio)
98
- self.episode_network_lifetime.append(self.network_lifetime)
99
- self.episode_average_latency.append(self.average_latency)
91
+ # if self.episode_count > 1 and os.getenv('PRINT_STATS') == 'True': # Statistics for the PPO algorithm during the training phase
92
+ # self.episode_returns.append(self.episode_return)
93
+ # self.episode_std_remaining_energy.append(np.std(self.remaining_energy))
94
+ # self.episode_mean_remaining_energy.append(np.mean(self.remaining_energy))
95
+ # self.episode_total_consumption_energy.append(np.sum(initial_energy - self.remaining_energy))
96
+ # self.episode_network_throughput.append(self.network_throughput)
97
+ # self.episode_packet_delivery_ratio.append(self.packet_delivery_ratio)
98
+ # self.episode_network_lifetime.append(self.network_lifetime)
99
+ # self.episode_average_latency.append(self.average_latency)
100
100
 
101
101
  # if self.episode_count > 1 and os.getenv('PRINT_STATS') == 'True-False': # Statistics for the PPO algorithm during the evaluation phase
102
102
  # print(f"Episode: {self.episode_count}")
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.1
2
2
  Name: gym-examples
3
- Version: 3.0.253
3
+ Version: 3.0.255
4
4
  Summary: A custom environment for multi-agent reinforcement learning focused on WSN routing.
5
5
  Home-page: https://github.com/gedji/CODES.git
6
6
  Author: Georges Djimefo
@@ -0,0 +1,7 @@
1
+ gym_examples/__init__.py,sha256=AWbW41IrAPwo2iJJSOYFUpGIhYeRMEF_j8MuxDKRdAE,166
2
+ gym_examples/envs/__init__.py,sha256=lgMe4pyOuUTgTBUddM0iwMlETsYTwFShny6ifm8PGM8,53
3
+ gym_examples/envs/wsn_env.py,sha256=HdGwa7pL3eVjHlkTnCMMF3LnM9tQzog-d5Ur41G80Ic,26448
4
+ gym_examples-3.0.255.dist-info/METADATA,sha256=Rik7EMZLHXOn8FMVeTDbseA4Cb6ArVxkameg76JHSGQ,412
5
+ gym_examples-3.0.255.dist-info/WHEEL,sha256=2wepM1nk4DS4eFpYrW1TTqPcoGNfHhhO_i5m4cOimbo,92
6
+ gym_examples-3.0.255.dist-info/top_level.txt,sha256=rJRksoAF32M6lTLBEwYzRdo4PgtejceaNnnZ3HeY_Rk,13
7
+ gym_examples-3.0.255.dist-info/RECORD,,
@@ -1,7 +0,0 @@
1
- gym_examples/__init__.py,sha256=X5GE915JrSF9FGS1atrA3RLqrF7xzZQX3qsHXbdMtL0,166
2
- gym_examples/envs/__init__.py,sha256=lgMe4pyOuUTgTBUddM0iwMlETsYTwFShny6ifm8PGM8,53
3
- gym_examples/envs/wsn_env.py,sha256=AbShz4cMLDOnMzWEFl1leAQA18LNlhqppqIS2pK4Edk,26430
4
- gym_examples-3.0.253.dist-info/METADATA,sha256=440oy3t2cxQwnQDSdVHzGD4VcxxztEjOeLxiMCqzQc0,412
5
- gym_examples-3.0.253.dist-info/WHEEL,sha256=2wepM1nk4DS4eFpYrW1TTqPcoGNfHhhO_i5m4cOimbo,92
6
- gym_examples-3.0.253.dist-info/top_level.txt,sha256=rJRksoAF32M6lTLBEwYzRdo4PgtejceaNnnZ3HeY_Rk,13
7
- gym_examples-3.0.253.dist-info/RECORD,,