gym-examples 3.0.182__py3-none-any.whl → 3.0.186__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
gym_examples/__init__.py CHANGED
@@ -6,4 +6,4 @@ register(
6
6
  max_episode_steps=50,
7
7
  )
8
8
 
9
- __version__ = "3.0.182"
9
+ __version__ = "3.0.186"
@@ -224,14 +224,19 @@ class WSNRoutingEnv(gym.Env):
224
224
  self.episode_network_lifetime.append(self.network_lifetime)
225
225
  self.episode_average_latency.append(self.average_latency)
226
226
 
227
- np.save(f"{base_back_up_dir}returns_QMIX_{self.version}.npy", np.array(self.episode_returns))
228
- np.save(f"{base_back_up_dir}std_remaining_energy_QMIX_{self.version}.npy", np.array(self.episode_std_remaining_energy))
229
- np.save(f"{base_back_up_dir}total_consumption_energy_QMIX_{self.version}.npy", np.array(self.episode_total_consumption_energy))
230
- np.save(f"{base_back_up_dir}mean_remaining_energy_QMIX_{self.version}.npy", np.array(self.episode_mean_remaining_energy))
231
- np.save(f"{base_back_up_dir}network_throughput_QMIX_{self.version}.npy", np.array(self.episode_network_throughput))
232
- np.save(f"{base_back_up_dir}packet_delivery_ratio_QMIX_{self.version}.npy", np.array(self.episode_packet_delivery_ratio))
233
- np.save(f"{base_back_up_dir}network_lifetime_QMIX_{self.version}.npy", np.array(self.episode_network_lifetime))
234
- np.save(f"{base_back_up_dir}average_latency_QMIX_{self.version}.npy", np.array(self.episode_average_latency))
227
+ metrics = {
228
+ "returns_PPO": self.episode_returns,
229
+ "std_remaining_energy_PPO": self.episode_std_remaining_energy,
230
+ "total_consumption_energy_PPO": self.episode_total_consumption_energy,
231
+ "mean_remaining_energy_PPO": self.episode_mean_remaining_energy,
232
+ "network_throughput_PPO": self.episode_network_throughput,
233
+ "packet_delivery_ratio_PPO": self.episode_packet_delivery_ratio,
234
+ "network_lifetime_PPO": self.episode_network_lifetime,
235
+ "average_latency_PPO": self.episode_average_latency
236
+ }
237
+
238
+ for metric_name, metric_value in metrics.items():
239
+ np.save(f"{base_back_up_dir}{metric_name}_{self.version}.npy", np.array(metric_value))
235
240
 
236
241
  return self._get_obs(), rewards, dones, {}
237
242
 
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.1
2
2
  Name: gym-examples
3
- Version: 3.0.182
3
+ Version: 3.0.186
4
4
  Summary: A custom environment for multi-agent reinforcement learning focused on WSN routing.
5
5
  Home-page: https://github.com/gedji/CODES.git
6
6
  Author: Georges Djimefo
@@ -0,0 +1,7 @@
1
+ gym_examples/__init__.py,sha256=U7dXZkWTF-yBDlPgEBQRIhkCuXyc-Dr80Rzz9D92bKQ,194
2
+ gym_examples/envs/__init__.py,sha256=lgMe4pyOuUTgTBUddM0iwMlETsYTwFShny6ifm8PGM8,53
3
+ gym_examples/envs/wsn_env.py,sha256=aYprA2YqcxEH11AFhQKZGpzX4azYgWWn2u1lMitidQ4,25701
4
+ gym_examples-3.0.186.dist-info/METADATA,sha256=icwRFP8ftYaNL4DXfq3HQaiP0mipmnSK7zwdoW4wsZc,412
5
+ gym_examples-3.0.186.dist-info/WHEEL,sha256=2wepM1nk4DS4eFpYrW1TTqPcoGNfHhhO_i5m4cOimbo,92
6
+ gym_examples-3.0.186.dist-info/top_level.txt,sha256=rJRksoAF32M6lTLBEwYzRdo4PgtejceaNnnZ3HeY_Rk,13
7
+ gym_examples-3.0.186.dist-info/RECORD,,
@@ -1,7 +0,0 @@
1
- gym_examples/__init__.py,sha256=RiS3AduI_PATwPjqyEMvdXeVB7NUbBrYUWBJchVfwuw,194
2
- gym_examples/envs/__init__.py,sha256=lgMe4pyOuUTgTBUddM0iwMlETsYTwFShny6ifm8PGM8,53
3
- gym_examples/envs/wsn_env.py,sha256=vJ94e6nzg1iSxSyfEYffTjhN7szbIz1ZaqU4tG_U0_Y,25901
4
- gym_examples-3.0.182.dist-info/METADATA,sha256=R7BDOPdQSi9IwPOQGfq9a8uqNBl0hYvwMHjJD2l-Jo8,412
5
- gym_examples-3.0.182.dist-info/WHEEL,sha256=2wepM1nk4DS4eFpYrW1TTqPcoGNfHhhO_i5m4cOimbo,92
6
- gym_examples-3.0.182.dist-info/top_level.txt,sha256=rJRksoAF32M6lTLBEwYzRdo4PgtejceaNnnZ3HeY_Rk,13
7
- gym_examples-3.0.182.dist-info/RECORD,,