gym-examples 3.0.35__py3-none-any.whl → 3.0.37__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- gym_examples/__init__.py +1 -1
- gym_examples/envs/wsn_env.py +8 -8
- {gym_examples-3.0.35.dist-info → gym_examples-3.0.37.dist-info}/METADATA +1 -1
- gym_examples-3.0.37.dist-info/RECORD +7 -0
- gym_examples-3.0.35.dist-info/RECORD +0 -7
- {gym_examples-3.0.35.dist-info → gym_examples-3.0.37.dist-info}/WHEEL +0 -0
- {gym_examples-3.0.35.dist-info → gym_examples-3.0.37.dist-info}/top_level.txt +0 -0
gym_examples/__init__.py
CHANGED
gym_examples/envs/wsn_env.py
CHANGED
@@ -146,6 +146,9 @@ class WSNRoutingEnv(gym.Env):
|
|
146
146
|
|
147
147
|
# Compute individual rewards
|
148
148
|
rewards[i] = self.compute_individual_rewards(i, action)
|
149
|
+
reward_packet_delivery_ratio = self.compute_reward_packet_delivery_ratio()
|
150
|
+
reward_latency = self.compute_reward_latency()
|
151
|
+
rewards[i] = np.append(rewards[i], [reward_packet_delivery_ratio, reward_latency])
|
149
152
|
# Update the number of packets
|
150
153
|
self.number_of_packets[action] += self.number_of_packets[i]
|
151
154
|
self.number_of_packets[i] = 0 # Reset the number of packets of the sensor i
|
@@ -165,14 +168,11 @@ class WSNRoutingEnv(gym.Env):
|
|
165
168
|
|
166
169
|
self.get_metrics()
|
167
170
|
|
168
|
-
reward_packet_delivery_ratio = self.compute_reward_packet_delivery_ratio()
|
169
|
-
reward_latency = self.compute_reward_latency()
|
170
|
-
rewards_metrics = [reward_packet_delivery_ratio, reward_latency]
|
171
|
-
rewards_metrics = self.compute_attention_rewards(rewards_metrics)
|
172
|
-
|
173
|
-
print(f"rewards: {rewards}, {rewards_metrics}")
|
174
|
-
raise ValueError("Stop")
|
175
|
-
rewards = np.array([r.detach().numpy() + rewards_metrics.detach().numpy() for r in rewards])
|
171
|
+
# reward_packet_delivery_ratio = self.compute_reward_packet_delivery_ratio()
|
172
|
+
# reward_latency = self.compute_reward_latency()
|
173
|
+
# rewards_metrics = [reward_packet_delivery_ratio, reward_latency]
|
174
|
+
# rewards_metrics = self.compute_attention_rewards(rewards_metrics)
|
175
|
+
# rewards = [torch.tensor(r, dtype=torch.float64) + rewards_metrics if isinstance(r, int) else r + rewards_metrics for r in rewards]
|
176
176
|
|
177
177
|
return self._get_obs(), rewards, dones, {}
|
178
178
|
|
@@ -0,0 +1,7 @@
|
|
1
|
+
gym_examples/__init__.py,sha256=LGX8WDZE4Bx_0vg5ndopkifzz7-4aW1m32F9g2-zc8s,193
|
2
|
+
gym_examples/envs/__init__.py,sha256=lgMe4pyOuUTgTBUddM0iwMlETsYTwFShny6ifm8PGM8,53
|
3
|
+
gym_examples/envs/wsn_env.py,sha256=X_Fb6iQ-LtXjDxDjdnuj9EyuAOKGxDQWHCX8LID0yNU,19238
|
4
|
+
gym_examples-3.0.37.dist-info/METADATA,sha256=EQgb3067i71u7IJTPEXN7KE4xcdb0EDCVgh35pj8N0U,411
|
5
|
+
gym_examples-3.0.37.dist-info/WHEEL,sha256=2wepM1nk4DS4eFpYrW1TTqPcoGNfHhhO_i5m4cOimbo,92
|
6
|
+
gym_examples-3.0.37.dist-info/top_level.txt,sha256=rJRksoAF32M6lTLBEwYzRdo4PgtejceaNnnZ3HeY_Rk,13
|
7
|
+
gym_examples-3.0.37.dist-info/RECORD,,
|
@@ -1,7 +0,0 @@
|
|
1
|
-
gym_examples/__init__.py,sha256=TUJSLkoJXyceTTwrQCXncmSEoUIfb3SPdA3bSvNFtYc,193
|
2
|
-
gym_examples/envs/__init__.py,sha256=lgMe4pyOuUTgTBUddM0iwMlETsYTwFShny6ifm8PGM8,53
|
3
|
-
gym_examples/envs/wsn_env.py,sha256=ENxEXTO436RvIMezy5HzvSEMFIjSbjIXU-tylfhcyns,19092
|
4
|
-
gym_examples-3.0.35.dist-info/METADATA,sha256=luDLCfgU0Hqa9gLk-HHOIqZVEaUlwzuBYsZ8iWXMbj4,411
|
5
|
-
gym_examples-3.0.35.dist-info/WHEEL,sha256=2wepM1nk4DS4eFpYrW1TTqPcoGNfHhhO_i5m4cOimbo,92
|
6
|
-
gym_examples-3.0.35.dist-info/top_level.txt,sha256=rJRksoAF32M6lTLBEwYzRdo4PgtejceaNnnZ3HeY_Rk,13
|
7
|
-
gym_examples-3.0.35.dist-info/RECORD,,
|
File without changes
|
File without changes
|