Spaces:
Running
Running
Added logging of rewards
Browse files- scripts/simulate.py +13 -6
scripts/simulate.py
CHANGED
@@ -265,13 +265,20 @@ def simulate_game(game_name: str, config: Dict[str, Any], seed: int) -> str:
|
|
265 |
opponents_list.append(opponent_str)
|
266 |
opponents = ", ".join(opponents_list)
|
267 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
268 |
agent_logger.log_game_result(
|
269 |
-
|
270 |
-
|
271 |
-
|
272 |
-
|
273 |
-
|
274 |
-
|
275 |
# Tensorboard logging
|
276 |
agent_type = "unknown"
|
277 |
agent_model = "None"
|
|
|
265 |
opponents_list.append(opponent_str)
|
266 |
opponents = ", ".join(opponents_list)
|
267 |
|
268 |
+
# Log reward to the rewards table
|
269 |
+
agent_logger.log_rewards(
|
270 |
+
game_name=game_name,
|
271 |
+
episode=episode + 1,
|
272 |
+
reward=reward
|
273 |
+
)
|
274 |
+
|
275 |
agent_logger.log_game_result(
|
276 |
+
game_name=game_name,
|
277 |
+
episode=episode + 1,
|
278 |
+
status=game_status,
|
279 |
+
reward=reward,
|
280 |
+
opponent=opponents
|
281 |
+
)
|
282 |
# Tensorboard logging
|
283 |
agent_type = "unknown"
|
284 |
agent_model = "None"
|