From 813c9d2c9147cccb5205ed13b30905813a33d537 Mon Sep 17 00:00:00 2001 From: romue Date: Thu, 17 Jun 2021 16:26:07 +0200 Subject: [PATCH] added own dqn --- algorithms/_base.py | 3 +-- 1 file changed, 1 insertion(+), 2 deletions(-) diff --git a/algorithms/_base.py b/algorithms/_base.py index e8639fe..dd3e458 100644 --- a/algorithms/_base.py +++ b/algorithms/_base.py @@ -183,7 +183,7 @@ class BaseQlearner: target_q = experience.reward + (1 - experience.done) * self.gamma * target_q_raw loss = torch.mean(self.reg_weight * pred_q + torch.pow(pred_q - target_q, 2)) - print(target_q) + #print(target_q) # log loss self.running_loss.append(loss.item()) @@ -210,7 +210,6 @@ if __name__ == '__main__': allow_no_op=False) env = SimpleFactory(dirt_properties=dirt_props, movement_properties=move_props, n_agents=N_AGENTS, pomdp_radius=2, max_steps=400, omit_agent_slice_in_obs=False) # env = DummyVecEnv([lambda: env]) - print(env) from stable_baselines3.dqn import DQN #dqn = RegDQN('MlpPolicy', env, verbose=True, buffer_size = 50000, learning_starts = 64, batch_size = 64,