From 44d3ca9db858e68c695d698a66a5f1b5fac3e415 Mon Sep 17 00:00:00 2001 From: MLErik <baerenjesus@gmail.com> Date: Mon, 24 Jun 2019 10:01:47 -0500 Subject: [PATCH] updated training file --- torch_training/training_navigation.py | 6 +++++- 1 file changed, 5 insertions(+), 1 deletion(-) diff --git a/torch_training/training_navigation.py b/torch_training/training_navigation.py index 634db69..de4b792 100644 --- a/torch_training/training_navigation.py +++ b/torch_training/training_navigation.py @@ -146,6 +146,9 @@ for trials in range(1, n_trials + 1): distance = norm_obs_clip(distance) agent_data = np.clip(agent_data, -1, 1) obs[a] = np.concatenate((np.concatenate((data, distance)), agent_data)) + agent_data = env.agents[a] + speed = 1 #np.random.randint(1,5) + agent_data.speed_data['speed'] = 1. / speed for i in range(2): time_obs.append(obs) @@ -153,6 +156,7 @@ for trials in range(1, n_trials + 1): for a in range(env.get_num_agents()): agent_obs[a] = np.concatenate((time_obs[0][a], time_obs[1][a])) + score = 0 env_done = 0 # Run episode @@ -167,7 +171,7 @@ for trials in range(1, n_trials + 1): if demo: eps = 0 # action = agent.act(np.array(obs[a]), eps=eps) - action = 2 #agent.act(agent_obs[a], eps=eps) + action = agent.act(agent_obs[a], eps=eps) action_prob[action] += 1 action_dict.update({a: action}) # Environment step -- GitLab