diff --git a/examples/flatland_2_0_example.py b/examples/flatland_2_0_example.py index 0d86561f0b2d3317f88dc25ef8d8fb824ade74d1..8106585c098e54d275016b33ff6df4617d517e43 100644 --- a/examples/flatland_2_0_example.py +++ b/examples/flatland_2_0_example.py @@ -113,7 +113,7 @@ for step in range(500): # reward and whether their are done next_obs, all_rewards, done, _ = env.step(action_dict) env_renderer.render_env(show=True, show_observations=False, show_predictions=False) - time.sleep(50) + time.sleep(1) frame_step += 1 # Update replay buffer and train agent for a in range(env.get_num_agents()):