diff --git a/examples/flatland_2_0_example.py b/examples/flatland_2_0_example.py
index 0d86561f0b2d3317f88dc25ef8d8fb824ade74d1..8106585c098e54d275016b33ff6df4617d517e43 100644
--- a/examples/flatland_2_0_example.py
+++ b/examples/flatland_2_0_example.py
@@ -113,7 +113,7 @@ for step in range(500):
     # reward and whether their are done
     next_obs, all_rewards, done, _ = env.step(action_dict)
     env_renderer.render_env(show=True, show_observations=False, show_predictions=False)
-    time.sleep(50)
+    time.sleep(1)
     frame_step += 1
     # Update replay buffer and train agent
     for a in range(env.get_num_agents()):