diff --git a/examples/introduction_flatland_2_1.py b/examples/introduction_flatland_2_1.py index 42917fb79e8d548f0ab29afb43c6de72ebb7d714..607bbf7405c4603643df9d1d685513a3a2d33a24 100644 --- a/examples/introduction_flatland_2_1.py +++ b/examples/introduction_flatland_2_1.py @@ -241,7 +241,7 @@ score = 0 # Run episode frame_step = 0 -for step in range(10): +for step in range(100): # Chose an action for each agent in the environment for a in range(env.get_num_agents()): action = controller.act(observations[a]) @@ -252,7 +252,7 @@ for step in range(10): next_obs, all_rewards, done, _ = env.step(action_dict) - # env_renderer.render_env(show=True, show_observations=False, show_predictions=False) + env_renderer.render_env(show=True, show_observations=False, show_predictions=False) frame_step += 1 # Update replay buffer and train agent for a in range(env.get_num_agents()):