diff --git a/examples/training_example.py b/examples/training_example.py index cdbd1ade672dee5478ca817ccad52aa48f992445..dd9ded92510be8ec5fa6c222b7259157db920430 100644 --- a/examples/training_example.py +++ b/examples/training_example.py @@ -76,8 +76,6 @@ for trials in range(1, n_trials + 1): for a in range(env.get_num_agents()): action = agent.act(obs[a]) action_dict.update({a: action}) - # Uncomment next line to print observation of an agent - # TreeObservation.util_print_obs_subtree((obs[a])) # Environment step which returns the observations for all agents, their corresponding # reward and whether their are done diff --git a/notebooks/Scene_Editor.ipynb b/notebooks/Scene_Editor.ipynb index a36b4e159a75edd2d18a5f7ed42ec41dbc5b9515..64bbad77d3a9b63caea062795cd0ab8f0cce72f3 100644 --- a/notebooks/Scene_Editor.ipynb +++ b/notebooks/Scene_Editor.ipynb @@ -70,7 +70,7 @@ { "data": { "application/vnd.jupyter.widget-view+json": { - "model_id": "4153480c3fbe4fd9864a4d52b416477c", + "model_id": "0b92e7084e37450cbfb3855dc3a58543", "version_major": 2, "version_minor": 0 },