diff --git a/examples/training_example.py b/examples/training_example.py
index cdbd1ade672dee5478ca817ccad52aa48f992445..dd9ded92510be8ec5fa6c222b7259157db920430 100644
--- a/examples/training_example.py
+++ b/examples/training_example.py
@@ -76,8 +76,6 @@ for trials in range(1, n_trials + 1):
         for a in range(env.get_num_agents()):
             action = agent.act(obs[a])
             action_dict.update({a: action})
-            # Uncomment next line to print observation of an agent
-            # TreeObservation.util_print_obs_subtree((obs[a]))
 
         # Environment step which returns the observations for all agents, their corresponding
         # reward and whether their are done
diff --git a/notebooks/Scene_Editor.ipynb b/notebooks/Scene_Editor.ipynb
index a36b4e159a75edd2d18a5f7ed42ec41dbc5b9515..64bbad77d3a9b63caea062795cd0ab8f0cce72f3 100644
--- a/notebooks/Scene_Editor.ipynb
+++ b/notebooks/Scene_Editor.ipynb
@@ -70,7 +70,7 @@
     {
      "data": {
       "application/vnd.jupyter.widget-view+json": {
-       "model_id": "4153480c3fbe4fd9864a4d52b416477c",
+       "model_id": "0b92e7084e37450cbfb3855dc3a58543",
        "version_major": 2,
        "version_minor": 0
       },