From cd27006ad0d4aee5a8c408bb5cc18973e3069948 Mon Sep 17 00:00:00 2001
From: Erik Nygren <erik.nygren@sbb.ch>
Date: Thu, 18 Jul 2019 16:00:16 -0400
Subject: [PATCH] added some fun hand drawn levels

---
 torch_training/multi_agent_inference.py  |   8 ++++----
 torch_training/railway/hard_crossing.pkl | Bin 0 -> 1169 bytes
 torch_training/railway/simple_avoid.pkl  | Bin 0 -> 849 bytes
 torch_training/railway/testing_stuff.pkl | Bin 0 -> 1226 bytes
 4 files changed, 4 insertions(+), 4 deletions(-)
 create mode 100644 torch_training/railway/hard_crossing.pkl
 create mode 100644 torch_training/railway/simple_avoid.pkl
 create mode 100644 torch_training/railway/testing_stuff.pkl

diff --git a/torch_training/multi_agent_inference.py b/torch_training/multi_agent_inference.py
index e399126..e8fd6d4 100644
--- a/torch_training/multi_agent_inference.py
+++ b/torch_training/multi_agent_inference.py
@@ -16,7 +16,7 @@ from utils.observation_utils import normalize_observation
 
 random.seed(3)
 np.random.seed(2)
-"""
+
 file_name = "./railway/complex_scene.pkl"
 env = RailEnv(width=10,
               height=20,
@@ -41,7 +41,7 @@ env = RailEnv(width=x_dim,
               obs_builder_object=TreeObsForRailEnv(max_depth=3, predictor=ShortestPathPredictorForRailEnv()),
               number_of_agents=n_agents)
 env.reset(True, True)
-
+"""
 tree_depth = 3
 observation_helper = TreeObsForRailEnv(max_depth=tree_depth, predictor=ShortestPathPredictorForRailEnv())
 env_renderer = RenderTool(env, gl="PILSVG", )
@@ -53,7 +53,7 @@ for i in range(tree_depth + 1):
 state_size = num_features_per_node * nr_nodes
 action_size = 5
 
-n_trials = 10
+n_trials = 1
 observation_radius = 10
 max_steps = int(3 * (env.height + env.width))
 eps = 1.
@@ -88,7 +88,7 @@ for trials in range(1, n_trials + 1):
 
     # Run episode
     for step in range(max_steps):
-        env_renderer.render_env(show=True, show_observations=False, show_predictions=False)
+        env_renderer.render_env(show=True, show_observations=False, show_predictions=True)
 
         if record_images:
             env_renderer.gl.save_image("./Images/Avoiding/flatland_frame_{:04d}.bmp".format(frame_step))
diff --git a/torch_training/railway/hard_crossing.pkl b/torch_training/railway/hard_crossing.pkl
new file mode 100644
index 0000000000000000000000000000000000000000..9a877c065dd852529868dd5fa9e92c4e3326d511
GIT binary patch
literal 1169
zcmb``KMsO06bA6uKZApVI=JCgba2p95Q3F7F=(KJI~On{EY9lY-~l{=;RYVUBlw_2
zjE3zIhNk_#Um=w26eW69JYgbB$;uIQUFD2d0R~M}F57}z#PzP*!Y<bz{!q48gLB~w
z8|w0^`8tn-ab5q;w@>-uFxMOWzTCp?Wd)41#s9D?HysAoAX2eO{nP|ThsPoZ5V`aq
zyY05AHhLTT>m&%P!K2yvobHrbJ*F=^aW_d2yR1$6*b3EVqxRZBkuxL(Ey<DOw4^|i
o(UKBLpe4a$xdhtJgQTD(Ig*@~6i6~!QX+{*nC_@w?9X&~1Ed~qz5oCK

literal 0
HcmV?d00001

diff --git a/torch_training/railway/simple_avoid.pkl b/torch_training/railway/simple_avoid.pkl
new file mode 100644
index 0000000000000000000000000000000000000000..8aed98cae65efeef035daf7caded46f979884845
GIT binary patch
literal 849
zcmZo!l3tXVGV2b52m=;?5I*a`be4q?iiF@wkrXs207cJ&Nw73fiUpwsu9rs;NzfCC
z4KtjPfdOGE+#(PIw=m2iu)`2eg}D<d!U#11;z%$FcQ2TM>|`XTA)JgPh~^=f*(eSM
zan>fLr{<Lu#}}6*mSiSRojQq`fq4?M5JU6Eg8bsllFa<P__U(LWDw&tO!Ly>g4EQM
z)Ak=A0((n}67!&HAu8j6x{_0Ka#AZYOBjGg3NZqWrK*t}!i<wRsB0t}(Bo{v)O97$
zNDgWmxeO9p6V|3=0wXLhIW;~vv7os12vc6l;Ug?1l?ADbZPF}`uw-ZEr5s^6!dje>
USdcn}flGwt2un(0N#YR(0GVI$^8f$<

literal 0
HcmV?d00001

diff --git a/torch_training/railway/testing_stuff.pkl b/torch_training/railway/testing_stuff.pkl
new file mode 100644
index 0000000000000000000000000000000000000000..ec94aae2913a576950e0b3f7dcfd7ae73f1d199f
GIT binary patch
literal 1226
zcmb`FPfEi;6vk(gNfmUVE-dciNp#_&^%5bZu>?w5l3BPE>H#EB5ZCI?g$M8iWp3ah
zJc8ewH<MK8!o`FT=Fj(i@An>Azs{R#_VO(*lu?r2GqLyCQK+f8iJh7~|NkO=H%Xeq
z6CuoXG{i}i9}XOic}U4eWG1B#_Eyz{9H07#sbAQ3fQAy@%eGsmAdgp{%RfzF%P;&O
zvdn$0$4eJL@bm!361l~MuGwWzxIHh19s*kz-N9$&gJWX4ltTAU!HWkyrTJEQ!LtBA
zSLwNN+e5gC*g}0me@n+7`Mj)ko3%RERq?#(gorye7VFQ|vaNKrth2i&FKDsl7$M%;
zRawrqS3i94rJKBVT2^Lo73E@4-d9>|I*G_6Y~*MxK^l}Y0_8zi$^?}7Lqg6{Au<XZ
W87t96M?onQ%yJZRmhb$VKK%mk;BrO)

literal 0
HcmV?d00001

-- 
GitLab