Skip to content
GitLab
Explore
Sign in
Primary navigation
Search or go to…
Project
M
marl-flatland
Manage
Activity
Members
Labels
Plan
Issues
Issue boards
Milestones
Wiki
Code
Merge requests
Repository
Branches
Commits
Tags
Repository graph
Compare revisions
Snippets
Deploy
Releases
Package Registry
Model registry
Operate
Terraform modules
Monitor
Incidents
Analyze
Value stream analytics
Contributor analytics
Repository analytics
Model experiments
Help
Help
Support
GitLab documentation
Compare GitLab plans
Community forum
Contribute to GitLab
Provide feedback
Keyboard shortcuts
?
Snippets
Groups
Projects
Show more breadcrumbs
manavsinghal157
marl-flatland
Commits
78d1c9ff
Commit
78d1c9ff
authored
4 years ago
by
Egli Adrian (IT-SCI-API-PFI)
Browse files
Options
Downloads
Patches
Plain Diff
DeadLockAvoidanceAgent
parent
bc9dca31
No related branches found
Branches containing commit
No related tags found
No related merge requests found
Changes
3
Hide whitespace changes
Inline
Side-by-side
Showing
3 changed files
run.py
+11
-2
11 additions, 2 deletions
run.py
utils/dead_lock_avoidance_agent.py
+1
-1
1 addition, 1 deletion
utils/dead_lock_avoidance_agent.py
utils/shortest_distance_walker.py
+87
-0
87 additions, 0 deletions
utils/shortest_distance_walker.py
with
99 additions
and
3 deletions
run.py
+
11
−
2
View file @
78d1c9ff
...
@@ -29,6 +29,7 @@ checkpoint = "./checkpoints/201103160541-1800.pth"
...
@@ -29,6 +29,7 @@ checkpoint = "./checkpoints/201103160541-1800.pth"
# Use last action cache
# Use last action cache
USE_ACTION_CACHE
=
True
USE_ACTION_CACHE
=
True
USE_DEAD_LOCK_AVOIDANCE_AGENT
=
False
# Observation parameters (must match training parameters!)
# Observation parameters (must match training parameters!)
observation_tree_depth
=
2
observation_tree_depth
=
2
...
@@ -82,8 +83,6 @@ while True:
...
@@ -82,8 +83,6 @@ while True:
nb_agents
=
len
(
local_env
.
agents
)
nb_agents
=
len
(
local_env
.
agents
)
max_nb_steps
=
local_env
.
_max_episode_steps
max_nb_steps
=
local_env
.
_max_episode_steps
policy
=
DeadLockAvoidanceAgent
(
local_env
)
tree_observation
.
set_env
(
local_env
)
tree_observation
.
set_env
(
local_env
)
tree_observation
.
reset
()
tree_observation
.
reset
()
observation
=
tree_observation
.
get_many
(
list
(
range
(
nb_agents
)))
observation
=
tree_observation
.
get_many
(
list
(
range
(
nb_agents
)))
...
@@ -105,6 +104,9 @@ while True:
...
@@ -105,6 +104,9 @@ while True:
agent_last_action
=
{}
agent_last_action
=
{}
nb_hit
=
0
nb_hit
=
0
if
USE_DEAD_LOCK_AVOIDANCE_AGENT
:
policy
=
DeadLockAvoidanceAgent
(
local_env
)
while
True
:
while
True
:
try
:
try
:
#####################################################################
#####################################################################
...
@@ -118,7 +120,14 @@ while True:
...
@@ -118,7 +120,14 @@ while True:
time_start
=
time
.
time
()
time_start
=
time
.
time
()
action_dict
=
{}
action_dict
=
{}
policy
.
start_step
()
policy
.
start_step
()
if
USE_DEAD_LOCK_AVOIDANCE_AGENT
:
observation
=
np
.
zeros
((
local_env
.
get_num_agents
(),
2
))
for
agent
in
range
(
nb_agents
):
for
agent
in
range
(
nb_agents
):
if
USE_DEAD_LOCK_AVOIDANCE_AGENT
:
observation
[
agent
][
0
]
=
agent
observation
[
agent
][
1
]
=
steps
if
info
[
'
action_required
'
][
agent
]:
if
info
[
'
action_required
'
][
agent
]:
if
agent
in
agent_last_obs
and
np
.
all
(
agent_last_obs
[
agent
]
==
observation
[
agent
]):
if
agent
in
agent_last_obs
and
np
.
all
(
agent_last_obs
[
agent
]
==
observation
[
agent
]):
# cache hit
# cache hit
...
...
This diff is collapsed.
Click to expand it.
utils/dead_lock_avoidance_agent.py
+
1
−
1
View file @
78d1c9ff
...
@@ -7,7 +7,7 @@ from flatland.envs.agent_utils import RailAgentStatus
...
@@ -7,7 +7,7 @@ from flatland.envs.agent_utils import RailAgentStatus
from
flatland.envs.rail_env
import
RailEnv
,
RailEnvActions
,
fast_count_nonzero
from
flatland.envs.rail_env
import
RailEnv
,
RailEnvActions
,
fast_count_nonzero
from
reinforcement_learning.policy
import
Policy
from
reinforcement_learning.policy
import
Policy
from
utils.shortest_
D
istance_walker
import
ShortestDistanceWalker
from
utils.shortest_
d
istance_walker
import
ShortestDistanceWalker
class
DeadlockAvoidanceObservation
(
DummyObservationBuilder
):
class
DeadlockAvoidanceObservation
(
DummyObservationBuilder
):
...
...
This diff is collapsed.
Click to expand it.
utils/shortest_distance_walker.py
0 → 100644
+
87
−
0
View file @
78d1c9ff
import
numpy
as
np
from
flatland.core.grid.grid4_utils
import
get_new_position
from
flatland.envs.rail_env
import
RailEnv
,
RailEnvActions
from
flatland.envs.rail_env
import
fast_count_nonzero
,
fast_argmax
class
ShortestDistanceWalker
:
def
__init__
(
self
,
env
:
RailEnv
):
self
.
env
=
env
def
walk
(
self
,
handle
,
position
,
direction
):
possible_transitions
=
self
.
env
.
rail
.
get_transitions
(
*
position
,
direction
)
num_transitions
=
fast_count_nonzero
(
possible_transitions
)
if
num_transitions
==
1
:
new_direction
=
fast_argmax
(
possible_transitions
)
new_position
=
get_new_position
(
position
,
new_direction
)
dist
=
self
.
env
.
distance_map
.
get
()[
handle
,
new_position
[
0
],
new_position
[
1
],
new_direction
]
return
new_position
,
new_direction
,
dist
,
RailEnvActions
.
MOVE_FORWARD
,
possible_transitions
else
:
min_distances
=
[]
positions
=
[]
directions
=
[]
for
new_direction
in
[(
direction
+
i
)
%
4
for
i
in
range
(
-
1
,
2
)]:
if
possible_transitions
[
new_direction
]:
new_position
=
get_new_position
(
position
,
new_direction
)
min_distances
.
append
(
self
.
env
.
distance_map
.
get
()[
handle
,
new_position
[
0
],
new_position
[
1
],
new_direction
])
positions
.
append
(
new_position
)
directions
.
append
(
new_direction
)
else
:
min_distances
.
append
(
np
.
inf
)
positions
.
append
(
None
)
directions
.
append
(
None
)
a
=
self
.
get_action
(
handle
,
min_distances
)
return
positions
[
a
],
directions
[
a
],
min_distances
[
a
],
a
+
1
,
possible_transitions
def
get_action
(
self
,
handle
,
min_distances
):
return
np
.
argmin
(
min_distances
)
def
callback
(
self
,
handle
,
agent
,
position
,
direction
,
action
,
possible_transitions
):
pass
def
get_agent_position_and_direction
(
self
,
handle
):
agent
=
self
.
env
.
agents
[
handle
]
if
agent
.
position
is
not
None
:
position
=
agent
.
position
else
:
position
=
agent
.
initial_position
direction
=
agent
.
direction
return
position
,
direction
def
walk_to_target
(
self
,
handle
,
position
=
None
,
direction
=
None
,
max_step
=
500
):
if
position
is
None
and
direction
is
None
:
position
,
direction
=
self
.
get_agent_position_and_direction
(
handle
)
elif
position
is
None
:
position
,
_
=
self
.
get_agent_position_and_direction
(
handle
)
elif
direction
is
None
:
_
,
direction
=
self
.
get_agent_position_and_direction
(
handle
)
agent
=
self
.
env
.
agents
[
handle
]
step
=
0
while
(
position
!=
agent
.
target
)
and
(
step
<
max_step
):
position
,
direction
,
dist
,
action
,
possible_transitions
=
self
.
walk
(
handle
,
position
,
direction
)
if
position
is
None
:
break
self
.
callback
(
handle
,
agent
,
position
,
direction
,
action
,
possible_transitions
)
step
+=
1
def
callback_one_step
(
self
,
handle
,
agent
,
position
,
direction
,
action
,
possible_transitions
):
pass
def
walk_one_step
(
self
,
handle
):
agent
=
self
.
env
.
agents
[
handle
]
if
agent
.
position
is
not
None
:
position
=
agent
.
position
else
:
position
=
agent
.
initial_position
direction
=
agent
.
direction
possible_transitions
=
(
0
,
1
,
0
,
0
)
if
(
position
!=
agent
.
target
):
new_position
,
new_direction
,
dist
,
action
,
possible_transitions
=
self
.
walk
(
handle
,
position
,
direction
)
if
new_position
is
None
:
return
position
,
direction
,
RailEnvActions
.
STOP_MOVING
,
possible_transitions
self
.
callback_one_step
(
handle
,
agent
,
new_position
,
new_direction
,
action
,
possible_transitions
)
return
new_position
,
new_direction
,
action
,
possible_transitions
This diff is collapsed.
Click to expand it.
Preview
0%
Loading
Try again
or
attach a new file
.
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Save comment
Cancel
Please
register
or
sign in
to comment