Skip to content
GitLab
Explore
Sign in
Primary navigation
Search or go to…
Project
Flatland
Manage
Activity
Members
Labels
Plan
Issues
Issue boards
Milestones
Wiki
Code
Merge requests
Repository
Branches
Commits
Tags
Repository graph
Compare revisions
Snippets
Build
Pipelines
Jobs
Pipeline schedules
Artifacts
Deploy
Releases
Model registry
Operate
Environments
Monitor
Incidents
Analyze
Value stream analytics
Contributor analytics
CI/CD analytics
Repository analytics
Model experiments
Help
Help
Support
GitLab documentation
Compare GitLab plans
Community forum
Contribute to GitLab
Provide feedback
Keyboard shortcuts
?
Snippets
Groups
Projects
Show more breadcrumbs
pranjal_dhole
Flatland
Commits
f9547092
Commit
f9547092
authored
5 years ago
by
Erik Nygren
Browse files
Options
Downloads
Patches
Plain Diff
fixed bug where agent can enter and move at the same time
parent
098e4652
No related branches found
No related tags found
No related merge requests found
Changes
2
Hide whitespace changes
Inline
Side-by-side
Showing
2 changed files
examples/introduction_flatland_2_1_1.py
+37
-3
37 additions, 3 deletions
examples/introduction_flatland_2_1_1.py
flatland/envs/rail_env.py
+2
-0
2 additions, 0 deletions
flatland/envs/rail_env.py
with
39 additions
and
3 deletions
examples/introduction_flatland_2_1_1.py
+
37
−
3
View file @
f9547092
...
@@ -132,7 +132,7 @@ controller = RandomAgent(218, env.action_space[0])
...
@@ -132,7 +132,7 @@ controller = RandomAgent(218, env.action_space[0])
# We start by looking at the information of each agent
# We start by looking at the information of each agent
# We can see the task assigned to the agent by looking at
# We can see the task assigned to the agent by looking at
print
(
"
Agents in the environment have to solve the following tasks:
\n
"
)
print
(
"
\n
Agents in the environment have to solve the following tasks:
\n
"
)
for
agent_idx
,
agent
in
enumerate
(
env
.
agents
):
for
agent_idx
,
agent
in
enumerate
(
env
.
agents
):
print
(
print
(
"
The agent with index {} has the task to go from its initial position {}, facing in the direction {} to its target at {}.
"
.
format
(
"
The agent with index {} has the task to go from its initial position {}, facing in the direction {} to its target at {}.
"
.
format
(
...
@@ -140,7 +140,9 @@ for agent_idx, agent in enumerate(env.agents):
...
@@ -140,7 +140,9 @@ for agent_idx, agent in enumerate(env.agents):
# The agent will always have a status indicating if it is currently present in the environment or done or active
# The agent will always have a status indicating if it is currently present in the environment or done or active
# For example we see that agent with index 0 is currently not active
# For example we see that agent with index 0 is currently not active
print
(
"
Their current statuses are:
\n
"
)
print
(
"
\n
Their current statuses are:
"
)
print
(
"
============================
"
)
for
agent_idx
,
agent
in
enumerate
(
env
.
agents
):
for
agent_idx
,
agent
in
enumerate
(
env
.
agents
):
print
(
"
Agent {} status is: {} with its current position being {}
"
.
format
(
agent_idx
,
str
(
agent
.
status
),
print
(
"
Agent {} status is: {} with its current position being {}
"
.
format
(
agent_idx
,
str
(
agent
.
status
),
str
(
agent
.
position
)))
str
(
agent
.
position
)))
...
@@ -149,7 +151,39 @@ for agent_idx, agent in enumerate(env.agents):
...
@@ -149,7 +151,39 @@ for agent_idx, agent in enumerate(env.agents):
# If the starting cell is free they will enter the level
# If the starting cell is free they will enter the level
# If multiple agents want to enter the same cell at the same time the lower index agent will enter first.
# If multiple agents want to enter the same cell at the same time the lower index agent will enter first.
# Let's check if there are any agents with the same start location
agents_with_same_start
=
[]
print
(
"
\n
The following agents have the same initial position:
"
)
print
(
"
============================
"
)
for
agent_idx
,
agent
in
enumerate
(
env
.
agents
):
for
agent_2_idx
,
agent2
in
enumerate
(
env
.
agents
):
if
agent_idx
!=
agent_2_idx
and
agent
.
initial_position
==
agent2
.
initial_position
:
print
(
"
Agent {} as the same initial position as agent {}
"
.
format
(
agent_idx
,
agent_2_idx
))
agents_with_same_start
.
append
(
agent_idx
)
# Lets try to enter with all of these agents at the same time
action_dict
=
{}
for
agent_id
in
agents_with_same_start
:
action_dict
[
agent_id
]
=
1
# Set agents to moving
print
(
"
\n
This happened when all tried to enter at the same time:
"
)
print
(
"
========================================================
"
)
for
agent_id
in
agents_with_same_start
:
print
(
"
Agent {} status is: {} with its current position being {}
"
.
format
(
agent_id
,
str
(
env
.
agents
[
agent_id
].
status
),
str
(
env
.
agents
[
agent_id
].
position
)))
# Do a step in the environment to see what agents entered:
env
.
step
(
action_dict
)
# Current state and position of the agents after all agents with same start position tried to move
print
(
"
\n
This happened when all tried to enter at the same time:
"
)
print
(
"
========================================================
"
)
for
agent_id
in
agents_with_same_start
:
print
(
"
Agent {} status is: {} with its current position being {} which is the same as the start position {} and orientation {}
"
.
format
(
agent_id
,
str
(
env
.
agents
[
agent_id
].
status
),
str
(
env
.
agents
[
agent_id
].
position
),
env
.
agents
[
agent_id
].
initial_position
,
env
.
agents
[
agent_id
].
direction
))
# Empty dictionary for all agent action
# Empty dictionary for all agent action
action_dict
=
dict
()
action_dict
=
dict
()
...
...
This diff is collapsed.
Click to expand it.
flatland/envs/rail_env.py
+
2
−
0
View file @
f9547092
...
@@ -458,6 +458,8 @@ class RailEnv(Environment):
...
@@ -458,6 +458,8 @@ class RailEnv(Environment):
RailEnvActions
.
MOVE_FORWARD
]
and
self
.
cell_free
(
agent
.
initial_position
):
RailEnvActions
.
MOVE_FORWARD
]
and
self
.
cell_free
(
agent
.
initial_position
):
agent
.
status
=
RailAgentStatus
.
ACTIVE
agent
.
status
=
RailAgentStatus
.
ACTIVE
agent
.
position
=
agent
.
initial_position
agent
.
position
=
agent
.
initial_position
self
.
rewards_dict
[
i_agent
]
+=
self
.
step_penalty
*
agent
.
speed_data
[
'
speed
'
]
return
else
:
else
:
# TODO: Here we need to check for the departure time in future releases with full schedules
# TODO: Here we need to check for the departure time in future releases with full schedules
self
.
rewards_dict
[
i_agent
]
+=
self
.
step_penalty
*
agent
.
speed_data
[
'
speed
'
]
self
.
rewards_dict
[
i_agent
]
+=
self
.
step_penalty
*
agent
.
speed_data
[
'
speed
'
]
...
...
This diff is collapsed.
Click to expand it.
Preview
0%
Loading
Try again
or
attach a new file
.
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Save comment
Cancel
Please
register
or
sign in
to comment