Welcome to the GOSEEK challenge page, which is run in conjunction with the Perception, Action, Learning Workshop at ICRA 2020.
For this competition, participants create a reinforcement learning (RL) agent that combines perception and high-level decision-making to search for objects placed within complex indoor environments from a Unity-based simulator. Simply put: like PACMAN, but in a realistic scene and with realistic perception capabilities. Several data modalities are provided from both the simulator ground truth and a perception pipeline (e.g., images, depth, agent location) to enable the participants to focus on the RL/search aspects. The contest will be hosted on the EvalAI platform, where participants will submit solutions, via docker containers run on AWS instances, for scoring.
Outline
The objective of this challenge is to navigate an agent through an office environment to collect randomly-spawned fruit as quickly as possible. Our teaser trailer (below) highlights several of the components of the challenge, such as the office environment, the target fruit, the perception pipeline, and our idealized robot's physical characteristics.
More specifically, the agent can select from one of four actions at each decision epoch: move forward 0.5 meters, turn left 8 degrees, turn right 8 degrees, and collect fruit within 2.0 meters of the agent's current position. Our robot is equipped with stereo cameras and an Inertial Measurement Unit (IMU), from which a state-of-the-art perception pipeline estimates three pieces of information that make up the agent's observation at each decision epoch: localization information (position and heading relative to start position), pixel-wise semantic labels for objects in the robot's field of view, and pixel-wise depth in the robot's field of view.
We provide two data sources for training:
Participants can use either or both of these sources for training their agents. Agent interfaces are identical between the two sources. We'll accept online submissions against either source (see below for details) and maintain a leaderboard for both. However, only evaluations against the Perception Pipeline will be used to declare an overall competition winner.
Agents are evaluated on the following criteria for each episode:
r
: recall of finding target fruit when the agents selects the collect action,p
: precision of finding target fruit when the agent selects the collect action,c
: number of collisions with objects in the scene, anda
: actions taken in the episode before all target fruit are collected or time expires.A single episode score is:
r + 0.1p - 0.1c/l - 0.1a/l
where l
is the maximum episode length (400). Note that an episode terminates early if all fruit are collected.
We use Monte Carlo evaluations to estimate an average episode score for the competition. Note that evaluations occur on withheld office scenes.
The current timeline for the competition is as follows:
Over the course of the competition, any important announcements or updates will be listed in this section. We recommend that you follow this repository to be alerted to these announcements.
tesse-gym
from the instructions.
Please also rebuild the goseek-base
docker image, as well as any of your submission images.Complete installation instructions can be found here, which lays out prerequisites, provides a link to download the competition simulator, and describes steps to install all required competition software. Users can also find an example for training an RL agent here, as well.
Participants will upload docker containers with their agents to EvalAI in order to be evaluated for the competition. The number of submissions is limited for each user, so we highly recommend performing local evaluations prior to submitting online solutions. This sections describes how to evaluate your agent locally, then submit online for a score.
Before proceeding, we recommend that you have read through and completed these instructions.
Modify Dockerfile
as appropriate for your agent.
See these instructions for modification details.
The example we've provided runs an agent that randomly selects actions at each step.
Build the docker image. Here we are naming the image submission
.
docker build -t submission .
Use test_locally.py
for local testing.
Assume you've named your docker image submission
as above, then evaluate your agent with Ground Truth data as follows.
python test_locally.py -s simulator/goseek-v0.1.4.x86_64 -i submission -g
Similarly, evaluate your agent with Perception Pipeline data as follows.
python test_locally.py -s simulator/goseek-v0.1.4.x86_64 -i submission -p
Install EvalAI-CLI: pip install evalai
.
Create on account on EvalAI's website and sign up for the GOSEEK-Challenge.
Follow the instructions on the submission tab to push your docker image. Note that we've provided four phases -- some to support development. Only the leader of the Competition Phase with Perception Pipeline will be declared the competition winner.
If you participate in GOSEEK and write a paper or a report about your entry, please cite:
First, we would like to thank Rishabh Jain and the rest of the team at EvalAI for providing their infrastructure and personal time to support this challenge. We must also acknowledge the team behind The Habitat challenge for being pathfinders of RL challenges with online submissions. Their challenge and associated infrastructure was inspiration for many of our own decisions.
DISTRIBUTION STATEMENT A. Approved for public release. Distribution is unlimited.
This material is based upon work supported by the Under Secretary of Defense for Research and Engineering under Air Force Contract No. FA8702-15-D-0001. Any opinions, findings, conclusions or recommendations expressed in this material are those of the author(s) and do not necessarily reflect the views of the Under Secretary of Defense for Research and Engineering.
(c) 2020 Massachusetts Institute of Technology.
MIT Proprietary, Subject to FAR52.227-11 Patent Rights - Ownership by the contractor (May 2014)
The software/firmware is provided to you on an As-Is basis
Delivered to the U.S. Government with Unlimited Rights, as defined in DFARS Part 252.227-7013 or 7014 (Feb 2014). Notwithstanding any copyright notice, U.S. Government rights in this work are defined by DFARS 252.227-7013 or DFARS 252.227-7014 as detailed above. Use of this work other than as specifically authorized by the U.S. Government may violate any copyrights that exist in this work.