Closed soyeonm closed 3 years ago
@soyeonm
Right, there are 3 repeats because each task has 3 language annotations. It's the same task that the agent needs to solve, but the instructions for achieving those tasks were written by 3 separate annotators. For the leaderboard, if you miss any of them, they will be counted as failures. So I would recommend running on all of them.
No sorry, for the purpose a standardized and fair evaluation, we require all submissions to use the same action space. The leaderboard server won't recognize LookUp_30
and instead you have to call LookUp_15
twice. Keeping the action space consistent is necessary for a fair comparison of path-length weighted scores across models.
va_interact
is part of the ALFRED API for interacting with objects, and you can surely use it for all the leaderboard submissions. The rule is referring to other metadata info from the simulator like ground-truth object names, locations, poses, and other properties which won't be available to an agent in a realistic embodied setting.
Hope this helps! Good luck!
Hello,
I have a few questions about the submission format to the leaderboard.
(Your repository seems to interpret "LookUp_x", "LookDown_x" as LookUp/Down by "AGENT_HORIZON_ADJ" defined in gen/constants.py)
Are our models for leaderboard submissions allowed to use this "va_interact" function (using ground truth segmentation mask from AI2THOR only when deciding which ObjectID to interact)?
Thank you all the time!