This repository is the main MineRL BASALT 2022 Competition submission template (only for the "basalt" track")!
You can find the "intro" track submission kit here.
MineRL BASALT is a competition on solving human-judged tasks. The tasks in this competition do not have a pre-defined reward function: the goal is to produce trajectories that are judged by real humans to be effective at solving a given task.
See the AICrowd competition page for further information.
This repository contains:
Other Resources:
minerl
package!In brief: you define your Python environment using Anaconda environment files, and AICrowd system will use the run.py
file to train and evaluate your agents.
You submit pretrained models, the evaluation code and the training code. Training code should produce the same models you upload as part of your submission. You must upload the training code; which will be used to retrain and validate finalist submissions.
Your evaluation code (test_<env_name>.py
) only needs to control the agent and accomplish the environment's task. The evaluation server will handle recording of videos. Do not change the number of episodes played or maximum number of steps executed: the submissions will fail otherwise.
Clone the github repository or press the "Use this Template" button on GitHub!
git clone https://github.com/minerllabs/basalt_2022_competition_submission_template.git
Install MineRL specific dependencies! Mainly, make sure you have Java JDK 8! See more details here
Specify your specific submission dependencies (PyTorch, Tensorflow, kittens, puppies, etc.)
Anaconda Environment. To make a submission you need to specify the environment using Anaconda environment files. It is also recommended you recreate the environment on your local machine. Make sure at least version 4.5.11
is required to correctly populate environment.yml
(By following instructions here). Then:
Create your new conda environment
cd basalt_2022_competition_submission_template
conda env create -f environment.yml
conda activate minerl
Your code specific dependencies
Add your own dependencies to the environment.yml
file. Remember to add any additional channels. PyTorch requires the channel pytorch
, for example.
You can also install them locally using
conda install <your-package>
Pip Packages If you need pip packages (not on conda), you can add them to the environment.yml
file (see the currently populated version):
Apt Packages If your training procedure or agent depends on specific Debian (Ubuntu, etc.) packages, add them to apt.txt
.
If above are too restrictive for defining your environment, see this Discourse topic for more information.
Please follow the example structure shared in the starter kit for the code structure. The different files and directories have the following meaning:
.
├── aicrowd.json # Submission meta information like your username
├── apt.txt # Packages to be installed inside docker image
├── config.py # Config for debugging submissions
├── data # DO NOT UPLOAD THE DATASET. This data will be available on the training instance.
│ ├── MineRLBasaltBuildVillageHouse-v0
│ │ ├── Player139-40dd711bfe2e-20220626-202049.jsonl
│ │ ├── Player139-40dd711bfe2e-20220626-202049.mp4
│ │ ├── Player139-62ae2702e774-20220626-201655.jsonl
│ │ ├── Player139-62ae2702e774-20220626-201655.mp4
│ │ └── ... rest of the files
│ ├── MineRLBasaltCreateVillageAnimalPen-v0
│ │ └── ... files as above
│ ├── MineRLBasaltFindCave-v0
│ │ └── ... files as above
│ ├── MineRLBasaltMakeWaterfall-v0
│ │ └── ... files as above
│ └── VPT-models
│ ├── foundation-model-1x.model
│ ├── foundation-model-1x.weights
│ ├── foundation-model-3x.model
│ ├── foundation-model-3x.weights
│ └── ... rest of the models
├── environment.yml # Conda environment description
├── LICENCE # Licence
├── run.py # The file that runs training and evaluation
├── test_BuildVillageHouse.py # IMPORTANT: Your testing/inference phase code.
├── test_CreateVillageAnimalPen.py # IMPORTANT: Your testing/inference phase code.
├── test_FindCave.py # IMPORTANT: Your testing/inference phase code.
├── test_MakeWaterfall.py # IMPORTANT: Your testing/inference phase code.
├── train # IMPORTANT: Your trained models MUST be saved inside this directory
└── train.py # IMPORTANT: Your training code. Running this should produce the same agent as you upload as part of the agent.
Finally, you must specify an AIcrowd submission JSON in aicrowd.json
to be scored!
The aicrowd.json
of each submission should contain the following content:
{
"challenge_id": "neurips-2022-minerl-basalt-competition",
"authors": ["your-aicrowd-username"],
"description": "sample description about your awesome agent",
"license": "MIT",
"gpu": true,
"debug": true,
"track": "basalt"
}
This JSON is used to map your submission to the said challenge, so please remember to use the correct challenge_id
as specified above.
By default, the debug
flag is set to true
. This makes evaluations run a single short episode. Please submit this way first to see if everything works fine on the AICrowd side. If it does, go ahead and submit with debug
set to false
.
Please specify if your code will use a GPU or not for the evaluation of your model.
You don't need to upload the provided BASALT dataset or pretrained models for the training stage, and it will be provided in online submissions at MINERL_DATA_ROOT
path (points to the data
directory), should you need them for training. Note that data
directory is not filled with data during initial evaluation phase.
Please only add files needed for the submission; too large submissions may fail randomly. Having large git history is fine.
See instructions here on how to download the BASALT dataset.
To make a submission, you will have to create a private repository on https://gitlab.aicrowd.com/.
You will have to add your SSH Keys to your GitLab account by following the instructions here. If you do not have SSH Keys, you will first need to generate one.
Then you can create a submission by making a tag push to your repository on https://gitlab.aicrowd.com/. Any tag push (where the tag name begins with "submission-") to your private repository is considered as a submission Then you can add the correct git remote, and finally submit by doing:
cd basalt_2022_competition_submission_template
# Add AIcrowd git remote endpoint
git remote add aicrowd git@gitlab.aicrowd.com:<YOUR_AICROWD_USER_NAME>/basalt_2022_competition_submission_template.git
git push aicrowd main
# Create a tag for your submission and push
git tag -am "submission-v0.1" submission-v0.1
git push aicrowd main
git push aicrowd submission-v0.1
# Note : If the contents of your repository (latest commit hash) do not change,
# then pushing a new tag will **not** trigger a new evaluation.
You now should be able to see the details of your submission at: https://gitlab.aicrowd.com/<YOUR_AICROWD_USER_NAME>/basalt_2022_competition_submission_template/issues/
Best of Luck :tada: :tada:
To upload large model files (e.g., your fine-tuned versions of the OpenAI VPT models, which can reach gigabytes), use git LFS. See instructions here.
You can perform local training and evaluation by simply running python run.py
, which then runs train.py
and all four test_<env_name>.py
scripts.
Note that you do not need to record videos in your code! AICrowd server will handle this. Your code only needs to play the games.
The quick-start kit was authored by Anssi Kanervisto, Karolis Ramanauskas and Shivam Khandelwal with help from William H. Guss
The BASALT competition is organized by the following team:
Advisors: