Preliminaries: These instructions are specific to XSEDE Bridges but other resources can be used if cuda, python3, and a NVIDIA P100 GPU are present, in which case 'module load' instructions can be skipped, which are specific to Bridges.
For Unix or Mac Users:
Login to bridges via ssh using a Unix or Mac command line terminal. Login is available to bridges directly or through the XSEDE portal. Please see the Bridges User's Guide.
For Windows Users:
Many tools are available for ssh access to bridges. Please see Ubuntu, MobaXterm or PuTTY
Once you have logged into Bridges, you can follow one of two methods for installing iceberg-penguins.
The lines below following a '$' are commands to enter (or cut and paste) into your terminal (note that all commands are case-sensitive, meaning capital and lowercase letters are differentiated.) Everything following '#' are comments to explain the reason for the command and should not be included in what you enter. Lines that do not start with '$' or '[penguins_env] $' are output you should expect to see.
$ pwd
/home/username
$ cd $SCRATCH # switch to your working space.
$ mkdir Penguins # create a directory to work in.
$ cd Penguins # move into your working directory.
$ module load cuda # load parallel computing architecture.
$ module load python3 # load correct python version.
$ virtualenv penguins_env # create a virtual environment to isolate your work from the default system.
$ source penguins_env/bin/activate # activate your environment. Notice the command line prompt changes to show your environment on the next line.
[penguins_env] $ pwd
/pylon5/group/username/Penguins
[penguins_env] $ export PYTHONPATH=<path>/penguins_env/lib/python3.5/site-packages # set a system variable to point python to your specific code. (Replace <path> with the results of pwd command above.
[penguins_env] $ pip install iceberg_penguins.search # pip is a python tool to extract the requested software (iceberg_penguins.search in this case) from a repository. (this may take several minutes).
$ git clone https://github.com/iceberg-project/Penguins.git
$ module load cuda
$ module load python3
$ virtualenv penguins_env
$ source penguins_env/bin/activate
[penguins_env] $ export PYTHONPATH=<path>/penguins_env/lib/python3.5/site-packages
[penguins_env] $ pip install . --upgrade
To test
[iceberg_penguins] $ deactivate # exit your virtual environment.
$ interact -p GPU-small # request a compute node. This package has been tested on P100 GPUs on bridges, but that does not exclude any other resource that offers the same GPUs. (this may take a minute or two or more to receive an allocation).
$ cd $SCRATCH/Penguins # make sure you are in the same directory where everything was set up before.
$ module load cuda # load parallel computing architecture, as before.
$ module load python3 # load correct python version, as before.
$ source penguins_env/bin/activate # activate your environment, no need to create a new environment because the Penguins tools are installed and isolated here.
[iceberg_penguins] $ iceberg_penguins.detect --help # this will display a help screen of available usage and parameters.
You can download to your local machine and use scp, ftp, rsync, or Globus to transfer to bridges.
The one provided here is at the epoch 300 of the model we will call "MY_MODEL".
Please put the model file here:
Then, follow the environment setup commands under 'To test' above. Finally, the script to run the prediction for a single PNG image tile is:
iceberg_penguins.detect [--params ...]
iceberg_penguins.detect --gpu-ids 0 --name MY_MODEL --epoch 300 --checkpoints_dir '../model_path/' --output test --input_im ../data/MY_IMG_TILE.png