This repository is the official implementation of PatchPerPix for Instance Segmentation.
Lisa Mais1, Peter Hirsch1, Dagmar Kainmueller, ECCV2020 1Authors contributed equally, listed in random order
We present a novel method for proposal free instance segmentation that can handle sophisticated object shapes that span large parts of an image and form dense object clusters with crossovers. Our method is based on predicting dense local shape descriptors, which we assemble to form instances. All instances are assembled simultaneously in one go. To the best of our knowledge, our method is the first non-iterative method that yields instances that are composed of learnt shape patches. We evaluate our method on a diverse range of data domains, where it defines the new state of the art on four benchmarks, namely the ISBI 2012 EM segmentation benchmark, the BBBC010 C. elegans dataset, and 2d as well as 3d fluorescence microscopy datasets of cell nuclei. We show furthermore that our method also applies to 3d light microscopy data of drosophila neurons, which exhibit extreme cases of complex shape clusters.
This package requires Python 3 and PyTorch.
Note Previous versions (e.g., for the experiments published in our ECCV 2020 paper) require TensorFlow 1.x. If you want to run older experiments please checkout the respective tag: eccv2020 If you have any questions, please open an issue (and mention that you're running the older code)
The recommended way is to install the package into your conda/python virtual environment. We recommend to use conda to install torch (tested with torch 1.13, but newer versions should work, too). The following instructions were tested on linux/ubuntu 20.04.
conda create --name ppp --yes
conda activate ppp
conda install python=3.9 pytorch-cuda torchvision torchaudio cudatoolkit -c pytorch -c nvidia --yes
git clone https://github.com/Kainmueller-Lab/PatchPerPix.git
cd PatchPerPix
PATH=/usr/local/cuda/bin:$PATH CUDA_ROOT=/usr/local/cuda pip install -e .
run_ppp.py
:train.py
: trains the networkpredict_no_gp.py
: prediction after trainingdecode.py
: if ppp+dec is used, decode the predicted patch encodings to the full patchesdefault.toml
: example configuration filedefault_train_code.toml
: example configuration file that uses ppp+dectorch_loss.py
: auxiliary file for the loss computationtorch_model.py
: auxiliary file for the torch model definitionThe code expects the data to be in the zarr format (https://zarr.readthedocs.io/en/stable/). It is similar to hdf5, but uses the underlying file system to enable parallel read and write). It expects all used arrays (e.g., raw image data and labels) to be placed in a single zarr file (organized into a hierarchy via groups, see zarr documentation).
The names of the arrays have to be set in the config file (e.g., raw_key
and gt_key
) appropriately (example zarr file).
The main script run_ppp.py
(in the experiments folder) can be used to control all aspects of the experiments.
Example call:
python run_ppp.py --setup setup01 --config flylight/setups/setup01/default_train_code.toml --do train validate_checkpoints predict decode label evaluate --app flylight --root ppp_experiments
With --do TASK
you can set the sub-task that should be executed (or all
for the whole pipeline), --root PATH
sets the output directory, --app APP
the experiment (e.g. flylight) and --setup SETUP
the specific setup of that experiment (e.g. setup01).
The command above creates a time stamped experiment folder under the path specified by --root
.
To continue training or for further validation or evaluation adapt the command. Change the --config
parameter to point to the config file in the created experiment folder and remove the --root
flag and replace it with the -id
flag and point it to the created experiment folder. The tasks specified after --do
depend on what you want to do:
python run_ppp.py --setup setup01 --config ppp_experiments/flylight_setup01_230614__123456/config.toml --do validate_checkpoints predict decode label evaluate --app wormbodies -id experiments/flylight_setup01_230614__123456
Task | Short Description |
---|---|
all |
equal to mknet train validate_checkpoints predict decode label postprocess evaluate |
infer |
equal to predict decode label evaluate |
mknet |
creates a graph of the network (only for tensorflow 1) |
train |
executes the training of the network |
validate_checkpoints |
performs validation (over stored model checkpoints and a set of hyperparameters) |
validate |
performs validation (for a specific model checkpoint and over a set of hyperparameters) |
predict |
executes the trained network in inference mode and computes predictions |
decode |
decodes predicted patch encodings to full patches (only if model was trained to output encodings) |
label |
computes final instances based on predicted patches |
postprocess |
post-processes predictions and predicted instances (optional, mostly for manual inspection of results) |
evaluate |
compares predicted instances to ground truth instances and computes quantitative evaluation |
(for more details on the results see PatchPerPix for Instance Segmentation)
(BBBC010: C. elegans live/dead assay)
($S = \frac{TP}{TP+FP+FN}$; TP, FP, FN computed per image; averaged across images; localized using IoU)
Method | avS[0.5:0.9:0.1] | S0.5 | S0.6 | S0.7 | S0.8 | S0.9 |
---|---|---|---|---|---|---|
Inst.Seg via Layering[1] | 0.754 | 0.936 | 0.919 | 0.865 | 0.761 | 0.290 |
PatchPerPix (ppp+dec) | 0.816 | 0.960 | 0.955 | 0.931 | 0.805 | 0.428 |
[1] results from: Instance Segmentation of Dense and Overlapping Objects via Layering
(server with leaderboard is down, but data is still available: ISBI 2012 Segmentation Challenge
Method | rRAND | rINF |
---|---|---|
PatchPerPix | 0.988290 | 0.991544 |
MWS[2] | 0.987922 | 0.991833 |
MWS-Dense | 0.979112 | 0.989625 |
[2] results from leaderboard (offline, see also The Mutex Watershed: Efficient, Parameter-Free Image Partitioning)
(Kaggle 2018 Data Science Bowl, train/val/test split defined by Cell Detection with Star-convex Polygons)
($S = \frac{TP}{TP+FP+FN}$; TP, FP, FN computed per image; averaged across images; localized using IoU)
Method | avS[0.5:0.9:0.1] | S0.1 | S0.2 | S0.3 | S0.4 | S0.5 | S0.6 | S0.7 | S0.8 | S0.9 |
---|---|---|---|---|---|---|---|---|---|---|
Mask R-CNN[3] | 0.594 | - | - | - | - | 0.832 | 0.773 | 0.684 | 0.489 | 0.189 |
StarDist[3] | 0.584 | - | - | - | - | 0.864 | 0.804 | 0.685 | 0.450 | 0.119 |
PatchPerPix | 0.693 | 0.919 | 0.919 | 0.915 | 0.898 | 0.868 | 0.827 | 0.755 | 0.635 | 0.379 |
[3] results from Cell Detection with Star-convex Polygons
(https://doi.org/10.5281/zenodo.5942574, train/val/test split defined by Star-convex Polyhedra for 3D Object Detection and Segmentation in Microscopy)
($S = \frac{TP}{TP+FP+FN}$; TP, FP, FN computed per image; averaged across images; localized using IoU)
Method | avS[0.5:0.9:0.1] | S0.1 | S0.2 | S0.3 | S0.4 | S0.5 | S0.6 | S0.7 | S0.8 | S0.9 |
---|---|---|---|---|---|---|---|---|---|---|
MALA[4] | 0.381 | 0.895 | 0.887 | 0.859 | 0.803 | 0.699 | 0.605 | 0.424 | 0.166 | 0.012 |
StarDist3d[5] | 0.406 | 0.936 | 0.926 | 0.905 | 0.855 | 0.765 | 0.647 | 0.460 | 0.154 | 0.004 |
3-label+cpv[6] | 0.425 | 0.937 | 0.930 | 0.907 | 0.848 | 0.750 | 0.641 | 0.473 | 0.224 | 0.035 |
PatchPerPix | 0.436 | 0.926 | 0.918 | 0.900 | 0.853 | 0.766 | 0.668 | 0.493 | 0.228 | 0.027 |
[4] Large Scale Image Segmentation with Structured Loss based Deep Learning for Connectome Reconstruction, we computed the results
[5] results from Star-convex Polyhedra for 3D Object Detection and Segmentation in Microscopy
[6] results from An Auxiliary Task for Learning Nuclei Segmentation in 3D Microscopy Images
(The FlyLight Instance Segmentation Datset, train/val/test split defined by tba)
Metrik | short description |
---|---|
S | average of avF1 and C |
avF1 | Multi-Threshold F1 Score |
C | Average ground Truth coverage |
CTP | Average true positive coverage |
FS | Number of false splits |
FM | Number of false merges |
(for a precise definition see tba)
Trained on completely labeled data, evaluated on completely labeled data and partly labeled data combined: |
Method | S | avF1 | C | CTP | FS | FM |
---|---|---|---|---|---|---|---|
PatchPerPix | |||||||
Trained on completely labeled and partly labeled data combined, evaluated on completely labeled data and partly labeled data combined: | Method | S | avF1 | C | CTP | FS | FM |
---|---|---|---|---|---|---|---|
PatchPerPix(+partly) | |||||||
If you would like to contribute, have encountered any issues or have any suggestions, please open an issue on this GitHub repository.
All contributions are welcome! The content in this repository is licensed under the MIT license.