WSI Superpixel Guided Labeling is a Girder 3 <https://github.com/girder>
plugin designed to be used in conjunction with HistomicsUI <https://github.com/DigitalSlideArchive/HistomicsUI>
and HistomicsTK <https://github.com/DigitalSlideArchive/HistomicsTK>
_ to facilitate active learning on whole slide images.
This plugin leverages the output of certain HistomicsTK/SlicerCLI jobs to allow end users to label superpixel regions of whole slide images to be used as input for machine learning algorithms.
An example algorithm is contained within the dsarchive/superpixel:latest
docker image. This can be used to generate superpixels, features, and machine learning models for active learning on a directory of images. See the installation instructions below for how to include the image as part of your Digital Slide Archive deployment.
Once the appropriate data is generated, a new view becomes available for labeling and retraining.
The recommended way to use this plugin is by adding it to the Digital Slide Archive's docker-compose
deployment. First, check out both this repository and digital_slide_archive
from Github, if you do not yet have a running instance of the Digital Slide Archive.
If you don't already use a provisioning yaml file as part of your DSA deployment, you'll want to create one, e.g. provision.local.yaml
. Make sure this file contains the following: ::
pip:
- /opt/wsi-superpixel-guided-labeling
rebuild-client: True
resources:
- model: collection
name: Tasks
creator: resource:admin
public: True
- model: folder
parent: resource:collection/Tasks
parentType: collection
name: "Slicer CLI Web Tasks"
creator: resource:admin
public: True
- model: collection
name: "Active Learning"
creator: resource:admin
public: True
- model: folder
parent: "resource:collection/Active Learning"
parentType: collection
name: Data
creator: resource:admin
public: True
metadata: {'active_learning': "true"}
metadata_update: True
- model: folder
parent: "resource:collection/Active Learning/Data"
parentType: folder
name: Annotations
creator: resource:admin
public: True
- model: folder
parent: "resource:collection/Active Learning/Data"
parentType: folder
name: Models
creator: resource:admin
public: True
- model: folder
parent: "resource:collection/Active Learning/Data"
parentType: folder
name: Features
creator: resource:admin
public: True
slicer-cli-image-pull:
- dsarchive/histomicstk:latest
- dsarchive/superpixel:latest
In digital_slide_archive/devops/dsa/
, you'll want to add or modify docker-compose.override.yaml
in the following manner: ::
---
version: '3'
services:
girder:
volumes:
- ./provision.local.yaml:/opt/digital_slide_archive/devops/dsa/provision.yaml
- <path>/<to>/wsi-superpixel-guided-labeling:/opt/wsi-superpixel-guided-labeling
Where <path>/<to>/wsi-superpixel-guided-labeling
is the path to this directory. These changes mount the source code for this plugin to the docker container so the plugin can be built. It also ensures the provisioning yaml will be used to install the plugin and perform some initial setup.
In the same directory, run DSA_USER=$(id -u):$(id -g) docker compose up
. (If that and docker compose version
fail you may first need to invoke sudo apt install docker-compose-plugin
.) Once the deployment is stood up, you can verify that everything has been provisioned correctly by visiting localhost:8080
.
To verify, login as the provisioned admin user and check that the "WSI Superpixel Guided Labeling" plugin is installed by navigating to the Admin console from the sidebar and clicking "Plugins." Additionally there should be a collection called "Active Learning", which should contain one folder called "Data", which in turn should contain folders "Annotations," "Features," and "Models."
After following the installation instructions, you should have a folder called Data in a collection called Active Learning, which looks like the following:
.. image:: docs/screenshots/active_learning_folder.png :alt: Active Learning/Data folder after provision
To enable launching the Active Learning UI from a folder, you'll need to set metadata on the folder. You can do this from here by clicking the blue plus button in the metadata header, selected Simple
, and adding the following metadata property. If you used the recommended provisioning values, this will have already been done.
.. image:: docs/screenshots/active_learning_metadata.png :alt: Metadata to add. Key: active_learning, Value: true
Upload the whole slide images you'd like to use for active learning to this folder, using the green upload button. Once your images have been uploaded, the Active Learning
button should appear in the top right. If not, try refreshing the page. Click the Active Learning
button to begin generating features and models for active learning.
.. image:: docs/screenshots/active_learning_button.png :alt: Button to launch the Active Learning workflow
The first step is to generate the superpixels and feature vectors. Using the form, you can control the approximate size of the superpixels generated (default is 100 pixels), and the magnification level at which to generate the superpixels (default is 5). Once you have chosen values for these fields, click the Generate Superpixels
button. This will start a background process which can be monitored from the Girder Admin Console. This job could take some time to finish, and will take longer the more images you have in your folder. Feel free to close the page or navigate away while the work is being done. If you remain on the page, you will be taken to the next step automatically once the job has finished.
.. image:: docs/screenshots/superpixel_generation.png :alt: The form for superpixel generation
Once superpixels and features have been generated, you will be able to create a set of categories for the superpixels, and label superpixels across your dataset to begin training the active learning models.
.. image:: docs/screenshots/initial_labels.png :alt: The initial label user interface
This view allows users to create new categories, and use those categories to label superpixels by interacting with the image viewer. The top form and buttons underneath are for creating the categories, and navigating between them. Right below that, the Image
drop down menu allows switching the current image shown in the image viewer. Clicking on a superpixel in the image viewer will label that superpixel with the currently displayed category. Clicking again on that superpixel will remove the label. A running total of superpixels labeled per category is available to the right of the image viewer.
Clicking on Begin Training
will kick off a background process to begin training the active learning model using the labels provided in this step. Once that task is completed, you will be presented with a new view containing predictions as described below.
.. image:: docs/screenshots/active_learning_view.png :alt: The active learning view
From here, you can label superpixel features using the film strip area at the bottom to retrain the model. Each block of the film strip depicts one superpixel. The bar at the top of each block shows the most recent prediction. Hovering over this section shows the confidence of that prediction. The superpixels shown are sorted so that users are shown the least confident predictions first. Users can add a label by either agreeing or disagreeing by using the radio buttons. If disagree is chosen, the drop down menu becomes active, and users can add a label by selecting the correct category from the drop down menu.
In order to clear all user inputs on this screen , a Reset All
button is provided. If the predictions for all of the visible blocks matches the actual class of the regions shown, there is an Agree to All
button. You can also view a color-coded pixelmap of the current batch of predictions by clicking the Show/hide Predictions
button.
After labeling some superpixels, a retrain can be triggered with the Retrain
button. This will kick off a job to generate a new batch of predictions, using the newly created labels as input. This job should not take as long as the first, since superpixel and feature generation only needs to be performed once. While that job is running, interactions with this view are disabled. Once the job is finished, new superpixels will be shown to the user for labeling.
The Superpixel CLI Docker image that is used for computing superpixels, extracting features, training models, and predicting labels can be updated without updating the whole system. This can be done by selecting Collections
-> Tasks
-> Slicer CLI Web Tasks
-> dsarchive/superpixel
-> latest
and then clicking on the Pull Latest
button near the upper right.
In the provisioning yaml file, if slicer-cli-image-pull
is used rather than slicer-cli-image
, this will also ensure the latest version of the docker image is available when the system is restarted with docker compose.