Project-MONAI / MONAI

AI Toolkit for Healthcare Imaging
https://monai.io/
Apache License 2.0
5.67k stars 1.04k forks source link

[Feature Request]: CLIP Driven Universal Model #5800

Open tangy5 opened 1 year ago

tangy5 commented 1 year ago

Contrastive Language-Image Pre-training (CLIP) Driven Models and Partially Supervised Learning for Medical Image Segmentation

This issue is to discuss adding the CLIP-Driven Universal Model Features to MONAI.

Potential assignee: @tangy5

CLIP-Driven Universal Model

Key features

The implementation will bring several new feature as follows:

  1. Universal Model: one model to detect and segment all abdominal organs and all types of tumors (Liver tumor, kidney tumor, Lung nodule, Pancreas tumor, hepatic vessel tumor, colon tumor).
  2. Language model (CLIP) and text-driven embeddings boost medical image analysis.
  3. Training Partial labelled datasets.
  4. Incremental learning: Users can continue to train new segmentation classes using the current trained model without catastrophic forgetting.

⏳ Dataset: The Universal Model is trained with following datasets

Screenshot from 2023-01-03 13-16-57

Implementation plans

More Details of the Feature Methodology:

  1. Universal Model: Screenshot from 2023-01-03 12-09-23

  2. CLIP Driven and text-driven segmentor: Screenshot from 2023-01-03 12-10-09

  3. Partial Supervised Learning: Screenshot from 2023-01-03 12-04-46

  4. Incremental Leraning:

Screenshot from 2023-01-03 12-11-14

Detailed steps of implantation will provide after open discussion.

Welcome all suggestions and comments!

@ljwztc @MrGiovanni

Da1daidaidai commented 1 year ago

looking forward for updating

tangy5 commented 1 year ago

looking forward for updating

Thanks, very appreciative, will continue to working on this, and provide useful API for CLIP related integration. Welcome comments and suggestions.