isl-org / lang-seg

Language-Driven Semantic Segmentation
MIT License
691 stars 85 forks source link

System requirements (GPU?) #1

Closed DanielTakeshi closed 2 years ago

DanielTakeshi commented 2 years ago

Hello,

This is great work @Boyiliee ! I'm excited to try this out.

I have a quick question: what kind of system requirements are necessary to train and run inference on this model? Specifically I am wondering about the type of GPU(s) needed to train LSeg.

Boyiliee commented 2 years ago

Hi @DanielTakeshi, thanks for your interest in LSeg!

For training, we typically use V100, but I assume others would also work if the memory is allowed.

For inference, both GPU/CPU work. If you want to use CPU, for example, you can follow https://github.com/isl-org/lang-seg/blob/main/lseg_app.py#L314. And of course, you can modify based on your machine and requirements.

Hope this helps!

Best, Boyi

DanielTakeshi commented 2 years ago

Thanks! Right now I have some machines with GPUs that have 12 GB of RAM each available, I wonder if that might not be enough memory (assuming we use 1 GPU). However I guess I just have to try the code to see if it work and maybe reduce batch sizes if there are errors.

Boyiliee commented 2 years ago

If you try the demo or inference test, I guess it should work!

But for training, we also provide different options for different model backbone implementation, for example,--backbone clip_vitb32_384, details in https://github.com/isl-org/lang-seg/blob/main/modules/models/lseg_net.py#L119. While I guess for 12GB, you might need to revise a little bit to reduce the model size for training ;)

Hope this helps! And please let me know if there are any other questions!