SHI-Labs / OneFormer

OneFormer: One Transformer to Rule Universal Image Segmentation, arxiv 2022 / CVPR 2023
https://praeclarumjj3.github.io/oneformer
MIT License
1.44k stars 130 forks source link

Custom dataset training clarification #5

Closed SkalskiP closed 1 year ago

SkalskiP commented 1 year ago

Hi team 👋!

First of all great project! I'm super excited to see that you used Detectron2 as framework of choice.

I'm trying to train my own model using custom dataset in COCO format. And for now I have 2 questions:

  1. Do I need to provide annotations for all tasks? Thats how I understand those guidelines. I'm mostly interested in instance segmentation task:
coco/
  annotations/
    instances_{train,val}2017.json
    panoptic_{train,val}2017.json
    caption_{train,val}2017.json
    # evaluate on instance labels derived from panoptic annotations
    panoptic2instances_val2017.json
  {train,val}2017/
    # image files that are mentioned in the corresponding json
  panoptic_{train,val}2017/  # png annotations
  panoptic_semseg_{train,val}2017/  # generated by the script mentioned below
  1. Do I need to train on 8x A100. I understand that you needed that much power when you trained from scratch, but if I use your checkpoint will 1x A100 be sufficient?
praeclarumjj3 commented 1 year ago

Hi @SkalskiP, thanks for your interest in our work. Please find the answers to your questions below. I have tried to be as comprehensive as possible.

  1. Annotations:

  2. Training Resources:

    • I haven't tried training on a single GPU myself. I suggest training on at least 2 or 4 A100 GPUs because the hyper-parameters are set for multi-GPU training. Also, you might encounter OOM on a single GPU.
praeclarumjj3 commented 1 year ago

Feel free to re-open if you face any issues.

Ankitajadhav611 commented 1 month ago

Hey, I am trying to train a custom dataset for instance segmentation which is in the coco data format. I am able to train the model, it lacks bbox coordinate while inference, thus unable to infer the results.