henghuiding / Vision-Language-Transformer

[ICCV2021 & TPAMI2023] Vision-Language Transformer and Query Generation for Referring Segmentation
MIT License
347 stars 23 forks source link

How to train on the custom data set? #3

Closed YunLongPan closed 3 years ago

changliu19 commented 3 years ago

Hi Yunlong,

Briefly speaking, you need to prepare a dataset annotation JSON file, then specify the path to the image folder and target mask folder in the config file. For details please check:

https://github.com/henghuiding/Vision-Language-Transformer/blob/9b24015566fa820e3eddbbd8942fa44512ec1b3c/data/data_process_v2.py#L77

https://github.com/henghuiding/Vision-Language-Transformer/blob/9b24015566fa820e3eddbbd8942fa44512ec1b3c/loader/loader.py#L93

YunLongPan commented 3 years ago

Thank you!

At 2021-10-06 18:30:04, "ntuLC" @.***> wrote:

Hi Yunlong,

Briefly speaking, you need to prepare a dataset annotation JSON file, then specify the path to the image folder and target mask folder in the config file. For details please check:

https://github.com/henghuiding/Vision-Language-Transformer/blob/9b24015566fa820e3eddbbd8942fa44512ec1b3c/data/data_process_v2.py#L77

https://github.com/henghuiding/Vision-Language-Transformer/blob/9b24015566fa820e3eddbbd8942fa44512ec1b3c/loader/loader.py#L93

— You are receiving this because you authored the thread. Reply to this email directly, view it on GitHub, or unsubscribe.