Epiphqny / VisTR

[CVPR2021 Oral] End-to-End Video Instance Segmentation with Transformers
https://arxiv.org/abs/2011.14503
Apache License 2.0
740 stars 96 forks source link

Paper reproduction #67

Open AI-student-wl opened 3 years ago

AI-student-wl commented 3 years ago

Hello, I am reproducing the results of your paper, but I have encountered some problems. Due to my limited level of in-depth learning, I ask you for help and report an error. I have put the screenshot below. Another question is where to put the mask picture. Thank you for your help. This is my data format VisTR ├── data │ ├── train │ ├--------JPEGImages │ ├------------.jpg │ ├── val │ ├--------JPEGImages │ ├------------.jpg │ ├── annotations │ │ ├── instances_train_sub.json │ │ ├── instances_val_sub.json ├── models

This is the command line I entered: python -m torch.distributed.launch --nproc_per_node=1 --use_env main.py --backbone resnet101 --ytvos_path /media/dmia/code1/why/VisTR/data --masks --pretrained_weights /media/dmia/code1/why/VisTR/models/resnet101.pth

23