Open ksv87 opened 8 months ago
Bump. I'm facing the same issue where the training time is around 5 days on 3 3090s with batch size 192 and size of dataset ~400k images. The number of workers is set to 8.
2024-09-28 05:43:37 | INFO | damo.apis.detector_trainer:368 - epoch: 1/40, iter: 300/1483, mem: 7203Mb, iter_time: 7.779s, model_time: 7.283s, total_loss: 2.5, loss_cls: 0.2, loss_bbox: 1.6, loss_dfl: 0.7, lr: 4.878e-05, size: (640, 640), ETA: 5 days, 10:22:08
As a result, I looked at the caching method in YOLOX and tried to apply it in DAMO-YOLO, the ETA was reduced from 2 days 18 hours to 8 hours, but I'm not sure that this did not affect the performance of the model - I'm conducting an experiment for evaluation.
It would be great if the authors added caching of the dataset
just too slow>............why?
Before Asking
[X] I have read the README carefully. 我已经仔细阅读了README上的操作指引。
[X] I want to train my custom dataset, and I have read the tutorials for finetune on your data carefully and organize my dataset correctly; 我想训练自定义数据集,我已经仔细阅读了训练自定义数据的教程,以及按照正确的目录结构存放数据集。
[X] I have pulled the latest code of main branch to run again and the problem still existed. 我已经拉取了主分支上最新的代码,重新运行之后,问题仍不能解决。
Search before asking
Question
the damo-yolo very similar with yolox, but traning on my custome dataset damo-yolo traning time x3 by comparison with yolox, why?
Additional
No response