Open GraceZhuuu opened 8 months ago
BTW, i modified here for error sampler option is mutually exclusive with shuffle
@@ -303,7 +303,7 @@ def train(local_rank, args):
train_dataset = DataSet(train_data_dir, img_transforms,vid_list=args.vid, resol=args.resol, frame_gap=args.frame_gap, )
train_sampler = torch.utils.data.distributed.DistributedSampler(train_dataset) if args.distributed else None
- train_dataloader = torch.utils.data.DataLoader(train_dataset, batch_size=args.batchSize, shuffle=True,
+ train_dataloader = torch.utils.data.DataLoader(train_dataset, batch_size=args.batchSize,
num_workers=args.workers, pin_memory=True, sampler=train_sampler, drop_last=True, worker_init_fn=worker_init_fn)
Thank you for your interest and notice! As the baselines used single-batch training, we have not focused on multi-GPU training. Does it perform the same with reduced training time? It would be very useful to implement it!
Thank you for your interest and notice! As the baselines used single-batch training, we have not focused on multi-GPU training. Does it perform the same with reduced training time? It would be very useful to implement it!
I have not run through the program under multiple GPUs, but it is okay under a single GPU. I am a novice and trying to solve it. Thanks again for your work.
Thanks for your great work.
I'm trying to training in parallel, singe machine and two GPUs. The program uses a single graphics card by default, then i set the
--distributed
but there occurs an error
i checked the code, here