Open enescingoz opened 1 year ago
I think your #gpus is not matching with the --sync_bn parameter in 'run.sh'. If running with 4 GPUs, change --sync_bn into 4, not 8. I didn't find the reason inside the code yet.
I think your #gpus is not matching with the --sync_bn parameter in 'run.sh'. If running with 4 GPUs, change --sync_bn into 4, not 8. I didn't find the reason inside the code yet.
I think the problem is related to sync_bn parameter in run.sh.
I changed this parameter to 4 but gives same error(cannot reshape array of size 1 into shape (8)). When i try to use this parameter as 1, i have cuda out of memory error.
RuntimeError: CUDA out of memory. Tried to allocate 292.00 MiB (GPU 0; 7.80 GiB total capacity; 4.92 GiB already allocated; 187.94 MiB free; 5.45 GiB reserved in total by PyTorch) If reserved memory is >> allocated memory try setting max_split_size_mb to avoid fragmentation.
I think your #gpus is not matching with the --sync_bn parameter in 'run.sh'. If running with 4 GPUs, change --sync_bn into 4, not 8. I didn't find the reason inside the code yet.
I want to know why it's like that, could you tell me? thank you! @doohyun-cho
Hello,
Im trying to run training process with the command mentioned in readme.
bash run.sh train bemapnet_nuscenes_swint 30
I completed all of the setup steps and my project directory same as mentioned in readme.
But i have this error
Python version: Python 3.8.10 CUDA Version: 11.8 Torch version: 2.0.1+cu117 (also tried with 1.10.1+cu111)