Closed MicroHest closed 2 years ago
Hi @MicroHest ,
We implemented and tested our code on the lab's slurm cluster, which features 2 socket E5 processor, 8 V100 GPUs (with 32G gRAM) and 512GB RAM, so it may be problematic when running on bare metal machines. Indeed, it requires much RAM, for we have to extract all the frames from the video, before sending them into the network. We're looking into this issue and hope it can run on some widely-used graphic cards like 1080Ti. Please kindly wait for a few days.
OK,thanks
Hi @MicroHest ,
Sorry for the delay. I have updated the code for bare-metal machines. Please pull the latest code and follow the instructions https://github.com/OpenPerceptionX/Openpilot-Deepdive#training-and-testing .
Also, please kindly note that the batch_size is actually the batch size per GPU. When setting it to 6, it consumes around 30G GPU memory, so it's interesting to know that you can run with batch_size=8
on 24GB 3090s without OOM.
As for RAM, when setting batch_size=6
and n_workers=4
, it consumes around 40 to 50 GB RAM per process. (That means, if you run with 8 cards, you will have to multiply it by 8).
I run this project. But it seems that I‘m either out of RAM or GPU memory.
At first, I run this project with the default setting according to the paper.
Then I tried a smaller batch size of 8 but failed with CUDA out of memory. Finally, I modified optimize_per_n_step to 20. Well, this time it works. But after a short time, I found my work process 2 exited unexcepted. After observing the top command's panel, I found my computer had run out of RAM before the process exited.
Finally, my computer works well in configuration below.
Could you tell me how much CPU RAM and GPU memory is recommended for this project? Or your computer info.