Sense-GVT / Fast-BEV

Fast-BEV: A Fast and Strong Bird’s-Eye View Perception Baseline
Other
567 stars 85 forks source link

Huge memory consumption during training #83

Open LaCandela opened 2 months ago

LaCandela commented 2 months ago

I noticed that during training the process consumes lots of CPU memory. For example, if I use 8 workers with batch size of 45 the memory consumption is more than 50GB and it is slowly increasing up to 60GB when my system shuts down. Has anyone noticed this before? What is the memory intensive part of the training pipeline? Is it related to some inefficient mmcv operation?

I start the training with this command: CUDA_VISIBLE_DEVICES=0 python tools/train.py --config configs/fastbev/exp/paper/fastbev_m0_r18_s256x704_v200x200x4_c192_d2_f4.py --work-dir /Fast-BEV/temp

Could it be a problem that I don't use slurm?