Open jclarkk opened 1 year ago
Hi,
This looks like an issue related to spawning processes in the data loaders (it's something Windows-specific). Can you try to set all num_workers
to 0 in https://github.com/google-research/nerf-from-image/blob/main/data/loaders.py ?
Thanks. I'm getting another process-related exception so I'll keep it on this topic:
I'm receiving a Dataloader worker exception on Debian: `RuntimeError: DataLoader worker (pid 5798) is killed by signal: Killed.
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/nerf-from-image/run.py", line 157, in
It was on: PRETTY_NAME="Debian GNU/Linux 10 (buster)" NAME="Debian GNU/Linux" VERSION_ID="10" VERSION="10 (buster)" VERSION_CODENAME=buster
Running on GCP with "a2-highgpu-1g" and NVIDIA A100 40GB.
Edit: Command used: python3 run.py --resume_from g_shapenet_chairs_pretrained --inv_export_demo_sample --gpus 1 --batch_size 4
Most likely, it's an out-of-memory error. Does this happen with the smaller datasets?
The script pre-loads the entire dataset into memory for performance reasons. You can try to either increase the memory of the VM or extract a sample of the dataset, if that's the issue.
RTX 3080 8GB Windows 10
Python: 3.10 Dependencies: torch==2.0.0+cu117 torchvision==0.15.1+cu117 imageio==2.28.0 opencv-python-headless==4.7.0.72 tensorboard==2.12.2 numpy==1.23.5 scikit-image==0.20.0 scipy==1.10.1 tqdm==4.65.0 lpips==0.1.4 pycocotools==2.0.6 pytorch-fid==0.3.0