Closed mnsmar closed 1 year ago
Should be fixed in 53a0a4f. The issue was that the code currently doesn't support multi-gpu inference, and pytorch lightning is too smart and picked it automatically since you have 3 GPUs.
okay thanks. Pulled the changes and testing it now. Will close the issue if it runs successfully.
fyi I notice that with 3 GPUs it was doing ~20it/s. With 1 GPU now is doing 19.7it/sec. I did not expect that small difference. I suspect either something wrong with iteration counting or with execution on multiple GPUs. Weird thing is that it was keeping all 3 GPUs at 100%. It previously took 11 hrs to complete. Let's see how long it takes now.
I used the command
python3 rnamodif/evaluation/run.py --workers 10 --datadir fast5/ --outfile out.tab --model 5eu_v1
but it seems to fail at the last step.