Open Jozdien opened 2 years ago
It doesn't seem to be the problem of the code. Can you check how much space you have on your disk? In traintest.py
, we save the output predictions of each epoch, which might take a few GBs for the training process, depending on how large your test set is.
I don't think it's the disk space because I'm testing this on a very small dataset (<20 samples). Could it be possible that something is writing a large amount of data to one particular file, and the filesize limits vary between systems (very speculative)?
I see. I would suggest running the ESC-50 recipe and see if the same error raises - it would be fast and easy to run; if you still see the error you can check your os; otherwise, you can check if your modification is correct or not.
traintest.py
does save the prediction files, which could be large, but considering you have only 20 samples, it is unlikely the case.
Yep, got the same error, although this time far later in the training process (ESC-50 ran for about 16 epochs before halting, on my dataset I don't recall it making any progress on training). Just to know what kind of specifications I should use, what amount of disk space do you recommend using?
I was trying to use the Audioset pretrained model for finetuning on a very small dataset to test it out on. At first the process would simply be killed with "Out of memory" in the log, but when I moved to a larger system, the process ran for longer before returning this error:
As far as I can tell, that OSError could indicate that the filesize has been exceeded, not just that the total memory is overflowing. I haven't changed
traintest.py
except for adding an elif condition for the finetuning dataset. Did you run into this error while finetuning, or does it seem like something you understand the cause of?