Open zichunxx opened 5 months ago
Hi @zichunxx, I will have a look in the next few days after some deadlines. Thank you
Have you tried with another buffer, like the standard ReplayBuffer
or the SequentialReplayBuffer
? Does it give you the same error?
Hi @zichunxx, I will have a look in the next few days after some deadlines. Thank you
No problem! I will try to fix it before you are done with your deadline.
Have you tried with another buffer, like the standard
ReplayBuffer
or theSequentialReplayBuffer
? Does it give you the same error?
I have tried with ReplayBuffer
and there is no OSError
. The above error seems to be triggered by too many .memmap
files generated on disk.
Hi @zichunxx, I tried yesterday on my machine and reached more than 200k steps without errors: how many steps can you print before the error is raised? PS I had to stop the experiment because I was running out of space on hdisk
Hi! The above error is triggered with 5000 steps and a buffer size 4990. Besides, I found this error only occurred when I ran the above program in the system terminal with conda env activated. If I tried this in the VSCode terminal, this error would not happen in 5000 steps, which bothered me a lot.
Hi!
I tried to store episodes with
EpisodeBuffer
andmemmap=True
to release RAM pressure but met this error:For traceback, the following minimal code snippet can reproduce the error
, where
memmap_dir
should be given.Could you please tell me what causes this problem?
Many thanks for considering my request.
Update:
I found this problem seems to be triggered by saving too many episodes on disk. (Please correct me if I'm wrong)
I tried with
EpisodeBuffer
because image observation almost consumes all RAM (64GB) during training, especially with frame stack. I want to complete this training process without upgrading the hardware. So I want to relieve the pressure on RAM withmemmap=True
but encounter the above problem. Any advice on this problem?Thanks in advance.