Closed mumulmaulana closed 6 months ago
Hi @mumulmaulana , Your issue most probably originates from your RAM; you don't have enough to pre-load all the ResNET features with dimension 2048. I ran my experiments on a workstation with 256GB of RAM, which could also accommodate the Baidu features of even larger dimensions (~8k AFAIK). The PCA512 version are reduced to a dimension of 512 with PCA, which consumes less RAM for fairly similar performances.
A solution for you would be to update the dataset class to load the features when sampling them in __getitem__
, instead of pre-loading them all on __init__
.
I hope that helps!
Thanks for responding! I will try the workaround first and get back with the result!
Hello! Thank you so much for providing this open repository for everyone!
I would like to ask about my issue running training for the Benchmark model. I tried training for the NetVLAD++ benchmark model in my system and ran into this issue:
This is the parameters for the training:
I have tried to reduce the batch_size and max_num_workers as well, but the problem persists. Also, not sure if it is needed, but it seems it only happens after it runs the test() procedure.
I found no trouble training with PCA512 features, so I figure there must be some resource issues. When I look at the System Monitor and GPU usage, this is the reading of the result.
I have tried increasing the swap file to more than 64GB, but it only results in my system freezing and crashing. Are you familiar with this issue? Is there any way for me to bypass this? Or, if it is no trouble, can you tell me the recommended specification to run this training? Thank you!
My system specs (RAM 16GB, SSD 2TB):