Closed AntreasAntoniou closed 1 month ago
Hey @AntreasAntoniou, I'm looking into the issue currently but the behavior is indeed very weird. Could you try out other datasets, maybe different splits of imagenet, and let me know if the behaviour persists?
There could also be a potential memory leak that's preventing s3 from reading data, could you share the output of htop after running this?
Closing as the original issue was fixed and the argument conflicting issue was fixed with PR https://github.com/activeloopai/deeplake/pull/2954
🐛🐛 Bug Report
I was trying to increase the performance of the dataloader by giving it more cache both in memory and disk, and somehow it backfired, as I now keep getting the following error even when I remove the cache requirements.
Note: I am using the skip/agreement branch as I am running my stuff at scale and do not want to accept all agreements every time, also, somehow the current way hub works does not play well with Hydra and the skip/agreement branch was a quick workaround by @davidbuniat to help me out
⚗️ Current Behavior
It seems that using large cache sizes breaks the dataloader.
Input Code
Expected behavior/code The dataloader should work, and use the cache to speed things up in later epochs
`
⚙️ Environment
Python
version(s): 3.8OS
: Ubuntu 20.04IDE
: [PyCharm]Packages
: [Tensorflow==2.1.2 - latest
, Pytorch==latest]🖼 Additional context/Screenshots (optional)
Add any other context about the problem here. If applicable, add screenshots to help explain.