Open nuwang opened 2 years ago
Following up on this, it looks like this could be a design issue. While on the surface, the use_cache
option works with s3fs (caching didn't work with geesefs), I'm not sure that the multiple processes wouldn't step on each other's toes. In CVMFS for example, they appear to be using a single mount per volume, which is then subsequently bind mounted per PVC: https://gitlab.cern.ch/cloud-infrastructure/cvmfs-csi/-/blob/master/pkg/cvmfs/nodeserver.go#L114. That way, there's only one cvmfs process per volume, and therefore, the cache is common. This same issue seems to exist in the ctrox csi driver. Ping @vitalif
Hi,
I'm trying to enable caching in the CSI driver. I've passed extra mountOptions as follows:
and they are being passed in correctly according to the logs:
However, the /tmp directory remains empty. Am I doing something wrong?
Also, with multiple pods mounting the same PVC, would the cache work correctly? I can see that there are multiple geesefs processes running, all pointing to the same cache path.
Finally, we want to use this with long-living, entirely read-only data (these are reference genomes and associated read-only data). This is why I set the
cache-to-disk-hits
to 1, assuming that caused the file to be cached on the very first read. Could you please recommend the best settings for very aggressive caching? I've noticed a lot of S3 calls being made for the same path even though that path for instance, has already been checked recently.