Closed yudhiesh closed 6 months ago
exit code 137
indicates its a memory error. I tried increasing the memory available on docker to 16GB
and 12CPU
. Which still throws the same error, therefore just to get things running I tried reducing the number of vectors to load from 10M
-> 100K
by changing the code here:
xb = ds.get_dataset()[:100_000, :]
print(xb.shape)
print("train")
index.train(xb)
But I noticed that this does not fix it nor does it print the shape of the dataset to confirm that its working. Diving into the Dockerfile for faiss, I do not see the faiss.py
file being copied anywhere. I checked the base neurips23
docker image as well and it too does not copy the code over.
I was able to get it to work within a VM Instance on AWS, in order to make sure the changes are added to the new docker image I just reran python install.py --neurips23track filter --algorithm faiss
prior to running the benchmark.
I am trying to leverage this framework to benchmark other Vector Databases for my own understanding. I am trying to run the filter track in order to understand the data flow for new custom databases but I am running into issues.
Steps to reproduce:
python create_dataset.py --dataset yfcc-10
python install.py --neurips23track filter --algorithm faiss
python run.py --algorithm faiss --neurips23track filter --dataset yfcc-10M
It fails at step 3 with the following error logs:
Any idea if I am missing anything here?