Open Kanishk-Kumar opened 5 months ago
Update:
Getting around same wait time for ingestion and exactly same for deletion when both llm and embedding mode is "mock". Also tried only keeping llm as mock, didn't work. I have also tried a different database:
poetry install --extras "llms-ollama ui vector-stores-postgres embeddings-ollama storage-nodestore-postgres"
Faced same issues there.
17:28:14.186 [INFO ] private_gpt.server.ingest.ingest_service - Ingesting file_names=['cleaned_50_rows.csv']
Parsing nodes: 0%| | 0/1 [00:00<?, ?it/s]17:28:14.237 [INFO ] uvicorn.access - 192.168.1.2:51586 - "GET /queue/data?session_hash=skizaephyc HTTP/1.0" 200
Parsing nodes: 100%|██████████| 1/1 [00:00<00:00, 8.14it/s]
Generating embeddings: 100%|██████████| 5259/5259 [00:00<00:00, 303480.11it/s]
17:28:15.617 [INFO ] private_gpt.components.ingest.ingest_component - Saving 1 files (1 documents / 5259 nodes)
17:28:48.822 [INFO ] private_gpt.server.ingest.ingest_service - Finished ingestion file_name=['cleaned_50_rows.csv']
17:28:48.903 [INFO ] uvicorn.access - 192.168.1.2:34634 - "POST /queue/join HTTP/1.0" 200
17:28:48.959 [INFO ] uvicorn.access - 192.168.1.2:34646 - "GET /queue/data?session_hash=skizaephyc HTTP/1.0" 200
17:28:49.022 [INFO ] uvicorn.access - 192.168.1.2:34650 - "POST /queue/join HTTP/1.0" 200
17:28:49.077 [INFO ] uvicorn.access - 192.168.1.2:34652 - "GET /queue/data?session_hash=skizaephyc HTTP/1.0" 200
17:29:09.919 [INFO ] uvicorn.access - 192.168.1.2:46896 - "POST /queue/join HTTP/1.0" 200
17:29:09.974 [INFO ] uvicorn.access - 192.168.1.2:46900 - "GET /queue/data?session_hash=skizaephyc HTTP/1.0" 200
17:29:11.599 [INFO ] uvicorn.access - 192.168.1.2:46908 - "POST /queue/join HTTP/1.0" 200
17:29:11.615 [INFO ] private_gpt.server.ingest.ingest_service - Deleting the ingested document=3d7949bf-84d4-42c8-8ad9-b3f48c8d10e1 in the doc and index store
17:29:11.658 [INFO ] uvicorn.access - 192.168.1.2:46922 - "GET /queue/data?session_hash=skizaephyc HTTP/1.0" 200
17:30:31.862 [INFO ] uvicorn.access - 192.168.1.2:38976 - "POST /queue/join HTTP/1.0" 200
17:30:31.925 [INFO ] uvicorn.access - 192.168.1.2:38992 - "GET /queue/data?session_hash=skizaephyc HTTP/1.0" 200
System specs: Intel(R) Core(TM) i9-14900K GeForce RTX 4070 Ti RAM: 128 GB
Lookup + inference speed is similar to ChatGPT, quite fast. But deleting 7.5 MB .txt file takes 7 hours, ingestion takes ~28 minutes for the same. Tried to reproduce in the log below using 958KB .csv file having one column with clean text (web articles) in each row:
Is it supposed to take this long? I see full GPU usage during first "Generating embeddings", then for subsequent "Generating embeddings" both GPU/CPU usage is less than 3%. VRAM usage 350 MB at most. Settings I'm using (settings-ollama.yaml):
Also tried:
And:
Same wait time in all cases. (Using model with less dimensions is fast, but subsequent "Generating embeddings" still slow.)