Closed JBGruber closed 5 months ago
Now we're quite a bit faster :grin:: ✔ embedded 1736 texts [1m 26.9s]
!
Technical background: I'm sending all texts to embed text at once via asynchronous requests. This essentially uses the queue of Ollama itself instead of queuing on the client side.
Example:
The embedding step ran for 10m 40.2s (
✔ embedded 1736 texts [10m 40.2s]
) while the same task took 2m 45s from the Ollama python library. Not sure what's going on.