Open rvjosh opened 9 months ago
Are you on a machine with GPU? It's pretty fast when running on GPU enabled machines.
We haven't seen a strong enough need to support generating embedding via API yet. But open to considering this if there's a strong enough reason to do so in the future
@rvjosh I've just worked on adding in support to use 3P embeddings API hosted on huggingface (if you configure it to be run as a Sentence Embedding
task). See PRs #609 , #616 . This should be fairly extensible with using the OpenAI service directly (eventually), but it would also work for any model via HuggingFace now. Will be available in the next release.
I know this is likely the opposite of the requests you all usually get on here, but I have an Obsidian vault that is ~2000 documents and it is taking a long time (40 min and counting to do the first 900 documents) to embed my vault running a Docker container locally. I see that you use the "thenlper/gte-small" model by default to do the embeddings - would you consider allowing for using 3P models to do the embeddings that are called via API such as OpenAI's text-ada-002? Thank you for the great app!