Open fire opened 4 months ago
That's definitely in scope. The best way to approach this would be to introduce the necessary embedding providers and to modify or create a new pipeline that shows an example of this in action.
I'm happy to team up on this.
I have two primary usecases:
For image embedding, do you think we can fit it into the pipeline here [https://github.com/SciPhi-AI/R2R/blob/main/r2r/pipelines/basic/ingestion.py] with a specific embedding provider, or do you think we need to fundamentally rework the structure of the codebase in some way?
I think multi-modal is an important use case and I am very interested in figuring out how to best support this.
I don't think I can drive multi-modal too much, but I'll see what spare time I can gather.
The obvious question are like what happens when we have two different embedding models like token integers, how do we sync them?
Hi,
I was wondering if it was in scope to embed media?