openml-labs / ai_search

RAG pipeline and summary for openml
https://openml-labs.github.io/ai_search/
0 stars 0 forks source link

reduce number of llm instances #40

Closed SubhadityaMukherjee closed 1 week ago

SubhadityaMukherjee commented 3 weeks ago

There are too many llm instances now. Because each is it's own service. This is not good and should be re-written. For instance, llama3 is used for deciding what kind of query it is, structured query analysis, and also generating results.

PGijsbers commented 2 weeks ago

But how do you propose to do this? Create a complex query that does multiple things (at the risk of being more unpredictable)? Replace an LLM by other methods?

SubhadityaMukherjee commented 1 week ago

Well turns out the issue was that the same LLM was being re-initalized multiple times which added to the time taken to get results. (which was the reason this was created) Other than that, it doesnt seem like there is anything else we can do, so I am closing this issue.