Closed vmesel closed 3 months ago
@vmesel the pr "only has" python code, don't you need to update the documentation?
how do i configure the models on different endpoints?
do i have to read the code to understand ?
the basis of open source is communication and simplification of use, we are getting both wrong
@avelino you are right! i missed that part
Hey, I've been wondering what the next big implementation we need to breed here inside @talkdai. After a quick chat with @avelino, we agreed that we need to start working on allowing users to use multiple models in the same deployment, making it less resource-expensive to have multiple LLMs and prompts.
This approach is quite simple, we need to support any LLM class that a user supplies to us through the
.toml
file and allow the user to choose the URL path for that model and also the prompt it should be using.A quick draft of this new session in the
.toml
file looks like this:The modification on dialog would be simple, on the loading of the project, our system should iterate through the toml endpoints and get all of the settings, setting up routers using the settings passed.