almutareb / InnovationPathfinderAI

GenAI Research Assistant for Innovation Labs
1 stars 2 forks source link

Stream model output to reduce percieved response time #31

Open almutareb opened 7 months ago

almutareb commented 7 months ago

Currently the model's response could be perceived as slow, as the assistant can take several iterations over different tools to find a suitable answer. To reduce the waiting time we should stream the agent's response.

almutareb commented 7 months ago

Check this: https://python.langchain.com/docs/modules/agents/how_to/streaming and this: https://medium.com/@shrinath.suresh/building-an-interactive-streaming-chatbot-with-langchain-transformers-and-gradio-93b97378353e

vonderwoman commented 7 months ago

The provided explanation details how to improve the user experience of the chatbot you described earlier by incorporating streaming functionalities