Open almutareb opened 7 months ago
Currently the model's response could be perceived as slow, as the assistant can take several iterations over different tools to find a suitable answer. To reduce the waiting time we should stream the agent's response.
Check this: https://python.langchain.com/docs/modules/agents/how_to/streaming and this: https://medium.com/@shrinath.suresh/building-an-interactive-streaming-chatbot-with-langchain-transformers-and-gradio-93b97378353e
The provided explanation details how to improve the user experience of the chatbot you described earlier by incorporating streaming functionalities
Currently the model's response could be perceived as slow, as the assistant can take several iterations over different tools to find a suitable answer. To reduce the waiting time we should stream the agent's response.