hi there, it is a great component, it works fine with my product.
I have a usecase:
The end user chat through the stream chat, I want the LLM to response the tool function call first, to drive the UI action, such as openCurrentDocument command, I will response the command in my vue3 fronend, but how to deliver the command json to my frontend? After that, the LLM will response "the document is opened" .
I have no idea how to simply use deep-chat to archive that (this is my prior choice), otherwise I need but another websocket channel by myself?
Hey @BeamLiu.
Forgive me but I don't seem to understand your use-case. Could you perhaps elaborate a little more with examples on what it is you are trying to achieve. Thanks!
hi there, it is a great component, it works fine with my product.
I have a usecase: The end user chat through the stream chat, I want the LLM to response the tool function call first, to drive the UI action, such as openCurrentDocument command, I will response the command in my vue3 fronend, but how to deliver the command json to my frontend? After that, the LLM will response "the document is opened" .
I have no idea how to simply use deep-chat to archive that (this is my prior choice), otherwise I need but another websocket channel by myself?