Open shiyuangu opened 5 months ago
The way I've done it now is to only extract the source nodes when processing the first chunk of the stream 🤷
Thanks for the feedback. This might be a bug, I'm working on to support getting all context during the Iterator
@shiyuangu this is how create-llama is currently handling the source nodes inside of a stream: https://github.com/run-llama/create-llama/blob/1ad25451a69e10183200dfaf4c7a8d6c8f800543/templates/types/streaming/nextjs/app/api/chat/llamaindex-stream.ts#L42
@himself65 so your idea is to remove the source nodes out of the stream and instead move them to some response context that can be read while streaming and that can contain more information? Sounds great
Hi all, I want to retrieve the source nodes from a stream response of ContextChatEngine.chat. When I look at the code, I observe that each chunk of the stream response has the source nodes attached to it in a duplicate manner. Is it a bug or a feature intentionally designed? What's the recommended way to extract the source codes from a streaming response? Thank you!
https://github.com/run-llama/LlamaIndexTS/blob/bcc3d0b4d1c7afc6c6e6c26f4f03a49d92e14d17/packages/core/src/engines/chat/ContextChatEngine.ts#L89