replicate / llama-chat

A boilerplate for creating a Llama 3 chat app
https://llama3.replicate.dev
Apache License 2.0
829 stars 306 forks source link

Stream output using server-sent events #7

Closed mattt closed 1 year ago

mattt commented 1 year ago

WIP

https://github.com/replicate/chat/assets/7659/290459d3-92d7-4bf7-9821-51e8147405cf

zeke commented 1 year ago

Moo!

bfirsh commented 1 year ago

This is fantastic. Does it stream token by token? How fast does first token arrive?

Ideally we want it to feel as snappy as https://labs.perplexity.ai/

bfirsh commented 1 year ago

Convo https://replicatehq.slack.com/archives/C04SGD84KC4/p1690191298627139