ngxson / wllama

WebAssembly binding for llama.cpp - Enabling in-browser LLM inference
https://huggingface.co/spaces/ngxson/wllama
MIT License
371 stars 18 forks source link

[Idea] Stream data from main thread to worker #43

Closed ngxson closed 4 months ago

ngxson commented 4 months ago

Data is now passing as Uint8Array. We can do better by using Streams: https://developer.mozilla.org/en-US/docs/Web/API/Streams_API/Using_readable_streams

We are now using Blob which already provides a ReadableStream.

flatsiedatsie commented 4 months ago

Where's the "on fire" emoji when you need it