Open almarklein opened 3 years ago
I apologize. My understanding that mpeg4 would be a huge savings in transmission performance (bytes sent/received) and would make up for any encoding/decoding time increase. This sounds like too much work, but feel free to close or leave this open if others want to comment on it in the future.
Regarding ffmpeg and writing to files, at least on a Linux system you should be able to define a named fifo (os.mkfifo
) to write to and read from. The kernel should not write anything to disk (95% sure).
My understanding that mpeg4 would be a huge savings in transmission performance (bytes sent/received) and would make up for any encoding/decoding time increase
I think this sentence is missing a word, making it ambiguous what you meant :)
"My understanding was that..."
In the other issue you mentioned mp4 video and my mind immediately went to "oh that will compress well so it'll be better in every single way except for maybe code complexity". It is clear that is isn't as much of a victory as I had hoped so I'm not as eager to figure it out.
I think it would be interesting to reopen this issue in 2024.
PyAV is readily installable on many platform and so is imageio-ffmpeg, I think it could really help reduce bandwidth requirements.
I was planning to experiment/benchmark some ideas for improving the performance. Related: @kushalkolar is interested in encoding to jpeg on the GPU.
I think there are a few interesting things about "video" encoders (including newer ones such as H264 and AV1):
I was personally interested in WebRTC since it promises to be "peer to peer". So if you have a server and a computer on a local connection, it will find the natural shortest path (or so it promises).
This seems to be a project that is inline https://github.com/aiortc/aiortc/ with the whole move to "async" which would be really interesting to see how to make it work.
I have to say, i spent all weekend with ChatGPT trying to get WebRTC to work and I somewhat failed.
One of the options mentioned in #3 is the use of an mpeg4 stream. I don't think it's worth the effort to look into it now, but let's collect some thoughts and findings here.
Front-end
Let's start with the easy part. Could be a
<video>
object, or using the webCRT API.MPEG4 encoding
We'd need to encode the images as an mpeg4 stream. Some options:
We'd need to set things up so that the encoding is suited for a stream (e.g. variable framerate, minimal buffering).
Streaming the video
If the server would host a web server, in the form of an URL endpoint that provides a streaming http connection, we can probably push the frames over it. Disclaimer: I've worked with video encoding and web servers, but never combined.
WebCRT
Browsers have an API for streaming video called WebCRT, which may (or may not) make the implementation easier. See e.g. https://github.com/maartenbreddels/ipywebrtc
We'd need an implementation of WebCRT in Python, and that should include mpeg4 encoding (see above). This would be a lot of code/work, unless there exists a library for that.
Something to keep in mind is that the current throttling mechanism schedules the server-side drawing based on how fast the client can display the images. This seems to work very well, and if we'd ever implement WebCRT we should make sure to have a similar coupling.
General thoughts
My current view is that it's not worth the effort. But let's keep the discussion open for the time being.