Describe the bug
The VideoFrame can be instantiated using a canvas in the Javascript SDK. How can the same be done in Android?
In Android, to create a VideoFrame we need to provide a buffer. Do we have any workaround to convert the canvas to a buffer?
To Reproduce
Post-frame processing e.g. blurring the outgoing video feed
Expected behavior
There should be a way to create the VideoFrame from the canvas.
Describe the bug The VideoFrame can be instantiated using a canvas in the Javascript SDK. How can the same be done in Android? In Android, to create a VideoFrame we need to provide a buffer. Do we have any workaround to convert the canvas to a buffer?
To Reproduce Post-frame processing e.g. blurring the outgoing video feed
Expected behavior There should be a way to create the VideoFrame from the canvas.
Additional context LiveKit track processing