Closed CasonTsai closed 4 months ago
Hey,
I haven't implemented that yet, however I think it's not that complicated, the headless server of Nvidia supports it. Feel free to contribute, I would appreciate!
Best regards
感谢无私奉献,能否进一步实现实时流式驱动?
Confirmed. Will work on it. The project will be improved and continued.
Started working on it. In a second step I will also move away from plain requests and build on fastSDK and media-toolkit. Second step will take longer though. I'll release an upgraded module first - can't tell you yet how long it takes, but I expect to finish in a few weeks.
@singelhero @CasonTsai can you give me a little bit more information where your input stream comes from? Do you expect a live microphone audio input? What are your aims / projects you're working on about? How would this feature look like in your opinion?
For the beginning I would implement chunking an audio file and feeding it to the stream.
Finished streaming implementation. @singelhero, @CasonTsai please check feature especially interesting wiht livelink. Please share result. Raise new issue if something is missing or misbehaving. Next steps: fastSDK
@w4hns1nn yes,
@singelhero @CasonTsai can you give me a little bit more information where your input stream comes from? Do you expect a live microphone audio input? What are your aims / projects you're working on about? How would this feature look like in your opinion?
For the beginning I would implement chunking an audio file and feeding it to the stream.
@singelhero @CasonTsai can you give me a little bit more information where your input stream comes from? Do you expect a live microphone audio input? What are your aims / projects you're working on about? How would this feature look like in your opinion?
For the beginning I would implement chunking an audio file and feeding it to the stream.
yeah ,i want to build a metahuman in unreal.the input stream from microphone audio input or audio data in memory .i found a solution in audio2face official doc,it use grpc in that document. excuse me,I want to know if there is a way to use audio2face without opening omniverse?
@CasonTsai I've added the streaming feature which uses gRPC. Check the updated readme for it. About the second question:
@CasonTsai I've added the streaming feature which uses gRPC. Check the updated readme for it. About the second question:
- please check if the headless server also uses omniverse.. py_audio2face runs with the headles..
- I plan something similar, but will be end of the year until i can publish something about
thanks
thank you for your great work,May I ask if this project supports sound streaming into audio2face to drive lip shapes, rather than in the form of voice files?