-
-
Problem:
The current implementation of Whisper in Open-WebUI uses a limited, robotic voice for all interactions.
While this is functional, it can be jarring and unnatural, making it difficult for …
-
Maybe we could use the Web Speech API to create a plugin that records spoken responses with automated speech recognition?
This could be based on the html-audio-response plugin and used to run tasks …
-
**What problem or use case are you trying to solve?**
For some people, typing is difficult. But because AI tools accept natural language input, it should be easy for those people to give instructions…
-
### OpenAI
According to the Models documentation the interfaces should be names as `ITextToSpeech` and `ISpeechRecognition` accordingly.
This issue is to keep track of this needed change, accor…
-
I'm looking to create custom [viseme](https://learn.microsoft.com/en-us/azure/ai-services/speech-service/speech-synthesis-markup-structure#viseme-element) animations for lip syncing in my project.
…
-
Hello! I was wondering if this is still being worked on? I'm looking into being able to pass text into uberduck api and receive back the tts audio. If this project is no longer supported, would you mi…
-
使用webui推理报错:
Traceback (most recent call last):
File "/data/workpace/fish-speech/fish_speech/webui/app.py", line 241, in inference
resp.raise_for_status()
File "/data/workpace/fish-speech/…
-
![image](https://github.com/SubtitleEdit/subtitleedit/assets/16167396/89e5e555-776f-451e-9e50-a92ae50337fe)
So, TTS done everything sentence by sentence, and when it is time to combine everything, on…
-
**Is your feature request related to a problem? Please describe.**
I'm building an app with androidjs, and I need the application to speak a text, and it should work offline
There are several onli…