-
### **Description**
We are going to fine-tune Meta's **MMS (Massively Multilingual Speech)** model for a Tibetan speaker named **Sherab** using Sherab's dataset. The process includes preparing Shera…
-
- Speech SDK log taken from a run that exhibits the reported issue.
Check here [https://gist.github.com/Elshaffei/cb1f13f1d79ccd6df0641b864420bc93](url)
- A stripped down, simplified version of y…
-
torch.cat(): expected a non-empty list of Tensors
File "D:\ai\ComfyUI-aki-v1.3-chumen0731\execution.py", line 317, in execute
output_data, output_ui, has_subgraph = get_output_data(obj, input_data…
-
Could you add a native speech to speech / audio-to-audio support with encoder (tokenizer) and decoder (back to audio waves)
I was able to implement a decoder only model, I first used audio codec to…
-
### Checklist
- [X] I have searched the [existing issues](https://github.com/streamlit/streamlit/issues) for similar feature requests.
- [X] I added a descriptive title and summary to this issue.
##…
-
### Describe the bug
Something in the way that it passes input into the speedy_speech model (tts_models/en/ljspeech/speedy-speech) is bugged and errors out for short inputs. It wants them to be a s…
-
## ❓ Questions and Help
Hi silero team!
When i try to use silero-vad using python, I felt it is good.
But if i use silero-vad using c++, i got quite different result between python and c++.
I …
-
Hi @trevorlinton
any plan about
``
feat. support?
Thank you
-
Mini-Omni提供了一个很棒的思路,可以将LLM结合TTS,与等待LLM流式返回后再传给TTS做合成相比,无疑在降低延时方面理论上有显著提升。
但对于输入的部分,跟调用ASR后得到文本,再将文本作为模型输入相比,将语音编码后直接输入到模型有什么效果上或者延时上的优势吗?
提出这样的问题主要是因为,我们在人机对话的过程中,如果要降低响应延时,怎么在vad方面做优化是一个很大的难点,如…
-
想使用sencevoice模型进行asr,并且区分说话人。加载一直报错。
加载模型代码如下
from funasr import AutoModel
from funasr.utils.postprocess_utils import rich_transcription_postprocess
model = AutoModel(model="/data/asr/SenseVoice…