-
Hi,
When I run
`python train_ms.py -c configs/config.json -m checkpoints`
I get the below stack trace. I've tried setting all versions of torch, CUDA, and other libraries compatible with each …
-
I know I previously mention edge-tts, which is cloud-based, fast and free, but under the GPL. I have recently been trying out https://github.com/rhasspy/piper/, which uses the VITS model and is under…
-
Hi, I managed to run VITS on Python 3.8.12 but when I upgraded to Python 3.10.12, I've got the following error:
Traceback (most recent call last):
File "/content/drive/MyDrive/vits/train.py", li…
-
### **Description**
We are going to fine-tune Meta's **MMS (Massively Multilingual Speech)** model for **Tibetan text-to-speech (TTS)** using our Tibetan dataset. The pipeline will cover data preproc…
-
I've tried normal speech dataset and generated very natual voice.
But how about training with emotional dataset? any one have a try?
-
>=36k steps and acapella slicing of samples from here (about 60>=)
-
Hi @Muzammal-Naseer and @cgarbin and @kahnchana
I've been reading the paper and looking into the code too. But was not able to fins the piece of codes related to visualization of figures Like F…
-
### Describe the issue
Hi,
I was trying to statically quantize [this Coqui VITS model](https://github.com/coqui-ai/TTS/blob/e5fb0d96279af9dc620add6c2e69992c8abd7f24/TTS/.models.json#L143) that I h…
-
Hi @KdaiP
I’m trying to add TensorBoard to visualize the MEL and audio as shown below. You can play back the audio to see the epoch.
I managed to get it working, but there is a lot of noise if …
-
Hello, I'm grateful for your research. I tried segment-anything with the code you shared. I though it would be fast, but it shows a very slow speed with an average of 13000ms. Can you tell me the reas…