-
I wish to generate caption text for images that I'll be providing. My understanding is that sample_speaker.py will do this. However, when I run it I get an error. Here's what I run in terminal, with t…
-
I've created a Training set based on LibriSpeech with two additional speakers. Ive pruned out speakers that have a small ammount of data, as well as the samples that are very short or very long, simi…
-
### System Info
- `transformers` version: 4.39.3
- Platform: macOS-14.4-arm64-arm-64bit
- Python version: 3.11.8
- Huggingface_hub version: 0.22.1
- Safetensors version: 0.4.2
- Accelerate versi…
-
When I try to run the basic inference demo I get mismatches for the dimensionality of the pretrained models and context_lstm. There isn't really any location in any of the project files where 1044 or …
-
## Abstract
- propose Vector Quantised Variational AutoEncoder (VQ-VAE)
- generative model that learns discrete representations
- prior is learnt rather than static
- solves the issue of "po…
-
Hello,
Based on your code, I added Korean tokens (using a Korean emotional dataset) to the tokenizer and fine-tuned the model with the LibriTTS R dataset. The Korean dataset is slightly less than 3…
-
Will support Mandrain?
-
File ~/work/python/tts/MeloTTS/melo/api.py:81, in TTS.tts_to_file(self, text, speaker_id, output_path, sdp_ratio, noise_scale, noise_scale_w, speed)
79 t = re.sub(r'([a-z])([A-Z])', r'\1 \2'…
-
Hi! I've just installed Amica and am trying to use it locally with Oobabooga's webui and Alltalk v2 standalone, both newly updated/installed, on a Windows 10 system. Sending a message to the webui API…
-
Still in colab when i run this cell as suggated by @atulpokharel-gp :
`from transformers import AutoTokenizer, AutoModelForTextToWaveform
tokenizer = AutoTokenizer.from_pretrained("facebook/mms-tt…