-
1. Which model does Yukarin uses for its training?
2. Are there any target voice training document specifications?
3. Would public voice datasets help with training?
4. Does this project work with…
-
There is a weird issue in the manylinux container (on github actions)
that fails the wheels for python3.10 & python3.11
Here is the detailed logs from the github actions console
```
+ /opt/py…
-
Hello,
I saw that you used `pad`, `audio_slice_frames`, `sample_frames` but I can't understand the usage of those params. Can you explain the meanings of them?
Also, `WaveRNN` model was using pa…
-
I am interested in vocal cloning from van der Oord et al and found this repo. I try to install on a Google Colab and get:
```
!pip install -r ZeroSpeech/requirements.txt
Collecting numpy==1.18.…
-
Some weird (TLS, timeout) errors have been appearing when trying to download datasets.
-
## 🐛 Bug
### To Reproduce
Steps to reproduce the behavior (**always include the command you ran**):
1. Run cmd
```
fairseq-train --fp16 $RESULT/quantized/fairseq-bin-data \
--task …
-
# Linguistic knowledge-driven tasks
## Task Objective
- Many of the tasks seem to be focusing on the application-specific downstream tasks.
- We aim to provide a set of tasks that are based on va…
-
When using `context: triphone-within` in the `params.yaml` file, the execution process calculates phoneme-within and phoneme-any contexts as well. In the eval call, the abx_sets are calculated, but …
-
Instructions at https://github.com/MarvinLvn/CPC2/blob/master/docs/training_and_eval.md state that
"$PATH_ITEM_FILE is the path to the .item file containing the triplet annotations" . However, whe…
-
Please check whether this paper is about 'Voice Conversion' or not.
## article info.
- title: **Unsupervised Acoustic Unit Representation Learning for Voice Conversion
using WaveNet Auto-encoders**
…