-
**conditional VAEで音声embeddingを獲得し、転写することで音声変換を行う。更にWGANを使ってクリアな音声を目指す。**
論文本体・著者
------------------
* Chin-Cheng Hsu, Hsin-Te Hwang, Yi-Chiao Wu, Yu Tsao, Hsin-Min Wang
* arXiv: https://arxiv.or…
-
Hi, seungwonpark,
I was trying to use Google's posted [test list](https://github.com/google/speaker-id/tree/master/publications/VoiceFilter/dataset/LibriSpeech) for LibriSpeech to reproduce their r…
-
### News
- Conferences
- Interspeech 2023 결과발표: 억셉되신 분들 모두 축하드립니다.
- NeurIPS 2023 Datasets & Benchmarks Track: 제출건수 300개 --> 984개 ㅠ_ㅠ (리뷰어, AC 모집합니다)
- [미군 AI드론, 가상훈련서 조종자 살해…'임무에 방해된다' 판단](…
-
I want to know whether we can use pretrained model of language other than english
in different Aligner with same Kaldi toolkit uses ANN as Acoustic model?
-
- https://arxiv.org/abs/2104.14830
- 2021 INTERSPEECH
多くの言語ファミリーでASR (automatic speech recognition)モデルを構築することは、大きな言語バリエーションと大きく不均衡なデータのため、マルチタスク学習の難しい問題です。
既存の研究では、高リソース言語から低リソース言語へのポジティブな移行が示されて…
e4exp updated
3 years ago
-
Hello. I trained tacotron and WaveNet on small dataset. Tacotron is about 100k, wavenet is about 130k. I use mulaw-quantize and GTA. Is this good result for small dataset or I just need train more? Or…
-
Hello,When will it be open source?
-
(metaeval_env) [rasaneno@nag07 vowel_discrimination]$ python pc_predictions_calculation/calculate_pc_predictions.py --input_features_path /scratch/gdmacr/interspeech_paper_2022/metaeval_tests/input_fe…
-
Just a general comment. Keep in mind this may be a dead end.
In particular, I am referring to the fact that snowfall has MmiTrainingGraphCompiler, while k2 does not: https://github.com/k2-fsa/snowf…
-
```
What steps will reproduce the problem?
1.install phonetisaurus on Fedora 20
1.run 'phonetisaurus-g2p' in the terminal
What is the expected output? What do you see instead?
I'm not sure what is th…