-
-
The currently followed architecture of is still too closely bound to traditional NLU based voice interaction concepts. While it aimed at including LLM with speech, LLM with multimodality, ... it is po…
-
I found that SampleRNN need to be run in parallel to get fast generation speed. It takes only about 500 seconds for generating 200 utterances, each with a length of 8 seconds speech. But it will be ve…
-
### News
- Conferences
- AAAI 2023 모두들 축하드립니다 (뒷북)
- ACL 2023: Softconf 1월 17일
- ICML 2023: Openreview 1월 26일 (한국 시간 27일 새벽 5시.. )
- CVPR 2023: 리뷰 1월 10일까지.. Aㅏ.....
- [ICML 2023에서 Chat…
-
The WaveNet paper mentioned very briefly (without any details) the possibility of using it for audio source separation. Examples of this include: extract the vocal part out of a piece of music, extra…
-
![img_2605](https://cloud.githubusercontent.com/assets/18543485/15139327/1ab91d84-16c8-11e6-92dc-879663d33fbb.PNG)
replace the text with this one, justify it left and right, and have white scroll bar…
-
# Deep visualizations
Deep visualizations is a technique to use neural networks as generative models for pictures/video. It would be cool to generate audio from a GAN but many tools already exist to …
-
### Checklist
- [ ] The issue exists after disabling all extensions
- [ ] The issue exists on a clean installation of webui
- [ ] The issue is caused by an extension, but I believe it is caused by a …
-
Hi there, 1st of all thanks for your awesome work !
Since we've "doxed" it in our [HyMPS project](https://github.com/forart/HyMPS#readme) (under the AUDIO \ [Tools page](https://github.com/forart/H…
-
First of all, congratulations on what looks like very exciting work!
I wanted to check your generative modelling application by running `autoregressive.py`, however, I had to change a couple of thi…