An advanced singing voice synthesis system with high fidelity, expressiveness, controllability and flexibility based on DiffSinger: Singing Voice Synthesis via Shallow Diffusion Mechanism
Refactoring the code, removing unnecessary parts, and adapting to OpenVPI/DiffSinger
The following parameters are recommended for use:
Please note that when dim>512, Layernorm will be enabled by default to ensure stability during training
LYNXNet(Linear Gated Depthwise Separable Convolution Network)
refer to: https://github.com/CNChTu/Diffusion-SVC/blob/v2.0_dev/diffusion/naive_v2/model_conformer_naive.py https://github.com/CNChTu/Diffusion-SVC/blob/v2.0_dev/diffusion/naive_v2/naive_v2_diff.py
Refactoring the code, removing unnecessary parts, and adapting to OpenVPI/DiffSinger
The following parameters are recommended for use: Please note that when dim>512, Layernorm will be enabled by default to ensure stability during training
TIPS:You can control the style of the generated results by modifying the 'activation'(LYNXNet.py,Line:129),