Aria-K-Alethia / BigCodec

Official implementation of the paper "BigCodec: Pushing the Limits of Low-Bitrate Neural Speech Codec"
https://aria-k-alethia.github.io/bigcodec-demo/
MIT License
86 stars 4 forks source link

Training data used for the released checkpoint #6

Open wincing2 opened 1 month ago

wincing2 commented 1 month ago

Thank you for releasing the code and the pretrained checkpoint. On what data was the checkpoint trained and how many hours are these data? Only LibriSpeech as stated in the paper?

Aria-K-Alethia commented 1 month ago

Hi,

Only LibriSpeech, of course. Everything stated in the paper is faithfully reflected in the code and ckpt.

Liujingxiu23 commented 1 month ago

@Aria-K-Alethia I see you claim in the paper that "trained on 8 NVIDIA A100 GPUs" and "model converges in about 600k steps.". I want to know about how many days to train to achieve 600k steps ?

Aria-K-Alethia commented 1 month ago

@Aria-K-Alethia I see you claim in the paper that "trained on 8 NVIDIA A100 GPUs" and "model converges in about 600k steps.". I want to know about how many days to train to achieve 600k steps ?

I don't remember clearly but I think 2~3 days. If possible, I suggest you increase batch size to 16 or more for a faster convergence.