Aria-K-Alethia / BigCodec

Official implementation of the paper "BigCodec: Pushing the Limits of Low-Bitrate Neural Speech Codec"
https://aria-k-alethia.github.io/bigcodec-demo/
MIT License
78 stars 4 forks source link

Training data used for the released checkpoint #6

Open wincing2 opened 1 week ago

wincing2 commented 1 week ago

Thank you for releasing the code and the pretrained checkpoint. On what data was the checkpoint trained and how many hours are these data? Only LibriSpeech as stated in the paper?

Aria-K-Alethia commented 1 week ago

Hi,

Only LibriSpeech, of course. Everything stated in the paper is faithfully reflected in the code and ckpt.

Liujingxiu23 commented 1 week ago

@Aria-K-Alethia I see you claim in the paper that "trained on 8 NVIDIA A100 GPUs" and "model converges in about 600k steps.". I want to know about how many days to train to achieve 600k steps ?

Aria-K-Alethia commented 1 week ago

@Aria-K-Alethia I see you claim in the paper that "trained on 8 NVIDIA A100 GPUs" and "model converges in about 600k steps.". I want to know about how many days to train to achieve 600k steps ?

I don't remember clearly but I think 2~3 days. If possible, I suggest you increase batch size to 16 or more for a faster convergence.