Closed manmay-nakhashi closed 2 months ago
to test or train just run by default it's using MushanW/GLOBE dataset 23,519 speakers and covers 164 accents
python3 train_e2.py
Note: need to add a path to vocab.json
@manmay-nakhashi Manmay! i remember you now from the natural speech work we did together some time ago
thanks for the PR! I will check it out tomorrow morning 😄
@manmay-nakhashi hey, looks good! 😄 do you want to try pulling and integrating the text as well?
Sure I'll do that.
@lucidrains it's ready
i'll write a inference script next so we can do some quick experiments.
nice! it looks good, but in the paper, they didn't use a tokenizer and just went character level
i was thinking we could just use utf character ids? (could remove the tokenizer and vocab.json
altogether) keep it simple
@lucidrains changes are done
@lucidrains resolved all the suggestions
@manmay-nakhashi thank you Manmay!