After training the mms-tts-urd_arabic-script on custom data, the final model is used for inference and works. However, the new model is not able to pronounce the words well. I thought it might be because the model is overfitted so wanted to load inference from the model checkpoints. However, it throws error with the standard huggingface pipeline code, specifically the absence of the config.json etc. On copying the last model files into the checkpoint directory the inference code works and generates an output .wav file but the file is blank with no audio. I am using the inference code provided by dunkerbunker in issue#1
After training the mms-tts-urd_arabic-script on custom data, the final model is used for inference and works. However, the new model is not able to pronounce the words well. I thought it might be because the model is overfitted so wanted to load inference from the model checkpoints. However, it throws error with the standard huggingface pipeline code, specifically the absence of the config.json etc. On copying the last model files into the checkpoint directory the inference code works and generates an output .wav file but the file is blank with no audio. I am using the inference code provided by dunkerbunker in issue#1