Closed SuperMaximus1984 closed 4 months ago
Hi @SuperMaximus1984
The issue you have is related to the samples you are using somehow:
> Filtering invalid eval samples!!
> Total eval samples after filtering: 0
Obviously, I don't know what your samples are like, how good the quality is etc. But My best guess is that its not been able to automatically break down your ORIGINAL supplied samples that you used in step one:
[!] Warning: The text length exceeds the character limit of 250 for language 'en', this might cause truncated audio.
What I would suggest is that you delete the current training data (which you can do on the final step, there is a button there). and manually break down your original sample(s) a bit in audacity or a similar audio editing package, then place those smaller samples in the "put-voice-samples-in-here" here directory. For example, lets say your original training sample is 10 minutes long, you could break it down into 5-10 smaller samples, which will help it further break down the original sample into smaller ones when it transcribes. After that, re-run the step 1.
I have introduced code in AllTalk V2's finetuning which works around this issue, so you can also look at using the v2 BETA.
Thanks
@erew123 Thanks, I'll try V2! Btw, I deleted the tmp data, changed the model to Whisper Large-V3 and was able to run the training smoothly.
Hi! Finetuning results in the below error. The dataset (1st stage) has been created without a problem. Could you please help to solve it?