Closed fcchit closed 6 months ago
Sorry for the inconvenience. I process the data and divide them into 34 frames with some hardcode since it runs only once. As metioned in data_libs/README.md
, i build a new cache named final_test for metric inferring.
So you need to run the self.cache_generation
func with is_test = False
on the test dataset to divide them into 34 frames.
Thanks for your quick reply! I use is_test = False
to generate the test data cache and now everything works fine. Although is_test = False
is a little unexpected, it solves the problem :)
Hello @zyhbili, I retrained LivelySpeaker, but the results of the evaluation metrics were not good. So I reprocessed the training data, but I found that the calculation of the mean and standard deviation of bvh is missing.
We use rot6d representation without normalization(i.e (x-mu)/std). So there is no need for the mean and standard deviation when processing the data, you can simply set them to mu=0 and std 1.
I see. Thanks for your quick replies.
Good job!
I am currently training LivelySpeaker on BEAT, but when I run the test_LivelySpeaker_beat.py, I find that the test data is not split into 34 frames, so the number of text tags is too large, exceeding the limit of 77 in the clip model, resulting in an error.
My data processing process is as follows,
aux_info
aux_info
,then I run test_LivelySpeaker_beat.py again, but the upper error occurs.
Do you have any idea about this? Looking forward to you quick reply, thank you!