Open A11en0 opened 7 months ago
It seems to have a bug in evaluate function as shown in following:
Since it only caculate the metric of last batch in the evaluation set, it maybe alter to scores = scores.mean().item()
scores = scores.mean().item()
After weeks of experiments, I also just realized this bug. I should have read the Issue page earlier :(
It seems to have a bug in evaluate function as shown in following:
Since it only caculate the metric of last batch in the evaluation set, it maybe alter to
scores = scores.mean().item()