MLCommons Algorithmic Efficiency is a benchmark and competition measuring neural network training speedups due to algorithmic improvements in both training algorithms and models.
For this PR, I added two assertions to validate the timing and evaluation consistency at the end of the training loop:
1) Duration consistency check: The total duration of training (submission time + logging time + evaluation time) is asserted to match the final evaluation result's total_duration, with a tolerance of 10 seconds.
2) Evaluation frequency check: Ensures the number of evaluations does not exceed the allowed evaluations based on the accumulated submission time and the workload's evaluation period.
For this PR, I added two assertions to validate the timing and evaluation consistency at the end of the training loop:
1) Duration consistency check: The total duration of training (submission time + logging time + evaluation time) is asserted to match the final evaluation result's total_duration, with a tolerance of 10 seconds.
2) Evaluation frequency check: Ensures the number of evaluations does not exceed the allowed evaluations based on the accumulated submission time and the workload's evaluation period.