Comparing with the planning evaluation numbers of table 1 in your paper, trained mini or small models from my side give lower evaluation numbers as shown in the below table.
I use the data which you shared in the following link.
I execute my training work through the first command example in To train and evaluate during training: and I checked that my training arguments are identical to your training arguments.
Are any additional steps required to re-produce your numbers? (Such as use pre-trained weights, data augmentation/processing or others.)
Comparing with the planning evaluation numbers of table 1 in your paper, trained mini or small models from my side give lower evaluation numbers as shown in the below table.
I use the data which you shared in the following link. I execute my training work through the first command example in To train and evaluate during training: and I checked that my training arguments are identical to your training arguments.
Are any additional steps required to re-produce your numbers? (Such as use pre-trained weights, data augmentation/processing or others.)
<html xmlns:m="http://schemas.microsoft.com/office/2004/12/omml" xmlns="http://www.w3.org/TR/REC-html40">
| 8sADE | 3sFDE | 5sFDE | 8sFDE | MR -- | -- | -- | -- | -- | -- Paper - Mini | 2.07 | 1.20 | 2.43 | 5.14 | 0.067 Re-prod - Mini | 3.12 | 1.93 | 3.86 | 6.99 | 0.162 Paper - Small | 1.91 | 1.05 | 2.22 | 4.83 | 0.049 Re-prod - Small | 2.08 | 1.19 | 2.47 | 5.05 | 0.070