Closed xiang-deng closed 4 years ago
The paper reports the results on the test set, and it is normal that the val set result is different from the test set. The readme page reports some of the val set results and you can use them as a reference.
But 67.98% is 20% higher than the 5-way 1-shot val result shown on the readme (although not the same model, but proto-adv should be better than proto right?). Can you confirm the above script is correct? Thanks!
The script is correct. You are using the BERT encoder so it is natural to beat the CNN encoder by a large margin. Note that the distribution of the validation set is a lot different from the test set so the results between them may differ a lot.
Hi, I'm using the build-in evaluation function, but I'm getting some strange results. Seems all results are better than what's reported in the paper. Is there anything we should pay attention to such as the selection of Q and val_step to have the same behavior as the official evaluation script in CodaLab? Right now For proto, I can get 49.87% accuracy after 3000 steps on 5-way 1-shot, at step 30000 it reports 67.98% eval accuracy.