Closed Shiva-Maruth-Alapati closed 3 years ago
Hi @Shiva-Alapati-1998 can you share a colab or script to reproduce the problem? The problem might be that there is a threshold after which the unsupervised float loss is not computed anymore. See the answer loss cutoff argument described in the paper https://arxiv.org/pdf/2004.02349.pdf
Potentially changing the answer_loss_cutoff parameter to a larger number can help in your use case.
Closing for the time being, please feel free to reopen if needed
Hello,
I am finetuning the tapas model with the COVID-19 dataset. When I am giving the 'encoding["float_answer"]=[[28756489.0],[10114.0],[2449]]' the model is always showing a training loss of 0, but when I remove one of the answers i.e encoding["float_answer"]=[[np.nan],[10114.0],[2449]] the model is getting trained. Initially, I thought the model is predicting it accurately so the loss is zero but it is not the case.