Open Amit-GH opened 3 years ago
I am also facing similar issue on my custom dataset. Evaluation works if the use_tpu is made false and code is run on GPU or CPU. But it takes way longer. Any thoughts on how to resolve this ?
I am also facing similar issue on my custom dataset. Evaluation works if the use_tpu is made false and code is run on GPU or CPU. But it takes way longer. Any thoughts on how to resolve this ?
Hi @prathameshk, can I ask how do you finetune the model on your custom dataset? I was thinking replace data_dir
by path_contains_tfrecords
, but I got error:
(0) Invalid argument: Feature: document (data type: string) is required but could not be found.
[[{{node ParseSingleExample/ParseExample/ParseExampleV2}}]]
[[MultiDeviceIteratorGetNextFromShard]]
[[RemoteCall]]
[[IteratorGetNext]]
[[Mean/_19475]]
Updates: I solved this problem by replacing the name_to_features fields with the actual fields in the tfrecord file.
If you haven't already, then check out the HuggingFace implementation of BigBird. That can be easier to use and integrate with your project.
I am using the script
roberta_base.sh
to train and test the model on PubMed summarization task. I am able to successfully train the model for multiple steps (5000) but it fails during evaluation time. Below is some of the error string.I am not too familiar with the code and about this error. Searched it online but didn't get much help. Hope you can help. Below is the script which I ran to reproduce this error: