Closed ambernorder closed 3 years ago
I do not have any real experience with Tensorflow and based on these terrible Tensorflow errors, debugging can de difficult. I do not remember having encountered this error before, so I cannot really help you.
In any case, I recommend you do not use this original BERT pretraining code but rather the Hugging Face Transformers library and their examples (https://github.com/huggingface/transformers/tree/master/examples/language-modeling). The PyTorch framework and the Transformers library are actually made to be usable by humans, so I'd recommend them in any applicable situation.
Hello,
I want to do extra pretraining on the bertje model on domain specific texts and I use the pretraining code from the original BERT code. I downloaded the model from the huggingface model hub and I need to use the .ckpt files. I cannot download the model via the code as I don't have access to the internet from where I work, so I have a folder of the bert-base-dutch-cased model.
When I try to run the pretraining code I get this error:
I can get the pretraining working with the original BERT checkpoints.
The command I use is:
python run_pretraining.py --bert_config_file="bert-base-dutch-cased/config.json" --input_file="tf_examples.tfrecord" --init_checkpoint="bert-base-dutch-cased/model.ckpt" --output_dir="output_dir" --max_seq_length=16 --max_predictions_per_seq=20 --do_train=True --do_eval=True --train_batch_size=1 --eval_batch_size=1 --learning_rate=1e-4 --num_train_steps=20 --num_warmup_steps=20 --save_checkpoints_setps=20 --iterations_per_loop=20 --max_level_steps=20
Do you maybe know what is going wrong?