allenai / scibert

A BERT model for scientific text.
https://arxiv.org/abs/1903.10676
Apache License 2.0
1.52k stars 217 forks source link

gradient_accumulation_batch_size missing in trainer #65

Open carbonz0 opened 5 years ago

carbonz0 commented 5 years ago

I found gradient_accumulation_batch_size exists in several scibert conf, such as https://github.com/allenai/scibert/blob/8562a120e6788dcbadbe05ef7fd4463dee17ee59/allennlp_config/ner.json but allennlp trainer doesn't have this param, https://github.com/allenai/allennlp/blob/master/allennlp/training/trainer.py

ibeltagy commented 5 years ago

Yes, AllenNLP doesn't support gradient accumulation. We have it implemented in our fork of allennlp (check requirements: https://github.com/allenai/scibert/blob/master/requirements.txt)

carbonz0 commented 5 years ago

got it, thank you!

xiaoruijiang commented 2 years ago

Hi, The installation command in the readme returned the following "allennlp version not found error". Can I ask what is the right way to install the tool and reproduce the results?

Thank you very much!

Obtaining allennlp from git+git://github.com/ibeltagy/allennlp@fp16_and_others#egg=allennlp (from -r requirements.txt (line 1)) WARNING: Discarding git+git://github.com/ibeltagy/allennlp@fp16_and_others#egg=allennlp. Command errored out with exit status 128: git rev-parse HEAD Check the logs for full command output. ERROR: Could not find a version that satisfies the requirement allennlp (unavailable) ERROR: No matching distribution found for allennlp (unavailable)