Closed haowenke closed 5 years ago
It seems that the code works fine on parsing and text classification while same errror occurs (KeyError: 'text') in ner and pico task.
Just pushed a fix. Please let us know if you have other issues. Thanks
Thanks, @ibeltagy. Was stuck at the same problem. The fix solved it. Although I ran into another issue.
2019-04-08 18:49:05,620 - INFO - allennlp.training.trainer - Beginning training.
2019-04-08 18:49:05,620 - INFO - allennlp.training.trainer - Epoch 0/74
2019-04-08 18:49:05,620 - INFO - allennlp.training.trainer - Peak CPU memory usage MB: 3521.184
2019-04-08 18:49:05,781 - INFO - allennlp.training.trainer - GPU 0 memory usage MB: 2018
2019-04-08 18:49:05,781 - INFO - allennlp.training.trainer - GPU 1 memory usage MB: 1
2019-04-08 18:49:05,781 - INFO - allennlp.training.trainer - GPU 2 memory usage MB: 1
2019-04-08 18:49:05,781 - INFO - allennlp.training.trainer - GPU 3 memory usage MB: 1
2019-04-08 18:49:05,785 - INFO - allennlp.training.trainer - Training
0%| | 0/62 [00:00<?, ?it/s]Traceback (most recent call last):
File "/opt/conda/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/opt/conda/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/opt/conda/lib/python3.6/site-packages/allennlp/run.py", line 21, in <module>
run()
File "/opt/conda/lib/python3.6/site-packages/allennlp/run.py", line 18, in run
main(prog="allennlp")
File "/opt/conda/lib/python3.6/site-packages/allennlp/commands/__init__.py", line 101, in main
args.func(args)
File "/opt/conda/lib/python3.6/site-packages/allennlp/commands/train.py", line 103, in train_model_from_args
args.force)
File "/opt/conda/lib/python3.6/site-packages/allennlp/commands/train.py", line 136, in train_model_from_file
return train_model(params, serialization_dir, file_friendly_logging, recover, force)
File "/opt/conda/lib/python3.6/site-packages/allennlp/commands/train.py", line 204, in train_model
metrics = trainer.train()
File "/opt/conda/lib/python3.6/site-packages/allennlp/training/trainer.py", line 480, in train
train_metrics = self._train_epoch(epoch)
File "/opt/conda/lib/python3.6/site-packages/allennlp/training/trainer.py", line 322, in _train_epoch
loss = self.batch_loss(batch_group, for_training=True)
File "/opt/conda/lib/python3.6/site-packages/allennlp/training/trainer.py", line 263, in batch_loss
output_dict = self.model(**batch)
File "/opt/conda/lib/python3.6/site-packages/torch/nn/modules/module.py", line 494, in __call__
result = self.forward(*input, **kwargs)
File "/opt/conda/lib/python3.6/site-packages/allennlp/models/crf_tagger.py", line 182, in forward
embedded_text_input = self.text_field_embedder(tokens)
File "/opt/conda/lib/python3.6/site-packages/torch/nn/modules/module.py", line 494, in __call__
result = self.forward(*input, **kwargs)
File "/opt/conda/lib/python3.6/site-packages/allennlp/modules/text_field_embedders/basic_text_field_embedder.py", line 110, in forward
token_vectors = embedder(*tensors)
File "/opt/conda/lib/python3.6/site-packages/torch/nn/modules/module.py", line 494, in __call__
result = self.forward(*input, **kwargs)
File "/opt/conda/lib/python3.6/site-packages/allennlp/modules/token_embedders/bert_token_embedder.py", line 91, in forward
attention_mask=util.combine_initial_dims(input_mask))
File "/opt/conda/lib/python3.6/site-packages/torch/nn/modules/module.py", line 494, in __call__
result = self.forward(*input, **kwargs)
File "/opt/conda/lib/python3.6/site-packages/pytorch_pretrained_bert/modeling.py", line 711, in forward
embedding_output = self.embeddings(input_ids, token_type_ids)
File "/opt/conda/lib/python3.6/site-packages/torch/nn/modules/module.py", line 494, in __call__
result = self.forward(*input, **kwargs)
File "/opt/conda/lib/python3.6/site-packages/pytorch_pretrained_bert/modeling.py", line 262, in forward
embeddings = self.LayerNorm(embeddings)
File "/opt/conda/lib/python3.6/site-packages/torch/nn/modules/module.py", line 494, in __call__
result = self.forward(*input, **kwargs)
File "/opt/conda/lib/python3.6/site-packages/apex-0.1-py3.6-linux-x86_64.egg/apex/normalization/fused_layer_norm.py", line 149, in forward
input, self.weight, self.bias)
File "/opt/conda/lib/python3.6/site-packages/apex-0.1-py3.6-linux-x86_64.egg/apex/normalization/fused_layer_norm.py", line 21, in forward
input_, self.normalized_shape, weight_, bias_, self.eps)
RuntimeError: Undefined backend is not a valid device type (backendToDeviceType at /opt/conda/lib/python3.6/site-packages/torch/include/c10/core/Backend.h:141)
frame #0: c10::Error::Error(c10::SourceLocation, std::string const&) + 0x45 (0x7f446433c505 in /opt/conda/lib/python3.6/site-packages/torch/lib/libc10.so)
frame #1: <unknown function> + 0x983b (0x7f44171a183b in /opt/conda/lib/python3.6/site-packages/apex-0.1-py3.6-linux-x86_64.egg/fused_layer_norm_cuda.cpython-36m-x86_64-linux-gnu.so)
frame #2: layer_norm_affine(at::Tensor, c10::ArrayRef<long>, at::Tensor, at::Tensor, double) + 0x58 (0x7f44171a43e8 in /opt/conda/lib/python3.6/site-packages/apex-0.1-py3.6-linux-x86_64.egg/fused_layer_norm_cuda.cpython-36m-x86_64-linux-gnu.so)
frame #3: <unknown function> + 0x19583 (0x7f44171b1583 in /opt/conda/lib/python3.6/site-packages/apex-0.1-py3.6-linux-x86_64.egg/fused_layer_norm_cuda.cpython-36m-x86_64-linux-gnu.so)
frame #4: <unknown function> + 0x1969e (0x7f44171b169e in /opt/conda/lib/python3.6/site-packages/apex-0.1-py3.6-linux-x86_64.egg/fused_layer_norm_cuda.cpython-36m-x86_64-linux-gnu.so)
frame #5: <unknown function> + 0x168f2 (0x7f44171ae8f2 in /opt/conda/lib/python3.6/site-packages/apex-0.1-py3.6-linux-x86_64.egg/fused_layer_norm_cuda.cpython-36m-x86_64-linux-gnu.so)
<omitting python frames>
frame #13: THPFunction_do_forward(THPFunction*, _object*) + 0x189 (0x7f448e54d129 in /opt/conda/lib/python3.6/site-packages/torch/lib/libtorch_python.so)
Any idea on how to fix this?
Nevermind, the fix mentioned here fixed it. :) Thanks again.
Hi, First of all, thanks for this great work. I am trying to reproduce the ner task mentioned on your paper. But some key error raised when I try to run following command.
./train_allennlp_local.sh outputs