使用作者自定义的CNerTokenizer会报错__init__() got an unexpected keyword argument 'max_len'
具体错误信息如下:
` File "BERT-NER-Pytorch-master/run_ner_softmax.py", line 549, in
main()
File "BERT-NER-Pytorch-master/run_ner_softmax.py", line 480, in main
cache_dir=args.cache_dir if args.cache_dir else None,)
File "BERT-NER-Pytorch-master\models\transformers\tokenization_utils.py", line 282, in from_pretrained
return cls._from_pretrained(*inputs, **kwargs)
File "BERT-NER-Pytorch-master\models\transformers\tokenization_utils.py", line 411, in _from_pretrained
tokenizer = cls(*init_inputs, **init_kwargs)
TypeError: init() got an unexpected keyword argument 'max_len'`
使用作者自定义的CNerTokenizer会报错
__init__() got an unexpected keyword argument 'max_len'
具体错误信息如下: ` File "BERT-NER-Pytorch-master/run_ner_softmax.py", line 549, inFile "BERT-NER-Pytorch-master/run_ner_softmax.py", line 480, in main
File "BERT-NER-Pytorch-master\models\transformers\tokenization_utils.py", line 282, in from_pretrained
File "BERT-NER-Pytorch-master\models\transformers\tokenization_utils.py", line 411, in _from_pretrained
TypeError: init() got an unexpected keyword argument 'max_len'`
P.S. 使用BertTokenizer不会报错。还想请问下作者为什么要自定义分词器呢?难道BertTokenizer不会将没有在词表中的单词转化为
<UNK>
吗?