Morizeyao / GPT2-Chinese

Chinese version of GPT2 training code, using BERT tokenizer.
MIT License
7.48k stars 1.7k forks source link

预训练模型的名字什么鬼 #296

Open AlisonDexter opened 6 months ago

AlisonDexter commented 6 months ago

ile "d:\Bert_pre\GPT_2\GPT2-Chinese-old_gpt_2_chinese_before_2021_4_22\GPT2-Chinese-old_gpt_2_chinese_before_2021_4_22\tokenizations\tokenization_bert.py", line 131, in init "model use tokenizer = BertTokenizer.from_pretrained(PRETRAINED_MODEL_NAME)".format(vocab_file))
ValueError: Can't find a vocabulary file at path 'cache/vocab_small.txt'. To load the vocabulary from a Google pretrained model use tokenizer = BertTokenizer.from_pretrained(PRETRAINED_MODEL_NAME) 这个报错是什么