ymcui / Chinese-BERT-wwm

Pre-Training with Whole Word Masking for Chinese BERT(中文BERT-wwm系列模型)
https://ieeexplore.ieee.org/document/9599397
Apache License 2.0
9.67k stars 1.39k forks source link

为什么RoBERTa版本训练速度更慢? #189

Closed sunyilgdx closed 3 years ago

sunyilgdx commented 3 years ago

老师您好,有一个不太成熟的问题,通过一段时间的简单使用,发现在相同的参数设置下(TensorFlow)RoBERTa版本的训练和推理速度明显慢于普通的BERT版本,不知道这是不是我使用时的问题,还是确实存在这方面的现象?

ymcui commented 3 years ago

你好,你是应用在什么任务呢?具体使用的是哪个模型?大概慢多少?

stale[bot] commented 3 years ago

This issue has been automatically marked as stale because it has not had recent activity. It will be closed if no further activity occurs. Thank you for your contributions.

stale[bot] commented 3 years ago

Closing the issue, since no updates observed. Feel free to re-open if you need any further assistance.