Closed YijiaXiao closed 3 years ago
Most likely yes it will be impacted by the amount of training data. Other important factors are the diversity of the training data and overlap between pretraining and the structures where you perform the downstream contact prediction.
Hi @tomsercu , thank you for your timely reply, it seems that pretraining do need large quantities of diverse data. Thank you :)
Hi, thank you for your great work! I have read the related paper <_MSA Transformer_>. In the paper, the authors mentioned that they used 26 million MSAs, which is a large amount and trained for 100k updates. I wonder does the amount of data and training steps matters a lot? If I want to pre-train one MSA transformer model on a much smaller dataset, say 1M MSAs, will the performance (e.g. contact prediction) of MSA model drop a lot? Thank you!