google-research / electra

ELECTRA: Pre-training Text Encoders as Discriminators Rather Than Generators
Apache License 2.0
2.33k stars 352 forks source link

Can you share models trained with all weights tied? #100

Open YovaKem opened 3 years ago

YovaKem commented 3 years ago

In the paper you say " On the other hand, tying all encoder weights caused little improvement while incurring the significant disadvantage of requiring the generator and discriminator to be the same size" Is it possible to share the generator and discriminator you trained to obtain this result?