Open zyk516 opened 1 month ago
Hello. I have not tried this method. Personally, I speculate that freezing the PLM during pretraining may still help the ID-based model perceive textual knowledge. However, since the PLM is not optimized, it may slightly weaken the performance of FLIP on downstream CTR tasks.
Hello I read the code&paper and find that PLM is trainable when in the pretraining stage.Do you try to freeze PLM and only train text projection layer in the pretraining stage? If you have tried, how effective is this training method? Thanks very much.