Closed samsara-ku closed 5 months ago
Is it okay to set just num_layers as 1 for TransformerDecoderLayer_w_Adapter class when training your model?
num_layers
TransformerDecoderLayer_w_Adapter
Some other models just use their num_layers as 6 or 8, is this just a simple curiosity.
We follow the configuration in FaceFormer, which uses just one layer of Transformer. You can add more layers, which can improve performance.
Is it okay to set just
num_layers
as 1 forTransformerDecoderLayer_w_Adapter
class when training your model?Some other models just use their
num_layers
as 6 or 8, is this just a simple curiosity.