Closed Levantespot closed 2 years ago
The MLP layer as defined here https://github.com/lhoyer/DAFormer/blob/8d6e710700ff5e6a053c77bfe384ba44d4672cbe/mmseg/models/decode_heads/segformer_head.py#L18 flattens its input over all pixels so that the same linear layer is applied to all pixels. Therefore, it has the same behavior as a 1x1 convolution. You can also recognize that when looking at the number of in_features
.
Thanks for your response!
In 3.2 of the paper:
However, the model built from
configs/daformer/gta2cs_uda_warm_fdthings_rcs_croppl_a999_daformer_mitb5_s0.py
uses MLP instead of conv1:I‘ll be grateful to your help.