kweonwooj / papers

summary of ML papers I've read
319 stars 34 forks source link

Pay Less Attention with Lightweight and Dynamic Convolutions #120

Open kweonwooj opened 5 years ago

kweonwooj commented 5 years ago

Abstract

Details

Background

Results

Personal Thoughts

Link : https://openreview.net/pdf?id=SkVhlh09tX Authors : Wu et al. 2018

demdecuong commented 4 years ago

As i guess, We have GLU to expand the dimension into nx2d then we go to conv to rescale it into nxd right ? I still dont understand how to apply the softmax in LightweightConv . We would softmax all the kernel weight of the Conv layer right ?

Moreover, i am not clear about the weight-sharing of the author since i try to re-implement this architecture.

Please give me some explanation .

Thank you so much.