Is there a specific reason why only the first vector out of the nn.Embedding is ever used? tgt8 ... tgt64 are always zeros at this stage so you end up picking the 0-th vector for each spatial position, or in other words qe8 .. qe64 will always be filled with identical repeating values.... the values in the Embedding will obviously change over time with training but they will always be repeated throughout the qe-s
Hi,
Is there a specific reason why only the first vector out of the nn.Embedding is ever used?
tgt8
...tgt64
are always zeros at this stage so you end up picking the 0-th vector for each spatial position, or in other wordsqe8
..qe64
will always be filled with identical repeating values.... the values in the Embedding will obviously change over time with training but they will always be repeated throughout theqe
-shttps://github.com/avishkarsaha/translating-images-into-maps/blob/92b9627bef43e9a50f136c13a438a02be9ceebb2/src/model/network.py#L1150
https://github.com/avishkarsaha/translating-images-into-maps/blob/92b9627bef43e9a50f136c13a438a02be9ceebb2/src/model/network.py#L1161
After permuting
qe8
to [batch, spatial, spatial, len_embedding]: