Open elk-april opened 3 years ago
Hi, I noticed that it: your code:
x = self.positional_embedding(x) # b,gh*gw+1,d x = self.transformer(x) # b,gh*gw+1,d
Vision Transformer(from https://github.com/lucidrains/vit-pytorch/blob/main/vit_pytorch/vit.py):
x += self.pos_embedding[:, :(n + 1)] x = self.dropout(x) x = self.transformer(x)
Actually, there are two differences:
Could you please tell me the reasons for these changes? Looking forward to your reply, thanks very much.
Hi, I noticed that it: your code:
Vision Transformer(from https://github.com/lucidrains/vit-pytorch/blob/main/vit_pytorch/vit.py):
Actually, there are two differences:
Could you please tell me the reasons for these changes? Looking forward to your reply, thanks very much.