I am using code now, and i wonder is there implemented add norm? I only find layer norm, but no add operation.
Here is code in h-transformer-1d.py line 489 ...
Is this a bug or something ? Thanks @Lucidrains
for ind in range(depth):
attn = attn_class(dim, dim_head = dim_head, heads = heads, block_size = block_size, pos_emb = self.pos_emb, **attn_kwargs)
ff = FeedForward(dim, mult = ff_mult)
I am using code now, and i wonder is there implemented add norm? I only find layer norm, but no add operation. Here is code in h-transformer-1d.py line 489 ... Is this a bug or something ? Thanks @Lucidrains
for ind in range(depth): attn = attn_class(dim, dim_head = dim_head, heads = heads, block_size = block_size, pos_emb = self.pos_emb, **attn_kwargs) ff = FeedForward(dim, mult = ff_mult)