Closed VJJJJJJ1 closed 1 week ago
hi, thank you for your great work, I have some questions:
- As you mentioned in the paper:
adding the attention module before the ConvNext module appears to be the optimal solution.
,but I found that indecoder/models.py
, the AttnBlock is contained in pos_net which is after convnext. They seem to be in the opposite order as in the paper.- I want to do streaming inference with wavtokenizer, I have replaced all the convolution layers in SEANetEncoder, SEANetDecoder, ConvNeXtBlock and pos_net(ResnetBlock+AttnBlock) with causal convolution layers(class SConv1d with causal=T). Unfortunately, the genetrator loss keeps increasing(see in the pic). Is there any wrong in the modified model?
thank you for your reply!
hi, thank you for your great work, I have some questions:
adding the attention module before the ConvNext module appears to be the optimal solution.
,but I found that indecoder/models.py
, the AttnBlock is contained in pos_net which is after convnext. They seem to be in the opposite order as in the paper.thank you for your reply!