Closed yiwei0730 closed 1 year ago
won't hurt from more conditioning, but i can make it a hyperparameter (how many layers per film condition)
Thanks for your reply,it may be a good effect of the naturalspeech 2. Is this github code can train enough or just need more programming?
@yiwei0730 it needs more work still
@lucidrains I saw the thesis. There is a part in 4.2.-> Specifically, we use a FiLM layer [38] at every 3 WaveNet layers to fuse the condition information processed by the second Q-K-V attention in the prompting mechanism in the diffusion model.
but i saw your model use the Film layer in any layers
class WavenetResBlock(nn.Module): def init( self, dim, *, dilation, kernel_size = 3, skip_conv = False, dim_cond_mult = None ): super().init()
Is this a wrong edit or is a specific trick to run the wavenet ?