Closed Dyongh613 closed 2 years ago
self.mlp = nn.Sequential( LinearNorm(residual_channels, residual_channels 4), Mish(), # return x torch.tanh(F.softplus(x)) LinearNorm(residual_channels * 4, residual_channels) )
class Mish(nn.Module): def forward(self, x): return x * torch.tanh(F.softplus(x))
Hi @qw1260497397 , it is for the embedding of diffusion time steps as described in the paper.
Close due to inactivity.
self.mlp = nn.Sequential( LinearNorm(residual_channels, residual_channels 4), Mish(), # return x torch.tanh(F.softplus(x)) LinearNorm(residual_channels * 4, residual_channels) )
class Mish(nn.Module): def forward(self, x): return x * torch.tanh(F.softplus(x))