Open PetrochukM opened 4 years ago
Related issue: https://github.com/magenta/ddsp/issues/12
Have you try it without a spectral convergence loss?
Am not sure if this answer is going to be helpful now. Am just sharing my understanding of SpectralConvergence Loss and it may or may not answer the question raised by you.
SpectralConvergence loss is more related to spectral loss. Spectral loss is just the L1/L2 norm over the difference in the log magnitude of the spectrogram of the signal. Whereas, SCL is Euclidean distance between magnitude of target and denoised signals normalized by euclidean length of target signal.
Since its euclidean it penalises largely, when even one of the frequency range is far apart from the target and forgiving for smaller differences.
Hi There!
I'd like to better understand spectral convergence loss. In the literature, these are the mentions I have found so far:
The above explanations for including the loss function are fairly vague and short. Furthermore, I am unable to find any mentions of a similar loss else were in literature. To better describe the loss, I searched for "relative spectral power" (f.y.i. since the spectrogram to the power of two is the "power spectrogram", and the sum of the power spectrogram is a "power spectral density").
Lastly, this paper from Google Brain "DDSP: DIFFERENTIABLE DIGITAL SIGNAL PROCESSING" https://arxiv.org/pdf/2001.04643.pdf just uses a spectrogram magnitude loss. It doesn't train with a spectral convergence loss and their results are pretty good.
Have you tried training without spectral convergence loss? Is there some more literature I am missing that validates it as a perceptual loss?