howardyclo / papernotes

My personal notes and surveys on DL, CV and NLP papers.
128 stars 6 forks source link

On Distinguishability Criteria for Estimating Generative Models #25

Open howardyclo opened 6 years ago

howardyclo commented 6 years ago

Metadata

howardyclo commented 6 years ago

Motivation to read this paper

After I read the paper "Adversarial Contrastive Estimation" (#23), which replaces the original fixed noise generator in noise contrastive estimation (NCE) with the dynamic noise generator using with GAN training, some questions like "How does NCE relate to GANs?", "NCE is closely related to MLE, and how about GANs?" naturally rises in my mind.

This paper compares MLE, NCE, GAN and gives several initial answers to:

In conclusion, the analysis shows that GANs are not as closely related to NCE as previously believed.

Notes:

Comparison (from NIPS 2016 Tutorial: Generative Adversarial Networks or watch video from 1:00:17)

Sorry for the inconsistent notation.

Similarities

Note: Asymptotically consistent estimator: See https://en.wikipedia.org/wiki/Consistent_estimator

Different p_{c}

Different goals

Different training objectives

Different stationary points when converges

Different convergence properties

NCE can implement MLE: Self-Contrastive Estimation (SCE)

Note: There is a error in the derivation of SCE's expected gradent: The equation 1/2 E_{x~p_g} log (p_g(x)) should be 1/2 E_{x~p_g} ∂/∂θ log (p_g(x)).

GANs cannot implement MLE

See derivation in the paper.

Reference