Open howardyclo opened 6 years ago
Neural text generation models are typically auto-regressive, trained via maximizing likelihood or perplexity and evaluate with validation perplexity. They claim such training and evaluation could result in poor sample quality (during sample generation, the model is often forced to condition on sequences that were never conditioned on at training time, leading to unpredictable dynamics in the hidden state of the RNN). Thus, they purpose to improve sample quality using actor-critic conditional GANs.
Metadata