Open emanjavacas opened 7 years ago
Yes, I think we can conclude that a better memory is the single greatest challenge we still have if we ever want to arrive at novel-length generation. The simpler suggestions are nice because they give authors a more direct impression that the model learns from them. At the same time, these are also a bit hacky and bypass the true challenge.
Giphart has made a number of points during the process and we should eventually take them into account. One of the open questions is how to make the text generation system aware of longer and longer contexts. I think a good place to start is Memory networks:
More recently less ambitioned and easier to implement are variants of pointer networks, that would allow the model to pick up on recently introduced words (even if not in the vocabulary):
Especially the last one is kind of crazily effective for so little of an addition.
Of course, all these models are formulated for word-level systems...
UPDATE: (should work for both word and char-level models):