Neural Text Degeneration

  • The Curious Case of Neural Text Degeneration
  • deep analysis into the properties of the most common decoding methods for open-ended language generation
  • surprising distributional differences between human text and machine text
  • decoding strategies alone can dramatically effect the quality of machine text, even when generated from exactly the same neural language model
  • likelihood maximizing decoding causes repetition and overly generic language usage
  • sampling methods without truncation risk sampling from the low-confidence tail of a model’s predicted distribution
  • Nucleus Sampling