On-the-Fly Attention Modularization for Neural Generation
Abstract:
Despite considerable advancements with deep neural language models (LMs), neural text generation still suffers from degeneration: generated text is repetitive, generic, self-inconsistent, and lacking commonsense. The empirical analyses on sentence-level attention patterns reveal that neural text degeneration may be associated with insuf...More
Code:
Data:
Full Text
Tags
Comments