On Sparsifying Encoder Outputs in Sequence-to-Sequence Models
Abstract:
Sequence-to-sequence models usually transfer all encoder outputs to the decoder for generation. In this work, by contrast, we hypothesize that these encoder outputs can be compressed to shorten the sequence delivered for decoding. We take Transformer as the testbed and introduce a layer of stochastic gates in-between the encoder and the...More
Code:
Data:
Tags
Comments