Papers
arxiv:2009.09672
Alleviating the Inequality of Attention Heads for Neural Machine Translation
Published on Sep 21, 2020
Authors:
Abstract
Recent studies show that the attention heads in Transformer are not equal. We relate this phenomenon to the imbalance training of multi-head attention and the model dependence on specific heads. To tackle this problem, we propose a simple masking method: HeadMask, in two specific ways. Experiments show that translation improvements are achieved on multiple language pairs. Subsequent empirical analyses also support our assumption and confirm the effectiveness of the method.
Models citing this paper 0
No model linking this paper
Cite arxiv.org/abs/2009.09672 in a model README.md to link it from this page.
Datasets citing this paper 0
No dataset linking this paper
Cite arxiv.org/abs/2009.09672 in a dataset README.md to link it from this page.
Spaces citing this paper 0
No Space linking this paper
Cite arxiv.org/abs/2009.09672 in a Space README.md to link it from this page.