Context-aware positional representation for self-attention networks

K Chen, R Wang, M Utiyama, E Sumita - Neurocomputing, 2021 - Elsevier
Neurocomputing, 2021Elsevier
In self-attention networks (SANs), positional embeddings are used to model order
dependencies between words in the input sentence and are added with word embeddings
to gain an input representation, which enables the SAN-based neural model to perform
(multi-head) and to stack (multi-layer) self-attentive functions in parallel to learn the
representation of the input sentence. However, this input representation only involves static
order dependencies based on discrete position indexes of words, that is, is independent of …
Abstract
In self-attention networks (SANs), positional embeddings are used to model order dependencies between words in the input sentence and are added with word embeddings to gain an input representation, which enables the SAN-based neural model to perform (multi-head) and to stack (multi-layer) self-attentive functions in parallel to learn the representation of the input sentence. However, this input representation only involves static order dependencies based on discrete position indexes of words, that is, is independent of context information, which may be weak in modeling the input sentence. To address this issue, we proposed a novel positional representation method to model order dependencies based on n-gram context or sentence context in the input sentence, which allows SANs to learn a more effective sentence representation. To validate the effectiveness of the proposed method, it is applied to the neural machine translation model, which adopts a typical SAN-based neural model. Experimental results on two widely used translation tasks, i.e., WMT14 English-to-German and WMT17 Chinese-to-English, showed that the proposed approach can significantly improve the translation performance over the strong Transformer baseline.
Elsevier
Showing the best result for this search. See all results