TY - JOUR
T1 - Recurrent attention unit: a new gated recurrent unit for long-term memory of important parts in sequential data
AU - Niu, Zhaoyang
AU - Zhong, Guoqiang
AU - Yue, Guohua
AU - Wang, Li-Na
AU - Yu, Hui
AU - Ling, Xiao
AU - Dong, Junyu
N1 -
Publisher Copyright:
© 2022 Elsevier B.V.
PY - 2023/1/14
Y1 - 2023/1/14
N2 - Gated recurrent unit (GRU) is a variant of the recurrent neural network (RNN). It has been widely used in many applications, such as handwriting recognition and natural language processing. However, GRU can only memorize the sequential information, but lacks the capability of adaptively paying attention to important parts in the sequences. In this paper, we propose a novel RNN model, called recurrent attention unit (RAU), which can seamlessly integrate the attention mechanism into the interior of the GRU cell by adding an attention gate. The attention gate enhances the ability of RAU to remember long-term information and pay attention to important parts in the sequential data. Extensive experiments on adding problem, image classification, sentiment classification and language modeling show that RAU consistently outperforms GRU and other related models.
AB - Gated recurrent unit (GRU) is a variant of the recurrent neural network (RNN). It has been widely used in many applications, such as handwriting recognition and natural language processing. However, GRU can only memorize the sequential information, but lacks the capability of adaptively paying attention to important parts in the sequences. In this paper, we propose a novel RNN model, called recurrent attention unit (RAU), which can seamlessly integrate the attention mechanism into the interior of the GRU cell by adding an attention gate. The attention gate enhances the ability of RAU to remember long-term information and pay attention to important parts in the sequential data. Extensive experiments on adding problem, image classification, sentiment classification and language modeling show that RAU consistently outperforms GRU and other related models.
KW - attention mechanism
KW - gated recurrent unit (GRU)
KW - memory
KW - recurrent attention unit (RAU)
KW - recurrent neural networks (RNNs)
UR - http://www.scopus.com/inward/record.url?scp=85143768575&partnerID=8YFLogxK
U2 - 10.1016/j.neucom.2022.10.050
DO - 10.1016/j.neucom.2022.10.050
M3 - Article
VL - 517
SP - 1
EP - 9
JO - Neurocomputing
JF - Neurocomputing
SN - 0925-2312
ER -