EA-LSTM: Evolutionary attention-based LSTM for time series prediction

Time series prediction with deep learning methods, especially Long Short-term Memory Neural Network (LSTM), have scored significant achievements in recent years. Despite the fact that LSTM can help to capture long-term dependencies, its ability to pay different degree of attention on sub-window feat...

Full description

Saved in:
Bibliographic Details
Published inKnowledge-based systems Vol. 181; p. 104785
Main Authors Li, Youru, Zhu, Zhenfeng, Kong, Deqiang, Han, Hua, Zhao, Yao
Format Journal Article
LanguageEnglish
Published Amsterdam Elsevier B.V 01.10.2019
Elsevier Science Ltd
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:Time series prediction with deep learning methods, especially Long Short-term Memory Neural Network (LSTM), have scored significant achievements in recent years. Despite the fact that LSTM can help to capture long-term dependencies, its ability to pay different degree of attention on sub-window feature within multiple time-steps is insufficient. To address this issue, an evolutionary attention-based LSTM training with competitive random search is proposed for multivariate time series prediction. By transferring shared parameters, an evolutionary attention learning approach is introduced to LSTM. Thus, like that for biological evolution, the pattern for importance-based attention sampling can be confirmed during temporal relationship mining. To refrain from being trapped into partial optimization like traditional gradient-based methods, an evolutionary computation inspired competitive random search method is proposed, which can well configure the parameters in the attention layer. Experimental results have illustrated that the proposed model can achieve competetive prediction performance compared with other baseline methods.
ISSN:0950-7051
1872-7409
DOI:10.1016/j.knosys.2019.05.028