Memory-based Transformer with shorter window and longer horizon for multivariate time series forecasting

•SWLHT obtains longer prediction length via a shorter input length.•An efficient memory module can capture the correlation between input segments.•A multi-segment forecasting method is proposed to reduce the accumulation of errors.•A linear autoregressive module is integrated to improve the robustne...

Full description

Saved in:
Bibliographic Details
Published inPattern recognition letters Vol. 160; pp. 26 - 33
Main Authors Liu, Yang, Wang, Zheng, Yu, Xinyang, Chen, Xin, Sun, Meijun
Format Journal Article
LanguageEnglish
Published Elsevier B.V 01.08.2022
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:•SWLHT obtains longer prediction length via a shorter input length.•An efficient memory module can capture the correlation between input segments.•A multi-segment forecasting method is proposed to reduce the accumulation of errors.•A linear autoregressive module is integrated to improve the robustness of the model. Multivariate time series forecasting is an important problem that spans many fields. One challenge of this problem is the complex and non-linear interdependence between time steps and different variables. Recent studies have shown that Transformer has potential in capturing long-term dependencies. However, in the field of time series forecasting, Transformer still has some problems to solve, such as prediction fragmentation and insensitivity to data scale. In addition, traditional forecasting models often require a large amount of input data to support the training of the model when predicting long-term data. However, it is hard to provide sufficient time series input data due to equipment damage or weather situation. To solve these limitations, a memory-based Transformer with shorter window and longer horizon is proposed, called SWLHT. It uses the memory mechanism to make the model no longer only rely on a single input, but can combine the previous forecast results to assist in capturing long-term dependencies, thereby avoiding the requirement of excessively long input sequence. Furthermore, the memory mechanism can alleviate the prediction fragmentation to some extent. The experimental results and comparison of baselines on several real-world multivariate time series datasets have verified the effectiveness of the proposed model.
ISSN:0167-8655
1872-7344
DOI:10.1016/j.patrec.2022.05.010