HyperAIHyperAI

Command Palette

Search for a command to run...

5 months ago

Informer: Beyond Efficient Transformer for Long Sequence Time-Series Forecasting

Haoyi Zhou; Shanghang Zhang; Jieqi Peng; Shuai Zhang; Jianxin Li; Hui Xiong; Wancai Zhang

Informer: Beyond Efficient Transformer for Long Sequence Time-Series Forecasting

Abstract

Many real-world applications require the prediction of long sequence time-series, such as electricity consumption planning. Long sequence time-series forecasting (LSTF) demands a high prediction capacity of the model, which is the ability to capture precise long-range dependency coupling between output and input efficiently. Recent studies have shown the potential of Transformer to increase the prediction capacity. However, there are several severe issues with Transformer that prevent it from being directly applicable to LSTF, including quadratic time complexity, high memory usage, and inherent limitation of the encoder-decoder architecture. To address these issues, we design an efficient transformer-based model for LSTF, named Informer, with three distinctive characteristics: (i) a $ProbSparse$ self-attention mechanism, which achieves $O(L \log L)$ in time complexity and memory usage, and has comparable performance on sequences' dependency alignment. (ii) the self-attention distilling highlights dominating attention by halving cascading layer input, and efficiently handles extreme long input sequences. (iii) the generative style decoder, while conceptually simple, predicts the long time-series sequences at one forward operation rather than a step-by-step way, which drastically improves the inference speed of long-sequence predictions. Extensive experiments on four large-scale datasets demonstrate that Informer significantly outperforms existing methods and provides a new solution to the LSTF problem.

Benchmarks

BenchmarkMethodologyMetrics
time-series-forecasting-on-etth1-168-1Informer
MAE: 0.722
MSE: 0.878
time-series-forecasting-on-etth1-168-2Informer
MAE: 0.337
MSE: 0.183
time-series-forecasting-on-etth1-24-1Informer
MAE: 0.523
MSE: 0.509
time-series-forecasting-on-etth1-24-2Informer
MAE: 0.152
MSE: 0.046
time-series-forecasting-on-etth1-336-1Informer
MAE: 0.753
MSE: 0.884
time-series-forecasting-on-etth1-336-2Informer
MAE: 0.346
MSE: 0.189
time-series-forecasting-on-etth1-48-1Informer
MAE: 0.563
MSE: 0.551
time-series-forecasting-on-etth1-48-2Informer
MAE: 0.274
MSE: 0.129
time-series-forecasting-on-etth1-720-1Informer
MAE: 0.768
MSE: 0.941
time-series-forecasting-on-etth1-720-2Informer
MAE: 0.357
MSE: 0.201
time-series-forecasting-on-etth2-168-1Informer
MAE: 0.996
MSE: 1.512
time-series-forecasting-on-etth2-168-2Informer
MAE: 0.306
MSE: 0.154
time-series-forecasting-on-etth2-24-1Informer
MAE: 0.523
MSE: 0.446
time-series-forecasting-on-etth2-24-2Informer
MAE: 0.213
MSE: 0.083
time-series-forecasting-on-etth2-336-1Informer
MAE: 1.035
MSE: 1.665
time-series-forecasting-on-etth2-336-2Informer
MAE: 0.323
MSE: 0.166
time-series-forecasting-on-etth2-48-1Informer
MAE: 0.733
MSE: 0.934
time-series-forecasting-on-etth2-48-2Informer
MAE: 0.249
MSE: 0.111
time-series-forecasting-on-etth2-720-1Informer
MAE: 1.209
MSE: 2.34
time-series-forecasting-on-etth2-720-2Informer
MAE: 0.338
MSE: 0.181

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp
Informer: Beyond Efficient Transformer for Long Sequence Time-Series Forecasting | Papers | HyperAI