HyperAIHyperAI

Command Palette

Search for a command to run...

3 months ago

TSMixer: Lightweight MLP-Mixer Model for Multivariate Time Series Forecasting

Vijay Ekambaram Arindam Jati Nam Nguyen Phanwadee Sinthong Jayant Kalagnanam

TSMixer: Lightweight MLP-Mixer Model for Multivariate Time Series Forecasting

Abstract

Transformers have gained popularity in time series forecasting for their ability to capture long-sequence interactions. However, their high memory and computing requirements pose a critical bottleneck for long-term forecasting. To address this, we propose TSMixer, a lightweight neural architecture exclusively composed of multi-layer perceptron (MLP) modules for multivariate forecasting and representation learning on patched time series. Inspired by MLP-Mixer's success in computer vision, we adapt it for time series, addressing challenges and introducing validated components for enhanced accuracy. This includes a novel design paradigm of attaching online reconciliation heads to the MLP-Mixer backbone, for explicitly modeling the time-series properties such as hierarchy and channel-correlations. We also propose a novel Hybrid channel modeling and infusion of a simple gating approach to effectively handle noisy channel interactions and generalization across diverse datasets. By incorporating these lightweight components, we significantly enhance the learning capability of simple MLP structures, outperforming complex Transformer models with minimal computing usage. Moreover, TSMixer's modular design enables compatibility with both supervised and masked self-supervised learning methods, making it a promising building block for time-series Foundation Models. TSMixer outperforms state-of-the-art MLP and Transformer models in forecasting by a considerable margin of 8-60%. It also outperforms the latest strong benchmarks of Patch-Transformer models (by 1-2%) with a significant reduction in memory and runtime (2-3X). The source code of our model is officially released as PatchTSMixer in the HuggingFace. Model: https://huggingface.co/docs/transformers/main/en/model_doc/patchtsmixer Examples: https://github.com/ibm/tsfm/#notebooks-links

Code Repositories

ibm/tsfm
Official
pytorch

Benchmarks

BenchmarkMethodologyMetrics
multivariate-time-series-forecasting-on-etth1TSMixer
MAE: 0.398
MSE: 0.368
multivariate-time-series-forecasting-on-etth1-1TSMixer
MAE: 0.418
MSE: 0.399
multivariate-time-series-forecasting-on-etth1-2TSMixer
MSE: 0.421
multivariate-time-series-forecasting-on-etth1-3TSMixer
MSE: 0.444
time-series-forecasting-on-electricity-336TSMixer
MSE: 0.158
time-series-forecasting-on-electricity-96TSMixer
MSE: 0.129
time-series-forecasting-on-etth1-192-1TSMixer
MAE: 0.418
MSE: 0.399
time-series-forecasting-on-etth1-336-1TSMixer
MAE: 0.436
MSE: 0.421
time-series-forecasting-on-etth1-720-1TSMixer
MAE: 0.467
MSE: 0.444
time-series-forecasting-on-etth1-96-1TSMixer
MAE: 0.398
MSE: 0.368
time-series-forecasting-on-etth1-96-4TSMixer
MAE: 0.398
MSE: 0.368
time-series-forecasting-on-etth2-192-1TSMixer
MAE: 0.374
MSE: 0.33
time-series-forecasting-on-etth2-336-1TSMixer
MAE: 0.401
MSE: 0.357
time-series-forecasting-on-etth2-720-1TSMixer
MAE: 0.436
MSE: 0.395
time-series-forecasting-on-etth2-96-1TSMixer
MAE: 0.337
MSE: 0.276
time-series-forecasting-on-ettm1-192-1TSMixer
MAE: 0.369
MSE: 0.333
time-series-forecasting-on-ettm1-336-1TSMixer
MAE: 0.385
MSE: 0.365
time-series-forecasting-on-ettm1-720-1TSMixer
MAE: 0.413
MSE: 0.416
time-series-forecasting-on-ettm1-96-1TSMixer
MAE: 0.346
MSE: 0.291
time-series-forecasting-on-ettm2-192-1TSMixer
MAE: 0.293
MSE: 0.219
time-series-forecasting-on-ettm2-336-1TSMixer
MAE: 0.329
MSE: 0.273
time-series-forecasting-on-ettm2-720-1TSMixer
MAE: 0.38
MSE: 0.358
time-series-forecasting-on-ettm2-96-1TSMixer
MAE: 0.255
MSE: 0.164
time-series-forecasting-on-traffic-192TSMixer
MSE: 0.377
time-series-forecasting-on-traffic-336TSMixer
MSE: 0.385
time-series-forecasting-on-traffic-720TSMixer
MSE: 0.424
time-series-forecasting-on-traffic-96TSMixer
MSE: 0.356
time-series-forecasting-on-weather-192TSMixer
MAE: 0.240
MSE: 0.191
time-series-forecasting-on-weather-336TSMixer
MAE: 0.279
MSE: 0.243
time-series-forecasting-on-weather-720TSMixer
MAE: 0.333
MSE: 0.316
time-series-forecasting-on-weather-96TSMixer
MAE: 0.197
MSE: 0.146

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp
TSMixer: Lightweight MLP-Mixer Model for Multivariate Time Series Forecasting | Papers | HyperAI