HyperAIHyperAI

Command Palette

Search for a command to run...

3 months ago

Unified Training of Universal Time Series Forecasting Transformers

Gerald Woo Chenghao Liu Akshat Kumar Caiming Xiong Silvio Savarese Doyen Sahoo

Unified Training of Universal Time Series Forecasting Transformers

Abstract

Deep learning for time series forecasting has traditionally operated within a one-model-per-dataset framework, limiting its potential to leverage the game-changing impact of large pre-trained models. The concept of universal forecasting, emerging from pre-training on a vast collection of time series datasets, envisions a single Large Time Series Model capable of addressing diverse downstream forecasting tasks. However, constructing such a model poses unique challenges specific to time series data: i) cross-frequency learning, ii) accommodating an arbitrary number of variates for multivariate time series, and iii) addressing the varying distributional properties inherent in large-scale data. To address these challenges, we present novel enhancements to the conventional time series Transformer architecture, resulting in our proposed Masked Encoder-based Universal Time Series Forecasting Transformer (Moirai). Trained on our newly introduced Large-scale Open Time Series Archive (LOTSA) featuring over 27B observations across nine domains, Moirai achieves competitive or superior performance as a zero-shot forecaster when compared to full-shot models. Code, data, and model weights can be found at https://github.com/SalesforceAIResearch/uni2ts.

Code Repositories

SalesforceAIResearch/uni2ts
Official
jax
Mentioned in GitHub

Benchmarks

BenchmarkMethodologyMetrics
time-series-forecasting-on-etth1-336-1MOIRAISmall
MAE: 0.429
MSE: 0.412
time-series-forecasting-on-etth1-336-1MOIRAIBase
MAE: 0.450
MSE: 0.456
time-series-forecasting-on-etth1-336-1MOIRAILarge
MAE: 0.474
MSE: 0.514

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp
Unified Training of Universal Time Series Forecasting Transformers | Papers | HyperAI