HyperAIHyperAI

Command Palette

Search for a command to run...

5 months ago

Learning Long-Term Dependencies in Irregularly-Sampled Time Series

Mathias Lechner; Ramin Hasani

Learning Long-Term Dependencies in Irregularly-Sampled Time Series

Abstract

Recurrent neural networks (RNNs) with continuous-time hidden states are a natural fit for modeling irregularly-sampled time series. These models, however, face difficulties when the input data possess long-term dependencies. We prove that similar to standard RNNs, the underlying reason for this issue is the vanishing or exploding of the gradient during training. This phenomenon is expressed by the ordinary differential equation (ODE) representation of the hidden state, regardless of the ODE solver's choice. We provide a solution by designing a new algorithm based on the long short-term memory (LSTM) that separates its memory from its time-continuous state. This way, we encode a continuous-time dynamical flow within the RNN, allowing it to respond to inputs arriving at arbitrary time-lags while ensuring a constant error propagation through the memory path. We call these RNN models ODE-LSTMs. We experimentally show that ODE-LSTMs outperform advanced RNN-based counterparts on non-uniformly sampled data with long-term dependencies. All code and data is available at https://github.com/mlech26l/ode-lstms.

Code Repositories

mlech26l/learning-long-term-irregular-ts
Official
tf
Mentioned in GitHub
mlech26l/ode-lstms
Official
tf
Mentioned in GitHub

Benchmarks

BenchmarkMethodologyMetrics
sequential-image-classification-on-sequentialODE-LSTM
Permuted Accuracy: 97.83%

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp
Learning Long-Term Dependencies in Irregularly-Sampled Time Series | Papers | HyperAI