HyperAIHyperAI

Command Palette

Search for a command to run...

5 months ago

Fusion-GCN: Multimodal Action Recognition using Graph Convolutional Networks

Michael Duhme; Raphael Memmesheimer; Dietrich Paulus

Fusion-GCN: Multimodal Action Recognition using Graph Convolutional Networks

Abstract

In this paper, we present Fusion-GCN, an approach for multimodal action recognition using Graph Convolutional Networks (GCNs). Action recognition methods based around GCNs recently yielded state-of-the-art performance for skeleton-based action recognition. With Fusion-GCN, we propose to integrate various sensor data modalities into a graph that is trained using a GCN model for multi-modal action recognition. Additional sensor measurements are incorporated into the graph representation, either on a channel dimension (introducing additional node attributes) or spatial dimension (introducing new nodes). Fusion-GCN was evaluated on two public available datasets, the UTD-MHAD- and MMACT datasets, and demonstrates flexible fusion of RGB sequences, inertial measurements and skeleton sequences. Our approach gets comparable results on the UTD-MHAD dataset and improves the baseline on the large-scale MMACT dataset by a significant margin of up to 12.37% (F1-Measure) with the fusion of skeleton estimates and accelerometer measurements.

Code Repositories

mduhme/fusion-gcn
Official
pytorch

Benchmarks

BenchmarkMethodologyMetrics
multimodal-activity-recognition-on-mmactFusion-GCN
F1-Score (Cross-Subject): 89.60

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp
Fusion-GCN: Multimodal Action Recognition using Graph Convolutional Networks | Papers | HyperAI