Command Palette
Search for a command to run...
Jointly learning heterogeneous features for rgb-d activity recognition
{Jian-Guo Zhang Jian-Huang Lai Wei-Shi Zheng Jian-Fang Hu}

Abstract
In this paper, we focus on heterogeneous features learning for RGB-D activity recognition. We find that features from different channels (RGB, depth) could share some similar hidden structures, and then propose a joint learning model to simultaneously explore the shared and feature-specific components as an instance of heterogeneous multi-task learning. The proposed model formed in a unified framework is capable of: 1) jointly mining a set of subspaces with the same dimensionality to exploit latent shared features across different feature channels, 2) meanwhile, quantifying the shared and feature-specific components of features in the subspaces, and 3) transferring feature-specific intermediate transforms (i-transforms) for learning fusion of heterogeneous features across datasets. To efficiently train the joint model, a three-step iterative optimization algorithm is proposed, followed by a simple inference model. Extensive experimental results on four activity datasets have demonstrated the efficacy of the proposed method. Anew RGB-D activity dataset focusing on human-object interaction is further contributed, which presents more challenges for RGB-D activity benchmarking.
Benchmarks
| Benchmark | Methodology | Metrics |
|---|---|---|
| skeleton-based-action-recognition-on-ntu-rgbd | Dynamic Skeletons | Accuracy (CS): 60.2 Accuracy (CV): 65.2 |
| skeleton-based-action-recognition-on-ntu-rgbd-1 | Dynamic Skeletons | Accuracy (Cross-Setup): 54.7% Accuracy (Cross-Subject): 50.8% |
| skeleton-based-action-recognition-on-sysu-3d | Dynamic Skeletons | Accuracy: 75.5% |
Build AI with AI
From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.