HyperAIHyperAI

Command Palette

Search for a command to run...

4 months ago

Towards Universal Representation for Unseen Action Recognition

Yi Zhu; Yang Long; Yu Guan; Shawn Newsam; Ling Shao

Towards Universal Representation for Unseen Action Recognition

Abstract

Unseen Action Recognition (UAR) aims to recognise novel action categories without training examples. While previous methods focus on inner-dataset seen/unseen splits, this paper proposes a pipeline using a large-scale training source to achieve a Universal Representation (UR) that can generalise to a more realistic Cross-Dataset UAR (CD-UAR) scenario. We first address UAR as a Generalised Multiple-Instance Learning (GMIL) problem and discover 'building-blocks' from the large-scale ActivityNet dataset using distribution kernels. Essential visual and semantic components are preserved in a shared space to achieve the UR that can efficiently generalise to new datasets. Predicted UR exemplars can be improved by a simple semantic adaptation, and then an unseen action can be directly recognised using UR during the test. Without further training, extensive experiments manifest significant improvements over the UCF101 and HMDB51 benchmarks.

Benchmarks

BenchmarkMethodologyMetrics
action-recognition-in-videos-on-activitynetCD-UAR
mAP: 53.8
action-recognition-in-videos-on-hmdb-51CD-UAR
Average accuracy of 3 splits: 51.8
action-recognition-in-videos-on-ucf101CD-UAR
3-fold Accuracy: 42.5
zero-shot-action-recognition-on-hmdb51UR
Top-1 Accuracy: 24.4
zero-shot-action-recognition-on-ucf101UR
Top-1 Accuracy: 17.5

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp
Towards Universal Representation for Unseen Action Recognition | Papers | HyperAI