Command Palette
Search for a command to run...
Georgia Gkioxari; Ross Girshick; Jitendra Malik

Abstract
There are multiple cues in an image which reveal what action a person is performing. For example, a jogger has a pose that is characteristic for jogging, but the scene (e.g. road, trail) and the presence of other joggers can be an additional source of information. In this work, we exploit the simple observation that actions are accompanied by contextual cues to build a strong action recognition system. We adapt RCNN to use more than one region for classification while still maintaining the ability to localize the action. We call our system RCNN. The action-specific models and the feature maps are trained jointly, allowing for action specific representations to emerge. RCNN achieves 90.2% mean AP on the PASAL VOC Action dataset, outperforming all other approaches in the field by a significant margin. Last, we show that RCNN is not limited to action recognition. In particular, RCNN can also be used to tackle fine-grained tasks such as attribute classification. We validate this claim by reporting state-of-the-art performance on the Berkeley Attributes of People dataset.
Code Repositories
Benchmarks
| Benchmark | Methodology | Metrics |
|---|---|---|
| human-object-interaction-detection-on-hico-1 | R*CNN | mAP: 28.5 |
| weakly-supervised-object-detection-on-4 | R*CNN | MAP: 0.99 |
| weakly-supervised-object-detection-on-hico | R*CNN | MAP: 2.15 |
Build AI with AI
From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.