Command Palette
Search for a command to run...
Roei Herzig Elad Ben-Avraham Karttikeya Mangalam Amir Bar Gal Chechik Anna Rohrbach Trevor Darrell Amir Globerson

Abstract
Recently, video transformers have shown great success in video understanding, exceeding CNN performance; yet existing video transformer models do not explicitly model objects, although objects can be essential for recognizing actions. In this work, we present Object-Region Video Transformers (ORViT), an \emph{object-centric} approach that extends video transformer layers with a block that directly incorporates object representations. The key idea is to fuse object-centric representations starting from early layers and propagate them into the transformer-layers, thus affecting the spatio-temporal representations throughout the network. Our ORViT block consists of two object-level streams: appearance and dynamics. In the appearance stream, an "Object-Region Attention" module applies self-attention over the patches and \emph{object regions}. In this way, visual object regions interact with uniform patch tokens and enrich them with contextualized object information. We further model object dynamics via a separate "Object-Dynamics Module", which captures trajectory interactions, and show how to integrate the two streams. We evaluate our model on four tasks and five datasets: compositional and few-shot action recognition on SomethingElse, spatio-temporal action detection on AVA, and standard action recognition on Something-Something V2, Diving48 and Epic-Kitchen100. We show strong performance improvement across all tasks and datasets considered, demonstrating the value of a model that incorporates object representations into a transformer architecture. For code and pretrained models, visit the project page at \url{https://roeiherz.github.io/ORViT/}
Code Repositories
Benchmarks
| Benchmark | Methodology | Metrics |
|---|---|---|
| action-recognition-in-videos-on-something | ORViT Mformer (ORViT blocks) | GFLOPs: N/A Parameters: N/A Top-1 Accuracy: 67.9 Top-5 Accuracy: 90.5 |
| action-recognition-in-videos-on-something | ORViT Mformer-L (ORViT blocks) | GFLOPs: N/A Parameters: N/A Top-1 Accuracy: 69.5 Top-5 Accuracy: 91.5 |
| action-recognition-on-ava-v2-2 | ORViT MViT-B, 16x4 (K400 pretraining) | mAP: 26.6 |
| action-recognition-on-diving-48 | ORViT TimeSformer | Accuracy: 88.0 |
| action-recognition-on-epic-kitchens-100 | ORViT Mformer-L (ORViT blocks) | Action@1: 45.7 Noun@1: 58.7 Verb@1: 68.4 |
Build AI with AI
From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.