Command Palette
Search for a command to run...
Voigtlaender Paul ; Krause Michael ; Osep Aljosa ; Luiten Jonathon ; Sekar Berin Balachandar Gnana ; Geiger Andreas ; Leibe Bastian

Abstract
This paper extends the popular task of multi-object tracking to multi-objecttracking and segmentation (MOTS). Towards this goal, we create densepixel-level annotations for two existing tracking datasets using asemi-automatic annotation procedure. Our new annotations comprise 65,213 pixelmasks for 977 distinct objects (cars and pedestrians) in 10,870 video frames.For evaluation, we extend existing multi-object tracking metrics to this newtask. Moreover, we propose a new baseline method which jointly addressesdetection, tracking, and segmentation with a single convolutional network. Wedemonstrate the value of our datasets by achieving improvements in performancewhen training on MOTS annotations. We believe that our datasets, metrics andbaseline will become a valuable resource towards developing multi-objecttracking approaches that go beyond 2D bounding boxes. We make our annotations,code, and models available at https://www.vision.rwth-aachen.de/page/mots.
Benchmarks
| Benchmark | Methodology | Metrics |
|---|---|---|
| multi-object-tracking-on-mots20 | Track R-CNN | IDF1: 42.4 sMOTSA: 40.6 |
| multiple-object-tracking-on-kitti-test-online | MOSTFusion | MOTA: 84.83 |
Build AI with AI
From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.