Command Palette
Search for a command to run...
Jia-Ren Chang; Yong-Sheng Chen

Abstract
Recent work has shown that depth estimation from a stereo pair of images can be formulated as a supervised learning task to be resolved with convolutional neural networks (CNNs). However, current architectures rely on patch-based Siamese networks, lacking the means to exploit context information for finding correspondence in illposed regions. To tackle this problem, we propose PSMNet, a pyramid stereo matching network consisting of two main modules: spatial pyramid pooling and 3D CNN. The spatial pyramid pooling module takes advantage of the capacity of global context information by aggregating context in different scales and locations to form a cost volume. The 3D CNN learns to regularize cost volume using stacked multiple hourglass networks in conjunction with intermediate supervision. The proposed approach was evaluated on several benchmark datasets. Our method ranked first in the KITTI 2012 and 2015 leaderboards before March 18, 2018. The codes of PSMNet are available at: https://github.com/JiaRenChang/PSMNet.
Code Repositories
Benchmarks
| Benchmark | Methodology | Metrics |
|---|---|---|
| omnnidirectional-stereo-depth-estimation-on | PSMNet | Depth-LRCE: 1.809 Depth-MAE: 2.509 Depth-MARE: 0.176 Depth-RMSE: 5.673 Disp-MAE: 0.286 Disp-MARE: 0.248 Disp-RMSE: 0.496 |
| stereo-lidar-fusion-on-kitti-depth-completion | PSMNet | RMSE: 884 |
Build AI with AI
From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.