HyperAIHyperAI

Command Palette

Search for a command to run...

5 months ago

360SD-Net: 360° Stereo Depth Estimation with Learnable Cost Volume

Ning-Hsu Wang; Bolivar Solarte; Yi-Hsuan Tsai; Wei-Chen Chiu; Min Sun

360SD-Net: 360° Stereo Depth Estimation with Learnable Cost Volume

Abstract

Recently, end-to-end trainable deep neural networks have significantly improved stereo depth estimation for perspective images. However, 360° images captured under equirectangular projection cannot benefit from directly adopting existing methods due to distortion introduced (i.e., lines in 3D are not projected onto lines in 2D). To tackle this issue, we present a novel architecture specifically designed for spherical disparity using the setting of top-bottom 360° camera pairs. Moreover, we propose to mitigate the distortion issue by (1) an additional input branch capturing the position and relation of each pixel in the spherical coordinate, and (2) a cost volume built upon a learnable shifting filter. Due to the lack of 360° stereo data, we collect two 360° stereo datasets from Matterport3D and Stanford3D for training and evaluation. Extensive experiments and ablation study are provided to validate our method against existing algorithms. Finally, we show promising results on real-world environments capturing images with two consumer-level cameras.

Code Repositories

albert100121/360SD-Net
pytorch
Mentioned in GitHub

Benchmarks

BenchmarkMethodologyMetrics
omnnidirectional-stereo-depth-estimation-on360SD-Net
Depth-LRCE: 0.904
Depth-MAE: 2.112
Depth-MARE: 0.152
Depth-RMSE: 5.077
Disp-MAE: 0.224
Disp-MARE: 0.191
Disp-RMSE: 0.419

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp