Unsupervised Learning for Stereo Depth Estimation using Efficient Correspondence Matching

Hui Wenbin, Kamata Sei-Ichiro

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

To avoid using costly ground-truth, several recent methods formulate the unsupervised depth estimation problem as the task of disparity estimation. However, most of them for monocular depth estimation cannot perform explicit depth maps, while even some methods for stereo depth estimation cannot efficiently use stereo-view information. In this work, we provide a novel cost volume construction approach (matching cost layer) to extract easy-to-learn cost volume by matching feature correspondences between stereo images, which can satisfy the unsupervised learning strategy better. We also propose a new network architecture that learns more reliable spatial location information of objects from fully fused stereo features. Extensive experiments of our method on the KITTI datasets demonstrate its superiority over the state-of-the-art methods.

Original languageEnglish
Title of host publicationICAIP 2021 - 2021 5th International Conference on Advances in Image Processing
PublisherAssociation for Computing Machinery
Pages30-34
Number of pages5
ISBN (Electronic)9781450385183
DOIs
Publication statusPublished - 2021 Nov 12
Event5th International Conference on Advances in Image Processing, ICAIP 2021 - Virtual, Online, China
Duration: 2021 Nov 122021 Nov 14

Publication series

NameACM International Conference Proceeding Series

Conference

Conference5th International Conference on Advances in Image Processing, ICAIP 2021
Country/TerritoryChina
CityVirtual, Online
Period21/11/1221/11/14

Keywords

  • Deep Learning
  • Stereo depth estimation
  • Stereo matching

ASJC Scopus subject areas

  • Human-Computer Interaction
  • Computer Networks and Communications
  • Computer Vision and Pattern Recognition
  • Software

Fingerprint

Dive into the research topics of 'Unsupervised Learning for Stereo Depth Estimation using Efficient Correspondence Matching'. Together they form a unique fingerprint.

Cite this