How Do Neural Networks See Depth in Single Images?

Research output: Chapter in Book/Conference proceedings/Edited volumeConference contributionScientificpeer-review

32 Downloads (Pure)

Abstract

Deep neural networks have lead to a breakthrough in depth estimation from single images. Recent work shows that the quality of these estimations is rapidly increasing. It is clear that neural networks can see depth in single images. However, to the best of our knowledge, no work currently exists that analyzes what these networks have learned. In this work we take four previously published networks and investigate what depth cues they exploit. We find that all networks ignore the apparent size of known obstacles in favor of their vertical position in the image. The use of the vertical position requires the camera pose to be known; however, we find that these networks only partially recognize changes in camera pitch and roll angles. Small changes in camera pitch are shown to disturb the estimated distance towards obstacles. The use of the vertical image position allows the networks to estimate depth towards arbitrary obstacles - even those not appearing in the training set - but may depend on features that are not universally present.
Original languageEnglish
Title of host publicationThe IEEE International Conference on Computer Vision (ICCV)
Pages2183-2191
Publication statusPublished - Oct 2019
EventThe IEEE International Conference on Computer Vision 2019 - Seoul, Korea, Seoul, Korea, Republic of
Duration: 27 Oct 20192 Nov 2019
http://iccv2019.thecvf.com/

Conference

ConferenceThe IEEE International Conference on Computer Vision 2019
Abbreviated titleICCV
CountryKorea, Republic of
CitySeoul
Period27/10/192/11/19
Internet address

Keywords

  • neural networks
  • monocular depth estimation
  • Depth perception

Fingerprint Dive into the research topics of 'How Do Neural Networks See Depth in Single Images?'. Together they form a unique fingerprint.

Cite this