[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
Next Article in Journal
Virtual and Real-Time Synchronous Interaction for Playing Table Tennis with Holograms in Mixed Reality
Previous Article in Journal
Crowd Counting with Semantic Scene Segmentation in Helicopter Footage
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Superb Monocular Depth Estimation Based on Transfer Learning and Surface Normal Guidance

1
Department of Mechanical Engineering and Automation, School of Mechanical and Aerospace Engineering, Jilin University, Changchun 130022, China
2
Research Center for Space Optical Engineering, Harbin Institute of Technology, P.O. Box 307, Harbin 150001, China
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(17), 4856; https://doi.org/10.3390/s20174856
Submission received: 3 August 2020 / Revised: 24 August 2020 / Accepted: 25 August 2020 / Published: 27 August 2020
(This article belongs to the Section Intelligent Sensors)
Graphical abstract
">
Figure 1
<p>The comparison of depth maps were produced by different methods. (<b>a</b>) Raw red-green-blue (RGB) images (<b>b</b>) Ground truth (GT) depth maps [<a href="#B14-sensors-20-04856" class="html-bibr">14</a>], (<b>c</b>) Depth maps from the state-of-the-art (SOTA) practice [<a href="#B7-sensors-20-04856" class="html-bibr">7</a>], (<b>d</b>) Depth maps from our depth prediction network.</p> ">
Figure 2
<p>The comparison of surface normal maps: (<b>a</b>) from left to right: RGB images, Ground-Truth (GT), surface normal maps produced by Qi et al. [<a href="#B10-sensors-20-04856" class="html-bibr">10</a>], ours. (<b>b</b>) Color-map definition: red represents left, green represents up, and blue represents outward.</p> ">
Figure 3
<p>General estimation framework. (<b>a</b>) Coarse depth estimation network; (<b>b</b>) red-green-blue-depth (RGB-D) surface normal network; (<b>c</b>) refinement network.</p> ">
Figure 4
<p>Encoder–decoder coarse depth network.</p> ">
Figure 5
<p>Generating coarse surface normal image. (<b>a</b>) Coarse depth (<math display="inline"><semantics> <mrow> <msup> <mi>D</mi> <mo>*</mo> </msup> </mrow> </semantics></math>); (<b>b</b>)<math display="inline"><semantics> <mrow> <mo> </mo> <msup> <mi>N</mi> <mo>*</mo> </msup> </mrow> </semantics></math>.</p> ">
Figure 6
<p>Surface normal adjustment network (Dense-net-121 based). (<b>a</b>) The general structure of the RGB-D surface normal network (RSN) network; (<b>b</b>) the architectures of up-projection units, fusion module and convolution blocks.</p> ">
Figure 7
<p>Comparison of point clouds from estimated depth maps between Alhashim [<a href="#B6-sensors-20-04856" class="html-bibr">6</a>] and ours. (<b>a</b>) Alhashim [<a href="#B6-sensors-20-04856" class="html-bibr">6</a>], (<b>b</b>) GT, (<b>c</b>) Ours. GT stands for 3D point cloud maps from ground truth images.</p> ">
Figure 8
<p>Qualitative results for depth estimation. (<b>a</b>) RGB images, (<b>b</b>) ground truth (GT), (<b>c</b>) Alhashim [<a href="#B2-sensors-20-04856" class="html-bibr">2</a>], (<b>d</b>) Our Coarse; (<b>e</b>) Refined Depth.</p> ">
Figure 9
<p>Qualitative results for surface normal estimation. (<b>a</b>) RGB images, (<b>b</b>) ground truth normal maps, (<b>c</b>), Geo-Net [<a href="#B10-sensors-20-04856" class="html-bibr">10</a>], (<b>d</b>) reconstructed from in-painted ground truth depth (<b>e</b>) ours, and (<b>f</b>) reconstructed from refined depth. All images are equally scaled for better visualization.</p> ">
Figure 10
<p>Comparison of time consumption: Runtimes of different up-sampling methods including. 2× bilinear interpolation [<a href="#B6-sensors-20-04856" class="html-bibr">6</a>], up and down projection [<a href="#B48-sensors-20-04856" class="html-bibr">48</a>], and up-projection [<a href="#B17-sensors-20-04856" class="html-bibr">17</a>].</p> ">
Figure 11
<p>Refined depth images generating from custom images (densenet-161 model).</p> ">
Figure A1
<p>Results of different methods on NYU V2 Depth. (<b>a</b>) Red-green-blue (RGB) images, (<b>b</b>) ground truth (GT), (<b>c</b>) Ranftl [<a href="#B29-sensors-20-04856" class="html-bibr">29</a>], (<b>d</b>) Hu [<a href="#B30-sensors-20-04856" class="html-bibr">30</a>], (<b>e</b>) Alhashim [<a href="#B3-sensors-20-04856" class="html-bibr">3</a>], (<b>f</b>) ours.</p> ">
Figure A2
<p>Three-dimensional point cloud maps comparison on NYU Depth V2.</p> ">
Figure A3
<p>Point cloud maps were reconstructed based on custom images.</p> ">
Versions Notes

Abstract

:
Accurately sensing the surrounding 3D scene is indispensable for drones or robots to execute path planning and navigation. In this paper, a novel monocular depth estimation method was proposed that primarily utilizes a lighter-weight Convolutional Neural Network (CNN) structure for coarse depth prediction and then refines the coarse depth images by combining surface normal guidance. Specifically, the coarse depth prediction network is designed as pre-trained encoder–decoder architecture for describing the 3D structure. When it comes to surface normal estimation, the deep learning network was designed as a two-stream encoder–decoder structure, which hierarchically merges red-green-blue-depth (RGB-D) images for capturing more accurate geometric boundaries. Relying on fewer network parameters and simpler learning structure, better detailed depth maps are produced than the existing states. Moreover, 3D point cloud maps reconstructed from depth prediction images confirm that our framework can be conveniently adopted as components of a monocular simultaneous localization and mapping (SLAM) paradigm.

Graphical Abstract">

Graphical Abstract

1. Introduction

1.1. Background

Compared with depth estimation methods relying on laser rangefinders or other optical instruments, the computer vision method does not require expensive optical equipment and repeated lens calibration. Therefore, image-based depth prediction has been extensively studied and widely applied to 3D scene understanding tasks, such as structure from motion (SFM) [1,2], simultaneous localization and mapping (SLAM) [3,4], 3D object detection [5], etc.
The computer vision method, i.e., image-based depth estimation, defines image depth as the distance from the object point corresponding to each pixel to the camera and exploits clues of images like linear perspective, focus, occlusion, texture, shadow, gradient, etc. for calculation. All the image-based methods can be summarized as two classes: stereo vision methods and monocular methods. The stereo vision methods are heavily dependent on natural light in a natural environment to collect images and is sensitive to changes in illumination angle and changes in illumination intensity. The differences in image matching of the two pictures will result in considerable differences from the matching algorithm. Compared with the stereo vision methods, the monocular vision systems [6,7,8,9] rarely encounter the aforementioned problems. Therefore, the monocular depth estimation method is more convenient for calibration or identification than the stereo vision approach.
Despite great achievements [6,10,11] in terms of monocular depth estimation, previous research has not given it enough attention [6,7,10,11,12]. First, not enough importance is attached to intrinsically consistency such as semantics labels, color clues of the red-green-blue (RGB) images, the surface normal of the depth images, and the linear perspective of the depth images. The state-of-the-art (SOTA) algorithms [7,9,11,13] leverage fairly complex deep neural networks that are too slow for real-time inference.

1.2. Ideas

In terms of the first content above, there are several methods for acquiring image depth based on supervised learning that can achieve ideal depth maps. However, establishing labeled datasets such as NYU depth V2 [14], Scan-Net [15], and Make-3D [16] is such a challenge for supervised learning methods, such that the datasets [14,15,16] have been constructed for over five years. The ground truth (GT) depth images from these datasets were thus captured by the outdated depth sensors. Therefore, the geometric accuracy of GT images definitely fluctuated, which would cause side effects on the precision of both training and testing processes. Therefore, we first predicted depth maps from RGB images based on an encoder–decoder structure and denoted those images as “coarse depth maps.” Then, surface normal maps were estimated and used for applying geometric guidance to the depth estimation process, i.e., the surface normal estimation was introduced in this paper so as to enable accurate regression of geometric structure and complex boundaries of 3D objects.
As for the second content, previous authors [6,17,18] chose to propose a lighter-weight model-based encoder–decoder network, which produced more accurate depth estimation maps and reduced computational complexity. Wofk et al. [17] demonstrated that a well-designed low latency network can maintain real-time depth estimation with economical computation costs. Alhashim et al. [6] proved that a very simple transfer learning-based decoder robustly achieves high-resolution depth maps. Previous studies [6,9,19] proved that the Dense-net is more suitable for depth estimation than models like SE-Net, Res-Net, and Mobile-net. However, experiments in references [6,9] showed that pre-trained deep learning structures based on Densenet-161, Densenet-169, and Densenet-201 models cannot afford real-time depth estimation. As a result, a lighter-weight depth estimation network Densenet-121 was adopted in this paper, which requires fewer parameters and inference iterations than other dense-net models. Moreover, in order to enhance the performance of the Densenet-121 model, the surface normal maps were jointly estimated with depth images.

1.3. Approach

The reform of current monocular depth framework can be summarized as follows. First, a novel depth estimation network called the coarse depth estimation (CDE) network was proposed, which leverages an efficient encoder–decoder network architecture. In terms of deep-feature-encoding layers, we introduced a pre-trained Densenet-121 model that was primitively proposed for image classification and object detection. As for the decoder, we chose the 2× bilinear proposed by Alhashim et al. [6] as the up-sampling method. Second, we put forward a network called the red-green-blue-depth (RGB-D) surface normal (RSN) network in order to obtain more accurate surface normal maps. The RSN network was designed as a two-stream encoder–decoder architecture, which produced surface normal maps by hierarchically fusing coarse depth produced by the CDE network and RGB images from a selected dataset. Then, the refinement network exploited surface normal maps produced by the RSN network for ameliorating the quality of coarse depth images.
Three-dimensional features can be represented with different formats, including point clouds, meshes, and volumetric grids. Specifically, the point cloud map is a set of data points preserving the original geometric information of 3D objects without any discretization. Three-dimensional point cloud maps were used as a criterion for evaluating the depth maps produced in our study.

1.4. Contributions

Different from previous studies [10,20], we remodeled the conventional depth estimation network layout. Repeated experiments proved that a deep network based on the low latency Densenet-121 model affords real-time coarse depth estimation and depth refinement depending on a Tesla M40 produced by NVIDA in Santa Clara, CA, U.S. with single 12G memory capacity. Benefiting from guidance of surface normal maps, the Densenet-121 based network obtained better depth maps than SOTA practices [6,7,9]. As shown in Figure 1, our network outperforms SOTA depth estimation on an NYU [14] dataset and produces higher-resolution results that capture object boundaries more faithfully.
Our second contribution is reflected in proposing an RGB-D surface normal network, which effectively captures the geometric relationships between RGB and depth images. Different from previous frameworks [21,22], we proposed a fusion network leveraging both RGB and coarse depth prediction images instead of using RGB images only. Images from different domains can complement each other in the surface normal estimation process, i.e., coarse depth images from the CDE network contribute to enhance geometry details, and RGB images make up the missing depth pixels. As shown in Figure 2, we achieved better surface normal maps than Qi et al. [10] based on the RGB-D surface normal network. Moreover, as we can see in Figure 1, with the geometrical guidance of surface normal, depth maps with distinct boundaries were acquired.

2. Related Work

Stereo depth estimation can be regarded as a well-posed process once ignoring the problem of occlusions and depth discontinuities. Moreover, depth estimation methods based on stereo vision achieved even more accurate and robust depth maps than RGB-D sensors [14,15]. Meanwhile, more precise depth features from our monocular method can also make contributions to better multi-view stereo reconstruction.
Monocular depth estimation has been considered by plenty of researchers [23,24,25,26,27,28,29], which usually defined the estimation as a regression of the depth map from a single RGB image. Eigen et al. [21] introduced the application of CNNs in monocular depth estimation, which inspired researchers to explore methods based on deep learning. At present, deep learning methods play a leading role in monocular depth estimation practices. Generally, deep learning methods can be divided into supervised approaches, self-supervised approaches, and unsupervised approaches. Supervised monocular depth methods achieved great breakthroughs relying on well-annotated ground-truth images offered by datasets [14,15,16]. For example, Liu et al. [28] combined CNNs with CRF to learn the super-pixel-wise connections between depth and RGB images. Different from the supervised method, self-supervised methods usually set up a separate camera pose estimation network [26,27] or a jointly calculated optical flow and camera pose [30]. Unsupervised practices learn the scene depth image synthesis [10] or ego-motion in monocular videos without using ground truth data [24,25]. Similar to the SOTA methods [2,7], in recent years, the supervised learning method was selected in this paper.
Multi-task/cross-task learning was designed based on intrinsic connections of physical elements of the research targets selected. Some recent works attempt to investigate the sharing of image features between different tasks [31,32,33]. Jiao et al. [34] jointly trained semantic labeling and depth estimation in their encoder–decoder network architecture. Zhang et al. [11] proposed a pattern-affinity propagation method for jointly predicting depth, surface normal and semantic segmentation. As for us, the proposed network first jointly predicts depth and surface normal and then takes advantage of surface normal maps to refine the predicted coarse depth maps.
Surface normal guidance has been introduced by previous studies [10,19,20], where they employed surface normal maps as 3D cues for improving the geometric quality of monocular depth images. Qi et al. [10] jointly calculated depth and surface normal from a single image, making the final estimation geometrically more precise. In work of Zeng et al. [19], a skip-connected architecture was proposed to fuse features from different layers for surface normal estimation. A novel 3D geometric feature virtual normal was proposed by Yin et al. [20] to refine the predicted depth maps. Surface normal estimation was adopted in this paper for calculating the angular difference between predicted depth images and ground-truth maps, which thus applied geometric restriction to depth images.
A transfer-learning-based deep framework was adopted by Zhang et al. [11] to obtain the state-of-the-art semantic segmentation maps. We adopt a pre-trained Densenet-121 model [5] as the backbone of the coarse depth estimation network in depth feature extraction. Our method benefits from the application of transfer learning, where we take advantage of encoders originally designed for 3D object detection by Alhashim [6].
Encoder–decoder networks were widely adopted in various computer vision tasks such as image classification, image segmentation, and 3D objects detection. In recent years, such architectures made significant contributions to both supervised learning and unsupervised learning-based practices of monocular depth estimation [3,30]. We devised a compendious but effective symmetrical encoder–decoder structure with skip connections. Repeated experiments indicated that our encoder–decoder network with simple structure can outperform the SOTA depth synthesis based on more complicated deep learning architecture [7,9].

3. Our Method

This section presents our method for monocular depth estimation. First, the general deep learning framework is introduced. Then, rational loss functions for overall training process are defined. Finally, we discuss the practice of 3D point cloud maps reconstruction.

3.1. Framework Overview

As shown in Figure 3, the whole deep learning structure consists of three parts: a coarse depth estimation network, an RGB-D surface normal network, and a refinement network.
The coarse depth estimation (CDE) network leverages efficient encoder–decoder network architecture. Coarse depth images generated from the CDE network are then fed to the RGB-D surface normal network for RGB-D fusion. Moreover, coarse depth images are also converted to coarse surface normal maps based on a fix-weight network [4]. For convenience, we describe a single RGB input image as   I C , a single in-painted ground truth depth image produced by Levin [35] as   D G T , an output coarse depth map as D * , and a coarse surface normal map recovered from D *   as   N * .
The RGB-D surface normal (RSN) network was designed for obtaining accurate surface normal maps, which functions as the refining coarse depth map ( D * ). As shown in Figure 3, the RSN network can be divided into two streams—the RGB stream and the depth stream. The latter can be further divided into depth branch and the confidence map branch. For the general architecture of RSN network, we define a single RGB input as   I F , a corresponding coarse depth input as   D * , and a surface normal output map as   N F .
The RGB stream and depth branch in Figure 3 operate respectively to generate RGB feature   R 1 ~   R 4 and relative sensor depth feature   D 1 ~   D 4 with hierarchical resolution. Then, the two branches cooperate to combine and fuse features from each branch.
The ground truth depth images ( D G T ) used for training the CDE network were produced following Levin et al. [35]. This method greatly contributes to fill pixel holes in sensor depth images from NYU-Depth-V2 [14]. However, this method cannot eliminate the side effects caused by lost pixels on accuracy of ground truth depth images and corresponding coarse depth images ( D * ) . Therefore, a confidence map network branch was set following the method proposed by Zeng et al. [11], which generates confidence maps to indicate whether side effects resulted from pixel holes on   D *   or not. Confidence maps [19] of depth image were produced by combining mask images [21] ( I M ) with relative coarse depth images ( D * ) and were denoted as   C 1 ~   C 4 according to resolution.
The refinement network actually servers as the convolution kernel function, which optimizes the coarse depth maps from the CDE network guided by surface normal maps from the RSN network. Finally, with the aid of the refinement network, a superbly accurate depth map   D was generated.

3.2. Coarse Depth Estimation (CDE) Network

Figure 4 shows the distinct structure of the encoder–decoder network for getting coarse depth maps ( D * ). For the encoder, the raw RGB image ( I C ) is encoded into a feature vector based on a Densenet-121 [5] model that has been pre-trained on Image-Net [36]. The feature vector is then transmitted to a sequence of up-sampling layers to produce D * with half resolution of   I C . As for decoding operation, the decoder network consists of four up-sampling units ( BU 1   and   USB 1 ~ USB 3 ) and relative concatenation (⨁) skip-connections. In decoding layers, the 2× bi-linear interpolation proposed by Alhashim [6] is adopted as an up-sampling method.
As shown in Figure 5, the coarse image D * is then converted to coarse surface normal image ( N * ) based on least square algorithm [37], and the inference network is just a fix-weight network [11].

3.3. RGB-D Surface Normal (RSN) Network

Both the RGB stream and depth branch leverage encoder–decoder network architecture. The two branches employ a pre-trained Densenet-121 model as encoding backbone [26], whose detailed structure is illustrated in the upper row of Figure 6. Generally, the encoder consists of same convolution layers with the raw Densenet-121 [5] encoder, except the last convolution blocks. The number of channels was reduced from 1024 to 512 via bottleneck layers aiming to reduce redundant parameters. A symmetric decoder equipped with concatenation connections for the refitted encoder was designed. Multi-scale up-sampling layers were introduced to decoder, which enables RGB-D images fuse in different scales. What’s more, common pooling masks were set to let the network learn more image features.
Different from CDE Network, the RSN Network leverages coarse depth images ( D * ) instead of in-painted [35] ground truth depth images ( D * ) as the depth input. The pre-trained RSN Network can thus be applied for estimating surface normal maps based on custom images.
As is shown in Figure 3, pixel holes in mask images ( I M ) suggest that there are lots of missing pixels in ground truth depth images, which inevitably causes deviation to supervised learning. Therefore, we adopted a multi-layer convolution network ( CB 2 ) for producing confidence map C l [19] of input depth images.   l   stands for scale value of images, i.e., if the resolution of 2D images can be denoted as   H × W , then the corresponding   l is defined as { ( l , H × W ) } = { ( 1 , 40 × 30 ) ; ( 2 , 80 × 60 ) ; ( 3 , 160 × 120 ) ; ( 4 , 320 × 240 ) } . The detailed structure of C B 2 is shown in Figure 6b. The depth branch also adopts Densenet-121 model–based encoding layers, and the fusion calculation takes place at the decoder side. As shown in Figure 6a, the depth features   D l *   are passed into the fusion module in each scale   l and re-weighted with the confidence map   C 1 ~   C 4 . Then re-weighted   D l *   are concatenated ( ) with color features with same resolution and transmitted to a de-convolution layer to produce surface maps based on RGB-D fusion. Finally, the convolution block ( C B 4 ) of RSN Network generates the surface map   N * . The RGB-D fusion algorithm [19] can be expressed as Equation (1).
N l =   d e c o n v [ I F l ( C l     D l * ) ]
The decoder layers of the RSN network were also designed based on the Densenet-121 model. Different from the CDE network, up-projection units [17] ( UPP 1 ~ UPP 4 and   UPP 1 ~ UPP 4 ) were employed instead of bilinear interpolation for boosting the surface normal estimation process. The detailed structure of up-projection units is shown in the upper row of Figure 6b.

3.4. Refinement Network

For a random pixel   i from the coarse depth images, we denote ( h i , w i ) as the location of pixel i in 2D space and ( x i ,   y i ,   d i * ) as the coordinate of corresponding 3D point, where   d i *   represents the coarse depth value. Similarly, surface normal can be noted as ( n x i ,   n y i ,   n d i * ). Then, a tangent plane p i [11] can be defined according to Equation (2).
n x i ( x x i ) + n y i ( y y i ) + n   d i * ( d d * ) = 0
A small 3D neighborhood ( H i ) of i was defined in previous studies [10,38]. For a random pixel j   H i , its depth value d j . The depth prediction value of pixel   i   can be computed as Equation (2) according to the pinhole camera model,
d j i = ( u j c x ) n x i / f x + ( v j c y ) n y i / f y + n   d i * ( u i c x ) n x i / f x + ( v i c y ) n y i / f y + n   d i * × d j
where f x     and f y   represent the focal length in x and y directions, respectively; C x     and   C y are coordinates of the lens principal points; and   K represents the convolution kernel operation. Then, in order to refine depth value of pixel   i , we applied kernel regression [10] to operate estimation on all pixels in H i   as
  d i = j   H i [ d j * K ( n i , n j ) d j i ] j   H i [ d i j K ( n i , n j ) ]
K ( n i , n j ) = n j T n i
where   d i is the refined depth value n i = [ n x i ,   n y i ,   n d i * ]. According to the linear kernel algorithm, the angle error between surface normal image ni and image nj decides whether pixels   i   and   j   are in the same tangent plane   p i or not. Therefore, the smaller angular error contributes to more accurate depth estimation   d j i .

3.5. Loss Function

This section presents loss functions for regression problems existing in depth and surface normal estimation. The loss functions evaluate the difference between the ground truth images and the prediction images generated by the deep learning network. For random pixel   i , we define the coarse depth map, refined depth map and ground-truth depth map as   d i * ,   d i , and   d i g t separately. Similarly, we denoted the coarse surface normal map, RGB-D fusion surface normal, and ground-truth surface normal as   n i * , n i F , and   n i g t separately. The total number of pixels is M, and the loss function for depth value is defined as Equation (6).
l d e p t h = 1 M ( i d i d i g t 2 2 + ω 1 i d i * d i g t 2 2 )
As shown in Equation (6), the loss function l d e p t h   is the point-wise loss defined according to depth values. On the one hand, l d e p t h   computes the sum of the L2 norm of the error vectors between coarse depth maps d i *   and ground truth images   d i g t . On the other hand, l d e p t h     calculates difference between refined depth maps d i   and ground truth images   d i g t   over all pixels. The loss for surface normal training process is defined as Equation (7),
l n o r m a l = 1 M ( μ l i n i F l n i g t 2 2 + ω 2 i n i * ( t ) n i g t 2 2 )
where l in Equation (7) stands for the scale value [19] of surface normal features and   ω 1 , ω 2 , and μ l   stand for weight parameter of loss functions. The loss function l n o r m a l     computes the L2 norm of the angular errors of the orientation vectors for each pixel. On the one hand, l n o r m a l   computes the L2 norm of the angular difference between coarse surface normal maps n i *   and ground-truth normal maps   n i g t . On the other hand, l n o r m a l     calculates the divergence between fusion normal maps n i F l   and ground-truth normal maps   n i g t . There are a lot of step edges such as texture crosshatches or object boundaries [39] in natural RGB images, which inevitably interfere with the accuracy of depth estimation. It is necessary to prevent such interference so as to deal with distorted or blurry problems around step edges. The l g r a d function defined by Hu et.al. [9] was adopted in this paper in order to penalize depth errors between neighboring pixels of depth images,
l g r a d = 1 M   i = 1 n ( F ( x ( τ i ) ) + F ( y ( τ i ) ) )
where   τ i = d i * d i g t 1 ,     x (   τ i ) and   x (   τ i ) represent spatial derivation of   τ i along the x and y directions, respectively.   F ( x ) stands for a log algorithm function of depth errors defined by Hu [30].
To be concluded, we defined a total loss denoted as   l t o t a l
l t o t a l = λ 1   l d e p t h   +   λ 2   l g r a d   +   λ 3   l n o r m a l
where l d e p t h represents pixel-wise loss for depth values, l g r a d   represents loss function for step edges, and   l n o r m a l represents function for surface normal.   λ 1 , λ 2 ,   and   λ 3   stand for initialed weights working for balancing effects of back propagation.

3.6. Recovering 3D Features from Estimated Depth

As is often the case [10,21], 3D space geometric constraints contribute to the capability of deep network in terms of depth estimation and corresponding 3D point cloud reconstruction. Therefore, the 3D point cloud maps shown in Figure 7 were reconstructed following Li et al. [40] to visualize the quality of depth maps predicted by our scheme.
We compare the 3D point-cloud maps produced by our scheme based on Densenet-121 with that of Dense-depth [6], which is based on Densenet-169 and does not introduce geometric constrictions. As shown in Figure 7 and Figure 8, our scheme outperforms Dense-depth [6] in terms of depth estimation results and the average quality of the relative 3D point-cloud maps. Therefore, the guidance of surface normal will definitely improve the quality of depth maps in terms of features in 3D space. What is more, better 3D point cloud maps can be recovered from depth maps with constriction of geometric features. As a result, the quality of point-cloud maps should also be adopted as a fundamental metric for evaluating the accuracy of depth estimation.

4. Experiments

In this section, we describe the implement details of experiments, evaluate the performance of our depth estimation scheme on NYU-Depth-V2 [14], and compare the prediction results against existing state-of-the-art (SOTA) methods. Moreover, we present results of ablation experiment to analyze the influence of the different parts of our proposed method.

4.1. Dataset

The NYU-Depth-V2 dataset [14] contains 407 K frames taken from 464 different indoor scenes, which were split into 249 training scenes and 215 testing scenes. Specially, 1449 RGB images were accurately labeled with depth images, in which 654 images are annotated for testing phase and others for training phase. All images were collected from videos captured by Kinect RGB-D sensor produced by Microsoft in Redmond, WA, the U.S. NYU-Depth-V2 widely serves as training datasets for supervised monocular depth prediction due to its accurate ground-truth (GT) depth labels and abundant image samples.
In some previous studies [6,9,10,20], depth estimation networks were trained on subsets sampled from the NYU dataset [14]. In practice [10,20], 30 K training frames were sampled from the raw NYU dataset [14]. The fewer frames used, the better-designed the deep learning schemes [4,6], which outperformed SOTA practices [7,21,41] that were trained on the entire NYU dataset. Moreover, the practice by Hu et al. [9] proved that the models are trained on subsets with more frames performing slightly better, but gains in accuracy did not justify the lower learning efficiency and higher system latency. Therefore, instead of the official splits image set, a subset with 30 K frames was used in this paper. All frames were randomly sampled from 249 training subsets.
The NYU dataset [14] does not provide ground-truth surface normal maps ( N G T ) ; previous studies [10,21] computed N G T   from in-pained depth images, and thus, the quality of the produced N G T   were up to the in-painting algorithm proposed by Levin [35]. Instead of the in-painting method, the method proposed by Hickson et al. [42] was leveraged for obtaining   N G T   in   this   paper . As for confidence map network, we utilized accurate binary mask images ( I M ) offered by Eigen et al. [21].

4.2. Implementation Details

We implemented both the coarse depth estimation (CDE) network and the RGB-D surface normal network (RSN) using the deep learning platform PyTorch by Huang et al. [43] operating on a Tesla M40 GPU produced by NVIDA in Santa Clara, CA, the U.S. with 12 GB capacity.
The encoder of the CDE network was designed based on the Densenet-121 model, which initializes weights for decoder layers. Specially, the last classification layers of the Densenet-121 model were removed.
The Adam [44] optimizer was adopted for training the CDE network, and learning rate was set as 0.0001. The raw RGB images with resolution of H × W   were down-sampled to H / 2 × W / 2 for boosting training processes and fitting the size of output depth images. We conducted the coarse depth estimation training phase with a batch size of six for 20 epochs. A small subset with 654 samples [14] was employed for testing performance of CDE network, and the batch size for testing phase was set as 32. The network finally produced depth images with resolution of H / 2 × W / 2 and an error evaluation index for supervising the training algorithm.
In order to avoid interference from over-fitting, four augmentation methods were employed following Cubuk et al. [45] for depth estimation training phase:
(1) The horizontal mirroring operation was applied to both RGB and depth images with a probability of 25%.
(2) Much rotation leads into invalid data for GT depth images [6], so input images in training process were rotated by slight degrees, which ranged from −2 to 2 with a probability of 25%.
(3) Contrast and brightness values of RGB images input were randomly scaled by (0.6, 1.2) with a probability of 50%.
(4) Both RGB and depth images were randomly resized to 320 × 256 with a probability of 50%.
The training process of surface normal was also performed on NYU Depth-V2 [14] and the number of training epochs was also set as 20 with batch size of six. The testing image-set for RSN Network consists of 654 sample RGB images and corresponding ground truth surface normal images produced following the practice by Hickson et al. [42]. For optimizer, the Adam [44] was selected with original learning rate of 1 × 10−4, initial parameters β   1   = 0.9, β   2 = 0.999 and weight decay of 1 × 10−4. The RSN network also produced surface normal images with the resolution of H/2 × W/2.
The refinement network functioned as the convolution kernel function [10], which does not demand any training process.
In all training experiments, weight   ω 1 for   l d e p t h   was set as 0.5 for balancing the importance of coarse depth estimation and depth refinement. Weight   ω 2   for   l g r a d was set as   ω 2 = 0.5   to balance the influence of two terms. As is often the case [6,7,9,10,20], the hyper-parameters are empirically set as a reasonable value for loss functions. In our study, parameter   ω 2   was set as 0.5 according to validations on an officially annotated subset with 1449 images. Weight   μ l   for   l n o r m a l was set as   μ l = 0.2 l ,   where l = {1,2,3,4}. The parameter   l was defined in Section 3.4, which stands for the scale value of images fed to the RSN network.
As shown in Equations (6)–(8),   l d e p t h calculates the depth value errors for each pixel, l n o r m a l calculates angular cosine errors for each pixel, and l g r a d calculates the gradient errors in the log domain. Therefore, the value of l d e p t h will be obviously bigger than l n o r m a l and   l g r a d . To mitigate this effect, the parameter λ 1   is set to a small value. According to test results on the annotated small subset, parameters { λ 1 ,   λ 2 ,   λ 3 }   were set as reasonable weights {   0.2 ,   1 ,   1   } .
As the proceeding of training process, the model gradually converges. The total number of training parameters for entire network based on Densenet-121 model was approximately 32.4 M. Training was performed for 600 K iterations on NYU Depth-V2 [14], which took 18 h to finish. With repeated inference learning, the value of loss functions   l g r a d , l n o r m a l , and   l d e p t h would converge to zero.

4.3. Evaluation Criteria

For quantitatively evaluation, three error metrics were used by the previous work of Eigen et al. [46]: absolute relative error ( A b s R e l ), root mean squared error (RMSE), and average ( L o g 10 ) error   ( E L o g 10 ) . Moreover, threshold accuracy (   T r e ) was selected as the accuracy metric. All metrics can be defined as:
m a x ( D i G T D i , D i D i G T ) = δ < T r e
A b s R e l = 1 S   i = 1 S | D i G T D i | 2 D i *
R M S E = 1 S i = 1 S | D i G T D i | 2
E L o g 10 = 1 S i = 1 S | L o g 10 D i G T L o g 10 D i |
As is shown in the above metrics, we denote D i G T as the ground truth depth corresponding to pixel   i , D i as the relative estimated depth, and S as the total number of pixels with available value in ground truth maps. Here, three different thresholds ( δ ,   δ 2 ,   δ 3 ) are set as ( 1.25 ,   1.25 2 ,   1.25 3 ) according to conventional works [2,46].
Three error metrics [21,27] were used for evaluating surface normal maps in this paper: mean of angle error (Mean), medians of angle error (Median), and root mean square error (RMSE). Moreover, three different thresholds (11.25°, 22.5°, 30°) [21] were used for calculating the specific angular error of pixels.

4.4. Benchmark Performance Comparison

We select the SOTA practice Dense-depth [6] for comparison, which adopted Densenet-169 [5] as the encoder of deep network. The pre-trained model functioned by extracting depth feature from RGB image input. As shown in Figure 8, Dense-depth outperforms the coarse depth estimation network based on Densenet-121 [5] proposed in this paper, while performs worse than our entire depth estimation scheme containing the surface normal guidance. Some depth estimation samples are shown in Figure 9, and more depth maps from our framework are listed in the Appendix A. All images are colorized for better visualization.
Table 1 lists depth estimation results generated by proposed method and previous masterpieces [6,7,9,21]. The quality of depth maps can be evaluated according to Equations (10)–(13).
According to results shown in Table 1 and Figure 8, the refined depth images have a better geometric quality than that of Dense-depth, which proves the necessity of surface normal guidance in monocular depth prediction tasks.
Table 2 compares the surface normal calculation results based on RGB image methods [33,37], surface normal with depth consistency method [10], and the RGB-D fusion method (ours). From Table 2, it can be seen that the surface normal scheme leveraging both depth and RGB image features is averagely superior to the schemes that employ depth images or RGB image only.
Instead of directly comparing errors between pixel values, Table 2 compares angular difference between orientation vectors for each pixel. The qualitative performance of previous research [10,22,37] is cited from the original papers. As listed in Table 2, with the aid of high-order geometric RGB-D fusion, the surface normal maps from our method outperforms Geo-Net [10]. From Table 2 and Figure 8, we can conclude that the method proposed in this paper can recover better shape from RGB images, which contributes to supply more accurately geometric details with depth images.

4.5. Computational Performance

Table 3 compares the computational efficiency of proposed depth estimation algorithm with that of state-of-the-art (SOTA) methods [6,7,9]. It is seen in Table 3 that our model achieves SOTA results on the RMSE metric. What is more, our model requires fewer training parameters, fewer training iterations, and fewer image samples in terms of the same training epochs. Furthermore, our model consumed less training time while operating on similar platforms.
In additional, we tested our pre-trained model and testing models released by references [6,9] on an Tesla M40 produced by NVIDA in Santa Clara, CA, the U.S. with single 12G memory capacity. As shown in Table 3, our model performed lower latency and higher accuracy.
All the data shown in Table 3 were cited from the original paper and corresponding released models [6,9].

4.6. Ablation Study

In this section, ablation studies are performed to verify each part of the proposed architecture in terms of performance on depth estimation.
In this experiment, the DenseNet-121 model was substituted with pre-trained DenseNet-161 both in the coarse depth estimation (CDE) network and the RGB-D surface normal (RSN) network. As shown in Table 4, the network based on the DenseNet-161 model [5] outperforms that of the DenseNet-121 model [5] in terms of qualitative accuracy metrics. However, according to the training parameters listed in Table 4, the growth of encoding layers in deep learning deep structure will introduce superfluous training costs.
Furthermore, as shown in Table 4, with the geometric constrictions from surface normal maps, the CDE network based on Densenet-121 model outperforms Dense-depth [6], whose decoder was designed based on deeper-model Densnet-169.
In this experiment, Mobilenet-V2 [47] was adopted as the backbone of depth prediction and surface normal estimation, which utilizes less training weight parameters and lower computational complexity than the Densenet-121 model. As shown in Table 5, the network based on Mobilenet-V2 apparently required less training cost and thus performed lower latency. However, Mobilenet-V2 based deep learning scheme produced worse depth maps than Densenet-121 model [5] at the same time. Therefore, Mobilenet-V2 [47] model-based networks can be conveniently embedded into simple mobile platforms such as a mobile phone and lite drone. The detailed structure of coarse depth estimation network based on Mobilenet-V2 [47] model was shown in Appendix B.
In this experiment, up-projection units in decoder layers of the RGB-D surface normal (RSN) network was substituted with the 2× bilinear up-sampling and up- and down-projection unit [48], respectively, for ablative comparison.
The up-projection unit [17] was designed for embed platforms. As shown in Figure 10, it achieves the lowest latency in this experiment. Therefore, it is suitable to be employed as the up-sampling strategy of the RSN network for boosting surface normal inference process.
The up- and down-projection unit applies super-resolution (SR) techniques to up-scaling the input image to a higher resolution. As shown in Figure 10, the value of absolute relative error ( A b s R e l ) suggests that the refined depth maps obtained by the up and down projection unit slightly outperforms 2× bilinear interpolation [6]. When using the up- and down-projection projection unit [48] in decoder layers, we found that the gains in performance did not justify the slow learning time and the extra GPU memory required. Therefore, the 2× bilinear up-sampling proposed by Alhashimet al. [6] functioned as an up-sampling method in our coarse depth estimation network.

4.7. Custom Results

To verify the deep learning network proposed herein, we took some videos of different indoor scenes with the monocular camera of a smart phone. Then, we randomly captured some RGB images from videos and resized them as the resolution of 640 × 480 for depth estimation.
As shown in Figure 11, depth maps predicted from our network perform distinct boundary and robust geometric shapes.

5. Conclusions

In this work, we designed a lighter-weight encoder–decoder deep learning network for depth estimation from monocular RGB images. The encoder layers were designed based on pre-trained deep learning model originally for image classification. Experiments proved that an effective encoder based on transfer learning and geometric guidance outperforms previous methods [7,49] employing complex feature extraction layers. Ablation studies suggested that employing different pre-trained models enabled our networks adapt to different platforms. For example, Mobilenet-V2 [3] can be a suitable model for simple mobile platforms such as smart phones and light micro drones, the Densenet-121 model can be used for mobile platforms equipped with a powerful GPU device, while a denser decoder based on Densenet161 can only be applied to platforms that can afford expensive training costs.
What is more, surface normal estimation was introduced to improve the quality of depth images. Benefiting from geometric guidance offered by surface normal maps, our network obviously achieved better depth images on benchmark NYU depth V2 [14] than state-of-the-art methods [4,7,9]. With the geometric constrictions from surface normal maps, superb 3D point-cloud maps were reconstructed from refined depth images.
Because this work greatly benefits from surface normal estimation, we believe that there are still many other possible geometric features that can be used to improve monocular depth estimation. Therefore, we will further study the effects of geometric features such as 3D objects boundary, semantic label, and image defocus.
Further research will also focus on how to apply pre-trained depth estimation network to 3D vision practices, such as augment reality (AR), simultaneous localization and mapping (SLAM), and structure from motion (SFM).

Author Contributions

Conceptualization, K.H. and X.Q.; Methodology, W.Z.; Software, K.H.; Validation, S.C., K.H. and W.Z.; Formal analysis, X.Q.; Data duration, F.Z., Z.C.; Writing—original draft preparation, K.H., S.C., and W.Z.; Writing—review and editing, K.H., H.Q., S.C., W.Z., and X.Q.; Funding acquisition, W.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Acknowledgments

The authors want to express gratitude to the editor and anonymous reviewers for their professional comments and valuable suggestions.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Figure A1. Results of different methods on NYU V2 Depth. (a) Red-green-blue (RGB) images, (b) ground truth (GT), (c) Ranftl [29], (d) Hu [30], (e) Alhashim [3], (f) ours.
Figure A1. Results of different methods on NYU V2 Depth. (a) Red-green-blue (RGB) images, (b) ground truth (GT), (c) Ranftl [29], (d) Hu [30], (e) Alhashim [3], (f) ours.
Sensors 20 04856 g0a1
Figure A2. Three-dimensional point cloud maps comparison on NYU Depth V2.
Figure A2. Three-dimensional point cloud maps comparison on NYU Depth V2.
Sensors 20 04856 g0a2
Figure A3. Point cloud maps were reconstructed based on custom images.
Figure A3. Point cloud maps were reconstructed based on custom images.
Sensors 20 04856 g0a3

Appendix B

Table A1. Details of coarse depth estimation architecture based on Densenet-161.
Table A1. Details of coarse depth estimation architecture based on Densenet-161.
LayerOutputOperator
Input640 × 480 × 3-
Convolution1 (CONV. 1)320 × 240 × 96CONV. 7 × 7
DenseBlock1 [26] (DB.1)160 × 120 × 96Avg. Pooling 2 × 2
DB2 80 × 60 × 192Avg. Pooling 2 × 2
DB3 40 × 30 × 384Avg. Pooling 2 × 2
DB4 20 × 15 × 1056Avg. Pooling 2 × 2
CONV. 220 × 15 × 2208CONV. 1 × 1
Bottle Layer20 × 15 × 1024Batch-Normal + ReLU + CONV. 1 × 1
UP-Sample1 (US1)40 × 30 × 1024Up-sample 2 × 2
Concatenating 140 × 30 × 1408Concatenating DB3 Avg. Pooling
US1-CONV. 1 [2]40 × 30 × 512CONV. 3 × 3
US1-CONV. 240 × 30 × 512CONV. 3 × 3
UP-Sample2 (US2)80 × 60 × 512Up-sample 2 × 2
Concatenating280 × 60 × 704Concatenating DB2 Avg. Pooling
US2-CONV. 180 × 60 × 256CONV. 3 × 3
US2-CONV. 280 × 60 × 256CONV. 3 × 3
UP-Sample3 (US3)160 × 120 × 256Up-sample 2 × 2
Concatenating3160 × 120 × 352Concatenating DB1 Avg. Pooling
US3-CONV. 1160 × 120 × 128CONV. 3 × 3
UP3-CONV. 2320 × 240 ×128CONV. 3 × 3
UP-Sample4 (US4)320 × 240 × 128Up-sample 2 × 2
Concatenating4320 × 240 × 224Concatenating CONV. 1 Avg. Pooling
US4-CONV. 1320 × 240 × 64CONV. 3 × 3
US4-CONV. 2320 × 240 × 64CONV. 3 × 3
CONV. 3320 × 240 ×1CONV. 3 × 3
Output240 × 320 × 1-
Table A2. Details of coarse depth estimation architecture based on Mobilenet-V2.
Table A2. Details of coarse depth estimation architecture based on Mobilenet-V2.
LayerOutputOperator
INPUT640 × 480 × 3-
Convolution1 (CONV.1)320 × 240 × 32CONV. 3 × 3
Inverted Residual [3] Block1 (IRB1)320 × 240 × 16Bottleneck × 1
Inverted Residual Block2 (IRB2)160 × 120 × 24Bottleneck × 6
Inverted Residual Block3 (IRB3)80 × 60 × 32Bottleneck × 6
Inverted Residual Block4 (IRB4) 80 × 60 × 64Bottleneck × 6
Inverted Residual Block5 (IRB5)40 × 30 × 96Bottleneck × 6
Inverted Residual Block6 (IRB6) 40 × 30 × 160Bottleneck × 6
Inverted Residual Block7 (IRB7)20 × 15 × 320Bottleneck × 6
CONV. 220 × 15 × 1280CONV. 1 × 1
CONV. 320 × 15 × 768CONV. 1 × 1
UP-Sample1 (US1)40 × 30 × 768Up-sample 2 × 2
Concatenating140 × 30 × 928Concatenating IRB6
US1-CONV. 140 × 30×384CONV. 3 × 3
US1-CONV. 240 × 30 × 384CONV. 3 × 3
UP-Sample2 (US2)80 × 60 × 384Up-sample 2 × 2
Concatenating280 × 60 × 448Concatenating IRB4
US2-CONV. 180 × 60 × 192CONV. 3 × 3
US2-CONV. 280 × 60 × 192CONV. 3 × 3
UP-Sample3 (US3)160 × 120 × 192Up-sample 2 × 2
Concatenating3160 × 120 × 216Concatenating IRB2
US3-CONV. 1160 × 120 × 96CONV. 3 × 3
UP3-CONV. 2160 × 120 × 96CONV. 3 × 3
UP-Sample4 (US4)320 × 240 × 96Up-sample 2 × 2
Concatenating4320 × 240 × 112Concatenating IRB1
US4-CONV. 1320 × 240 ×64CONV. 3 × 3
US4-CONV. 2320 × 240 × 64CONV. 3 × 3
CONV. 4320 × 240 × 1CONV. 1 × 1
Output320 × 240 × 1-

References

  1. Guizilini, V.; Ambrus, R.; Pillai, S.; Gaidon, A. Packnet-sfm: 3d packing for self-supervised monocular depth estimation. arXiv 2019, arXiv:1905.02693. [Google Scholar]
  2. Ummenhofer, B.; Zhou, H.; Uhrig, J.; Mayer, N.; Ilg, E.; Dosovitskiy, A.; Brox, T. Demon: Depth and motion network for learning monocular stereo. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 5622–5631. [Google Scholar]
  3. Zhou, H.; Ummenhofer, B.; Brox, T. Deeptam: Deep tracking and mapping. In Proceedings of the European Conference on Computer Visio (ECCV), Munich, Germany, 8–14 September 2018; pp. 822–838. [Google Scholar]
  4. Yu, C.; Liu, Z.; Liu, X.; Xie, F.; Yang, Y.; Wei, Q.; Fei, Q. DS-SLAM: A Semantic Visual SLAM towards Dynamic Environments. In Proceedings of the 2018 IEEE International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 1–5 October 2018; pp. 1168–1174. [Google Scholar]
  5. Huang, G.; Liu, Z.; Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2261–2269. [Google Scholar]
  6. Alhashim, I.; Wonka, P. High quality monocular depth estimation via transfer learning. arXiv 2018, arXiv:1812.11941. [Google Scholar]
  7. Fu, H.; Gong, M.; Wang, C.; Batmanghelich, K.; Tao, D. Deep ordinal regression network for monocular depth estimation. In Proceedings of the 2018 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2018, Salt Lake City, UT, USA, 18–22 June 2018; pp. 2002–2011. [Google Scholar]
  8. Chen, L.; Tang, W.; Wan, T.R.; Nigel, W.J. Self-supervised monocular image depth learning and confidence estimation. arXiv 2018, arXiv:1803.05530. [Google Scholar] [CrossRef] [Green Version]
  9. Hu, J.J.; Ozay, M.; Zhang, Y.; Okatani, T. Revisiting Single Image Depth Estimation: Toward Higher Resolution Maps with Accurate Object Boundaries. arXiv 2019, arXiv:1803.08673. [Google Scholar]
  10. Qi, X.; Liao, R.; Liu, Z.; Urtasun, R.; Jia, J. GeoNet: Geometric Neural Network for Joint Depth and Surface Normal Estimation. In Proceedings of the 2018 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2018, Salt Lake City, UT, USA, 18–22 June 2018; pp. 283–291. [Google Scholar]
  11. Zhang, Z.; Cui, Z.; Xu, C.; Yan, Y.; Sebe, N.; Yang, J. Pattern-affinitive propagation across depth, surface normal and semantic segmentation. In Proceedings of the 2019 IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 4106–4115. [Google Scholar]
  12. Lee, J.-H.; Kim, C.-S. Monocular Depth Estimation Using Relative Depth Maps. In Proceedings of the 2019 IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 9729–9738. [Google Scholar]
  13. Chen, L.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 40, 834–848. [Google Scholar] [CrossRef] [PubMed]
  14. Silberman, N.; Hoiem, D.; Kohli, P.; Fergus, R. Indoor segmentation and support inference from rgbd images. In Proceedings of the European Conference on Computer Vision, Florence, Italy, 7–13 October 2012; pp. 746–760. [Google Scholar]
  15. Dai, A.; Chang, A.X.; Savva, M.; Halber, M.; Funkhouser, T.; Nießner, M. ScanNet: Richly annotated 3D reconstructions of indoor scenes. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2432–2443. [Google Scholar]
  16. Chang, A.; Dai, A.; Funkhouser, T.; Halber, M.; Niessner, M.; Savva, M.; Song, S.; Zeng, A.; Zhang, Y. Matterport3D: Learning from RGB-D data in indoor environments. In Proceedings of the International Conference on 3D Vision (3DV), Qingdao, China, 10–12 October 2017; pp. 667–676. [Google Scholar]
  17. Wofk, D.; Ma, F.C.; Yang, T.J.; Karaman, S.; Vivienne, S. FastDepth: Fast Monocular Depth Estimation on Embedded Systems. arXiv 2019, arXiv:1903.03273. [Google Scholar]
  18. Nekrasov, V.; Shen, C.H.; Reid, I. Light-Weight RefineNet for Real-Time Semantic Segmentation. arXiv 2018, arXiv:1810.03272. [Google Scholar]
  19. Zeng, J.; Tong, Y.; Huang, Y.; Yan, Q.; Sun, W.; Chen, J.; Wang, Y. Deep Surface Normal Estimation With Hierarchical RGB-D Fusion. In Proceedings of the 2019 IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 6146–6155. [Google Scholar]
  20. Yin, W.; Liu, Y.; Shen, C.; Yan, Y. Enforcing Geometric Constraints of Virtual Normal for Depth Prediction. In Proceedings of the 2019 IEEE International Conference on Computer Vision, Seoul, Korea, 27–28 October 2019; pp. 5683–5692. [Google Scholar]
  21. Eigen, D.; Fergus, R. Predicting Depth, Surface Normals and Semantic Labels with a Common Multi-scale Convolutional Architecture. In Proceedings of the 2015 IEEE International Conference on Computer Vision, Santiago, Chile, 13–16 December 2015; pp. 2650–2658. [Google Scholar]
  22. Wang, X.; Fouhey, D.; GuptaIn, A. Designing deep networks for surface normal estimation. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 8–10 June 2015; pp. 539–547. [Google Scholar]
  23. Garg, R.; BG, V.K.; Carneiro, G.; Reid, I. Unsupervised CNN for single view depth estimation: Geometry to the rescue. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; pp. 740–756. [Google Scholar]
  24. Wang, C.; Miguel Buenaposada, J.; Zhu, R.; Lucey, S. Learning depth from monocular videos using direct methods. In Proceedings of the 2018 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2018, Salt Lake City, UT, USA, 18–22 June 2018; pp. 2022–2030. [Google Scholar]
  25. Yang, Z.; Wang, P.; Wang, Y.; Xu, W.; Nevatia, R. Every Pixel Counts: Unsupervised Geometry Learning with Holistic 3D Motion Understanding. arXiv 2018, arXiv:1806.10556. [Google Scholar]
  26. Zhou, T.H.; Brown, M.; Noah, S.; David, L. Unsupervised Learning of Depth and Ego-Motion from Video. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 6612–6619. [Google Scholar]
  27. Li, R.; Wang, S.; Long, Z.; Gu, D. UnDeepVO: Monocular Visual Odometry through Unsupervised Deep Learning. arXiv 2017, arXiv:1709.06841. [Google Scholar]
  28. Liu, F.; Chunhua, S.; Guosheng, L. Deep convolutional neural fields for depth estimation from a single image. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 5162–5170. [Google Scholar]
  29. Lee, J.H.; Han, M.K.; Kong, D.W.; Suh, H. From Big to Small: Multi-Scale Local Planar Guidance for Monocular Depth Estimation. arXiv 2019, arXiv:1907.10326. [Google Scholar]
  30. Janai, J.; Guney, F.; Ranjan, A.; Black, M.; Geiger, A. Unsupervised Learning of Multi-Frame Optical Flow with Occlusions. In Proceedings of the European Conference on Computer Visio (ECCV), Munich, Germany, 8–14 September 2018; pp. 713–731. [Google Scholar]
  31. Misra, I.; Shrivastava, A.; Gupta, A.; Hebert, M. Cross-stitch networks for multitask learning. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 3994–4003. [Google Scholar]
  32. Kwak, D.-H.; Lee, S.-H. A Novel Method for Estimating Monocular Depth Using Cycle GAN and Segmentation. Sensors 2020, 20, 2567. [Google Scholar] [CrossRef] [PubMed]
  33. Xu, D.; Ouyang, W.; Wang, X.; Sebe, N. PAD-Net: Multi-tasks Guided Prediction and Distillation Network for Simultaneous Depth Estimation and Scene Parsing. In Proceedings of the 2018 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2018, Salt Lake City, UT, USA, 18–22 June 2018; pp. 675–684. [Google Scholar]
  34. Jiao, J.B.; Cao, Y.; Song, Y.B.; Lau, R. Look Deeper into Depth: Monocular Depth Estimation with Semantic Booster and Attention-Driven Loss. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; Springer: Cham, Switzerland, 2016; pp. 55–71. [Google Scholar]
  35. Levin, A.; Lischinski, D.; Weiss, Y. Colorization using optimization. ACM Trans. Graph. 2004, 23, 689–694. [Google Scholar] [CrossRef] [Green Version]
  36. Deng, J.; Dong, W.; Socher, R.; Li, L.-J.; Li, K.; Li, F.-F. ImageNet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR 2009), Miami, FL, USA, 20–25 June 2009. [Google Scholar]
  37. Fouhey, D.F.; Gupta, A.; Hebert, M. Data-driven 3d primitives for single image understanding. In Proceedings of the 2013 International conference on computer vision, Sydney, Australia, 1–8 December 2013; pp. 3392–3399. [Google Scholar]
  38. Sun, L.H.; Wang, J.S.; Yun, H.; Zhu, Q.; Yin, B.C. Surface Normal Data Guided Depth Recovery with Graph Laplacian Regularization. In Proceedings of the 2019 ACM Multimedia Asia (MMAsia ′19), Beijing, China, 16–18 December 2019; pp. 1–6. [Google Scholar]
  39. Huang, J.; Lee, A.B.; Mumford, D. Statistics of range images. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Hilton Head, SC, USA, 13–15 June 2000; pp. 324–331. [Google Scholar]
  40. Xiaoyang, L. Depth2pointCloud. Available online: https://github.com/ZJULiXiaoyang/depth2pointCloud (accessed on 19 July 2020).
  41. Laina, I.; Rupprecht, C.; Belagiannis, V.; Tombari, F.; Navab, N. Deeper Depth Prediction with Fully Convolutional Residual Networks. In Proceedings of the 2016 Fourth International Conference on 3D Vision (3DV), Stanford, CA, USA, 25–28 October 2016; pp. 239–248. [Google Scholar]
  42. Hickson, S.; Raveendran, K.; Fathi, A.; Murphy, K.; Essa, I. Floors are Flat: Leveraging Semantics for Real-Time Surface Normal Prediction. arXiv 2019, arXiv:1906.06792. [Google Scholar]
  43. Paszke, A.; Gross, S.; Chintala, S.; Chanan, G.; Yang, E.; DeVito, Z. Automatic differentiation in PyTorch. In Proceedings of the Advances in Neural Information Processing Systems Workshops, Long Beach, CA, USA, 4–9 December 2017; pp. 1–4. [Google Scholar]
  44. Kingma, D.P.; Ba, J.L. Adam: A method for stochastic optimization. In Proceedings of the 2015 International Conference on Learning Representations, San Diego, CA, USA, 7–9 May 2015; pp. 1–41. [Google Scholar]
  45. Cubuk, E.; Zoph, B.; Mane, D.; Vasudevan, V.; Le, Q.V. Autoaugment: Learning augmentation policies from data. arXiv 2018, arXiv:1805.09501. [Google Scholar]
  46. Eigen, D.; Puhrsch, C.; Fergus, R. Depth Map Prediction from a Single Image Using a Multi-scale Deep Network. In Proceedings of the 27th International Conference on Neural Information Processing Systems, Montreal, QC, Canada, 8–13 December 2014; pp. 2366–2374. [Google Scholar]
  47. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, M.; Chen, L. Mobilenetv2: Inverted residuals and linear bottlenecks. In Proceedings of the 2018 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2018, Salt Lake City, UT, USA, 18–22 June 2018; pp. 4510–4520. [Google Scholar]
  48. Haris, M.; Shakhnarovich, G.; Ukita, N. Deep back-projection networks for super-resolution. In Proceedings of the 2018 IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 1664–1673. [Google Scholar]
  49. Ranftl, R.; Lasinger, K.; Hafner, D.; Schindler, K.; Koltun, V. Towards Robust Monocular Depth Estimation: Mixing Datasets for Zero-shot Cross-dataset Transfer. arXiv 2019, arXiv:1907.01341. [Google Scholar]
Figure 1. The comparison of depth maps were produced by different methods. (a) Raw red-green-blue (RGB) images (b) Ground truth (GT) depth maps [14], (c) Depth maps from the state-of-the-art (SOTA) practice [7], (d) Depth maps from our depth prediction network.
Figure 1. The comparison of depth maps were produced by different methods. (a) Raw red-green-blue (RGB) images (b) Ground truth (GT) depth maps [14], (c) Depth maps from the state-of-the-art (SOTA) practice [7], (d) Depth maps from our depth prediction network.
Sensors 20 04856 g001
Figure 2. The comparison of surface normal maps: (a) from left to right: RGB images, Ground-Truth (GT), surface normal maps produced by Qi et al. [10], ours. (b) Color-map definition: red represents left, green represents up, and blue represents outward.
Figure 2. The comparison of surface normal maps: (a) from left to right: RGB images, Ground-Truth (GT), surface normal maps produced by Qi et al. [10], ours. (b) Color-map definition: red represents left, green represents up, and blue represents outward.
Sensors 20 04856 g002
Figure 3. General estimation framework. (a) Coarse depth estimation network; (b) red-green-blue-depth (RGB-D) surface normal network; (c) refinement network.
Figure 3. General estimation framework. (a) Coarse depth estimation network; (b) red-green-blue-depth (RGB-D) surface normal network; (c) refinement network.
Sensors 20 04856 g003
Figure 4. Encoder–decoder coarse depth network.
Figure 4. Encoder–decoder coarse depth network.
Sensors 20 04856 g004
Figure 5. Generating coarse surface normal image. (a) Coarse depth ( D * ); (b)   N * .
Figure 5. Generating coarse surface normal image. (a) Coarse depth ( D * ); (b)   N * .
Sensors 20 04856 g005
Figure 6. Surface normal adjustment network (Dense-net-121 based). (a) The general structure of the RGB-D surface normal network (RSN) network; (b) the architectures of up-projection units, fusion module and convolution blocks.
Figure 6. Surface normal adjustment network (Dense-net-121 based). (a) The general structure of the RGB-D surface normal network (RSN) network; (b) the architectures of up-projection units, fusion module and convolution blocks.
Sensors 20 04856 g006
Figure 7. Comparison of point clouds from estimated depth maps between Alhashim [6] and ours. (a) Alhashim [6], (b) GT, (c) Ours. GT stands for 3D point cloud maps from ground truth images.
Figure 7. Comparison of point clouds from estimated depth maps between Alhashim [6] and ours. (a) Alhashim [6], (b) GT, (c) Ours. GT stands for 3D point cloud maps from ground truth images.
Sensors 20 04856 g007
Figure 8. Qualitative results for depth estimation. (a) RGB images, (b) ground truth (GT), (c) Alhashim [2], (d) Our Coarse; (e) Refined Depth.
Figure 8. Qualitative results for depth estimation. (a) RGB images, (b) ground truth (GT), (c) Alhashim [2], (d) Our Coarse; (e) Refined Depth.
Sensors 20 04856 g008
Figure 9. Qualitative results for surface normal estimation. (a) RGB images, (b) ground truth normal maps, (c), Geo-Net [10], (d) reconstructed from in-painted ground truth depth (e) ours, and (f) reconstructed from refined depth. All images are equally scaled for better visualization.
Figure 9. Qualitative results for surface normal estimation. (a) RGB images, (b) ground truth normal maps, (c), Geo-Net [10], (d) reconstructed from in-painted ground truth depth (e) ours, and (f) reconstructed from refined depth. All images are equally scaled for better visualization.
Sensors 20 04856 g009
Figure 10. Comparison of time consumption: Runtimes of different up-sampling methods including. 2× bilinear interpolation [6], up and down projection [48], and up-projection [17].
Figure 10. Comparison of time consumption: Runtimes of different up-sampling methods including. 2× bilinear interpolation [6], up and down projection [48], and up-projection [17].
Sensors 20 04856 g010
Figure 11. Refined depth images generating from custom images (densenet-161 model).
Figure 11. Refined depth images generating from custom images (densenet-161 model).
Sensors 20 04856 g011
Table 1. Comparison of depth estimation accuracy with previous methods on NYU Depth-V2 [14].
Table 1. Comparison of depth estimation accuracy with previous methods on NYU Depth-V2 [14].
Method Error (Lower Is Better)Accuracy (Higher Is Better)
AbsREL.RMSE.ELog10.δ < 1.25δ < 1.252δ < 1.253
Eigen [21]0.1580.641-0.7690.9500.988
Fu [7]0.1150.5090.0510.828 0.9650.992
Alhashim [6] (Densenet-169)0.1230.4650.0530.8460.9740.994
Hu [9] (Densenet-161)0.1230.5440.0530.855 0.9720.993
Ours (Densenet-121-coarse)0.1370.5720.0560.8390.9620.988
Ours (Densenet-121-refined)0.1220.4590.0510.8590.9720.993
Table 2. Comparison of surface normal accuracy with previous methods on NYU Depth-V2 [14].
Table 2. Comparison of surface normal accuracy with previous methods on NYU Depth-V2 [14].
MethodError (Lower Is Better)Accuracy (Higher Is Better)
Mean.Median.RMSE. 11.25 ° 22.5 ° 30 °
3DP(MW) [37]36.319.246.639.252.957.8
Wang [22]26.914.8-42.061.268.2
Qi [10]19.011.826.948.471.579.5
Ours (RGB-D fusion)20.611.025.647.973.281.8
Table 3. Comparison of computational efficiency and performance on the NYU dataset [14]. The “Parm.’”stands for training parameters, and “Iters.” represents the number of training iterations.
Table 3. Comparison of computational efficiency and performance on the NYU dataset [14]. The “Parm.’”stands for training parameters, and “Iters.” represents the number of training iterations.
MethodRMSE.FramesEpochsTraining
Time
(h)
Iters.Inference
Time
(s)
Parms.
Fu [7]0.509120K--3M-110M
Alhashim [6]0.46550K20201M0.26542.6M
Hu [9]
(Se-Net)
0.53050K20-1M0.352-
Ours0.45930K2018600k0.21732.4M
Table 4. Ablation study of encoder layers. Parm. stands for training parameters; DN stands for Dense-net.
Table 4. Ablation study of encoder layers. Parm. stands for training parameters; DN stands for Dense-net.
MethodParm.Error (Lower Is Better)Accuracy (Higher Is Better)
Abs-RELRMSE.ELog10.δ < 1.25δ < 1.252δ < 1.253
DN-121-Refined32.4 M0.1220.4590.0510.8590.9720.993
Dense-depth [2] (DN-169)42.6 M0.1230.4650.0530.8460.9740.994
DN-161-Refined67.2 M0.1160.4460.0490.8670.9760.994
Table 5. Ablation study of encoder layer numbers. Parm. stands for training parameters.
Table 5. Ablation study of encoder layer numbers. Parm. stands for training parameters.
MethodParm.Error (Lower Is Better)Accuracy (Higher Is Better)
AbsRELRMSE.ELog10.δ < 1.25δ < 1.252δ < 1.253
Ours(121-refined)32.4 M0.1220.4590.0510.8590.9720.993
Ours(V2-Refined)4.7 M0.1960.5190.1460.8110.9510.979

Share and Cite

MDPI and ACS Style

Huang, K.; Qu, X.; Chen, S.; Chen, Z.; Zhang, W.; Qi, H.; Zhao, F. Superb Monocular Depth Estimation Based on Transfer Learning and Surface Normal Guidance. Sensors 2020, 20, 4856. https://doi.org/10.3390/s20174856

AMA Style

Huang K, Qu X, Chen S, Chen Z, Zhang W, Qi H, Zhao F. Superb Monocular Depth Estimation Based on Transfer Learning and Surface Normal Guidance. Sensors. 2020; 20(17):4856. https://doi.org/10.3390/s20174856

Chicago/Turabian Style

Huang, Kang, Xingtian Qu, Shouqian Chen, Zhen Chen, Wang Zhang, Haogang Qi, and Fengshang Zhao. 2020. "Superb Monocular Depth Estimation Based on Transfer Learning and Surface Normal Guidance" Sensors 20, no. 17: 4856. https://doi.org/10.3390/s20174856

APA Style

Huang, K., Qu, X., Chen, S., Chen, Z., Zhang, W., Qi, H., & Zhao, F. (2020). Superb Monocular Depth Estimation Based on Transfer Learning and Surface Normal Guidance. Sensors, 20(17), 4856. https://doi.org/10.3390/s20174856

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop