Next Article in Journal
Enhancing Atmospheric Monitoring Capabilities: A Comparison of Low- and High-Cost GNSS Networks for Tropospheric Estimations
Previous Article in Journal
Accurate Calculation of Upper Biomass Volume of Single Trees Using Matrixial Representation of LiDAR Data
Previous Article in Special Issue
Vision through Obstacles—3D Geometric Reconstruction and Evaluation of Neural Radiance Fields (NeRFs)
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

SIM-MultiDepth: Self-Supervised Indoor Monocular Multi-Frame Depth Estimation Based on Texture-Aware Masking

1
School of Instrumentation and Optoelectronic Engineering, Key Laboratory of Precision Opto-Mechatronics Technology, Ministry of Education, Beihang University, Beijing 100191, China
2
Qingdao Research Institute, Beihang University, Qingdao 266104, China
3
Institute of Artificial Intelligence, Beihang University, Beijing 100191, China
4
Aerospace Optical-Microwave Integrated Precision Intelligent Sensing, Key Laboratory of Ministry of Industry and Information Technology, Beihang University, Beijing 100191, China
5
School of Automation Science and Electrical Engineering, Beihang University, Beijing 100191, China
6
Hangzhou Research Institute, Beihang University, Hangzhou 310051, China
7
Nanchang Institute of Technology, Nanchang 330044, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2024, 16(12), 2221; https://doi.org/10.3390/rs16122221
Submission received: 8 April 2024 / Revised: 10 June 2024 / Accepted: 14 June 2024 / Published: 19 June 2024
(This article belongs to the Special Issue Photogrammetry Meets AI)

Abstract

:
Self-supervised monocular depth estimation methods have become the focus of research since ground truth data are not required. Current single-image-based works only leverage appearance-based features, thus achieving a limited performance. Deep learning based multiview stereo works facilitate the research on multi-frame depth estimation methods. Some multi-frame methods build cost volumes and take multiple frames as inputs at the time of test to fully utilize geometric cues between adjacent frames. Nevertheless, low-textured regions, which are dominant in indoor scenes, tend to cause unreliable depth hypotheses in the cost volume. Few self-supervised multi-frame methods have been used to conduct research on the issue of low-texture areas in indoor scenes. To handle this issue, we propose SIM-MultiDepth, a self-supervised indoor monocular multi-frame depth estimation framework. A self-supervised single-frame depth estimation network is introduced to learn the relative poses and supervise the multi-frame depth learning. A texture-aware depth consistency loss is designed considering the calculation of the patch-based photometric loss. Only the areas where multi-frame depth prediction is considered unreliable in low-texture regions are supervised by the single-frame network. This approach helps improve the depth estimation accuracy. The experimental results on the NYU Depth V2 dataset validate the effectiveness of SIM-MultiDepth. The zero-shot generalization studies on the 7-Scenes and Campus Indoor datasets aid in the analysis of the application characteristics of SIM-MultiDepth.

1. Introduction

Depth estimation has been widely applied in many downstream computer vision tasks, such as automatic driving, robot navigation, and augmented reality. Compared with obtaining depth values from active sensors, monocular depth estimation using a single camera is much more flexible. In terms of whether labeled data are required during training, monocular depth estimation methods are divided into supervised and self-supervised ones. Because of the benefit of not requiring ground truth, self-supervised methods have become the focus of research.
Many self-supervised monocular depth estimation methods [1,2,3] adopt a single image as the test input. Single-frame methods rely on appearance-based cues to predict scene depth. Although multiple frames are used in training, the geometric matching information between adjacent frames is still ignored while testing. To fully utilize geometric cues, some studies [4,5,6] have used multiple frames to estimate the depth at the time of the test. Taking inspiration from deep learning based multiview stereo (MVS) [7], these methods build 3D cost volumes based on depth hypotheses and differentiable warping. However, dynamic and low-textured areas often cause wrong matchings to occur, which results in mistakes in the cost volume. Further, the whole training process is subject to overfitting. With photometric stereo methods [8,9], in which the 3D surface normal is recovered from multiple images with directional lighting, such low-textured areas, caused by non-Lambertian reflection, can be mitigated by the use of max-pooling layers, which fuse visual cues from multiple images [10]. However, it is difficult for such methods to adapt to complex lighting conditions in the estimation of depth. To solve these problems, researchers [4,11,12,13] have leveraged a single-frame depth estimation network to predict camera pose and supervise the training of the multi-frame network. However, current multi-frame methods mainly focus on predicting depth on outdoor datasets such as KITTI [14] and Cityscapes [15]. Few self-supervised multi-frame works perform depth estimation on indoor datasets such as NYU Depth V2 [16] and 7-Scenes [17].
We propose a self-supervised indoor monocular multi-frame depth estimation framework called SIM-MultiDepth. Following the ManyDepth [4] architecture, we build a cost volume with feature map warping and depth hypotheses. The cost volume and target frame feature are then processed by the depth decoder to predict depth.
In indoor depth estimation, a great challenge concerns the existence of many low-textured areas, such as tables, walls, and floors. The multi-frame network suffers severely from such regions and often falls into cost volume overfitting. Inspired by previous methods, we introduce a single-frame network to calculate camera poses for the multi-frame network to solve the overfitting issue. Additionally, the single-frame network supervises the multi-frame depth learning in selected areas to improve the overall performance. Figure 1 provides a comparison of the performances of the methods.
Different from ManyDepth, we designed a new single-frame supervision method. We selected the intersection of two parts of areas as the supervised region. The first part includes the pixels not involved in calculating the patch-based photometric loss. These areas can be considered as lacking unique textures, i.e., low-textured areas. The second part includes those areas where multi-frame depth prediction results are considered unreliable. In this way, the low-textured areas for which multi-frame prediction may be inaccurate will be supervised by single-frame depth. The proposed texture-aware masking manner has proven effective in the experiments on the NYU Depth V2 dataset [16]. The experiments, regarding some limitations, were performed to illustrate the capabilities and constraints of the framework. Zero-shot generalization experiments were also conducted on the 7-Scenes [17] and our collected Campus Indoor datasets [18]. The scenarios applicable for multi-frame and single-frame networks were further investigated.
In brief, the following three contributions were made:
  • We propose a novel self-supervised indoor monocular multi-frame depth estimation framework called SIM-MultiDepth. To solve the overfitting problem and achieve a better performance in low-textured areas, a single-frame depth estimation network is introduced to compute camera poses and serve as supervision.
  • Considering the patch-based photometric loss, a texture-aware masking supervision strategy is well-designed. The corresponding depth consistency loss guarantees that the points with discriminative features are only involved in geometric reasoning, instead of being forced to be consistent with the single-frame depth.
  • The experiments on the NYU Depth V2 dataset illustrate the effectiveness of our proposed SIM-MultiDepth and the texture-aware masking strategy. All evaluation metrics improved compared with the single-frame method. The generalization experiments on the 7-Scenes and Campus Indoor datasets also reveal the characteristics of the multi-frame and single-frame methods.

2. Related Work

2.1. Indoor Monocular Single-Frame Depth Estimation

Monocular single-frame depth estimation aims to predict the depths of pixels in a single image. Many supervised methods [19,20,21,22,23,24,25] use pixel-wise regression to estimate depth. IEBins [26] treat depth estimation as classification–regression and leverages iterative elastic bins to calculate accurate depths. Recent advances in the attention mechanism have led many studies [27,28,29,30,31,32,33,34,35] to replace previously used Convolutional Neural Network (CNN) modules with Transformer [36]. Diffusion models have also been applied in supervised monocular depth estimation tasks [37,38]. Since there often exists high geometric regularity in indoor scenes, many supervised works design appropriate geometry-related loss terms to achieve better performance. Surface normal loss is applied in some works [39,40]. Plane-aware constraints are also introduced in many methods, such as PlaneNet [41], PlaneReg [42], P3Depth [43], and NDDepth [44].
Self-supervised monocular single-frame depth estimation methods usually adopt photometric consistency loss as the mainly used loss term. SfMLearner [1] first proposes self-supervised depth learning with view synthesis as supervision. Monodepth2 [2] is another classical method that reduces the impact of occlusion and moving object issues in auto-driving datasets such as KITTI. However, Monodepth2 fails to achieve a satisfactory performance in indoor scenarios because there exist many low-textured regions. Simple photometric consistency loss can lead to a large number of false matchings. To improve accuracy on indoor datasets, P2Net [3] adopts patch-based photometric consistency loss and planar consistency loss based on Monodepth2 architecture. Similarly, StructDepth [45] uses Manhattan normal loss and coplanar loss as extra supervision. F2Depth [18] leverages an improved optical flow network to provide accurate pixel motion in low-textured areas. Based on SC-Depth V1 [46], SC-Depth V2 [47] introduces an Auto-Rectify Network to handle the issue of excessive camera rotation in indoor scenes. For the same problem related to camera rotation, MonoIndoor [48] and MonoIndoor++ [49] use a residual pose estimation module after the initial pose estimation. From the expert network DPT [50,51], DistDepth [52] distills structure to the student network. GasMono [53] utilizes the structure-from-motion tool COLMAP [54] to perform coarse pose estimations and enhances feature processing in low-textured regions with Transformer.
For monocular depth estimation from a single image, the sequence information between adjacent frames is ignored. Thus, monocular multi-frame depth estimation that makes use of geometric cues has received more attention.

2.2. Monocular Multi-Frame Depth Estimation

Different from single-frame depth estimation, monocular multi-frame depth estimation exploits sequential multiple frames to estimate scene depth. Test-time refinement methods (e.g., [55]) iteratively refine the single-frame model for global temporal consistency during testing. These kinds of approaches will increase the computational cost. Another group of methods (e.g., [56]) are based on recurrent networks. The recently proposed MAMo [57] leverages memory and an attention mechanism to augment single-frame depth estimation networks into multi-frame depth estimation. Although these works use multiple frames to estimate depth at the time of test, geometric information is still not utilized explicitly.
To perform explicit geometry reasoning, some multiview stereo (MVS) studies [7,58,59] construct a 3D cost volume by image differentiable warping. However, these methods assume camera poses are known for training. For monocular depth estimation based on consecutive frames, the relative poses between adjacent frames are usually unknown.
Some supervised approaches [60,61] require relative pose as input to cost volumes at test time. Some methods [62,63] do not need known camera poses, but the number of frames used is unchangeable. Dynamic areas in outdoor scenarios tend to violate multiview consistency and cause false depth hypotheses in the cost volume. Recently, in [64] dynamic areas were managed by proposing a cross-cue fusion (CCF) module. The CCF module fuses single-frame and multi-frame cues to enhance both representations.
To eliminate the dependence on ground truth and known poses, ManyDepth [4] proposes a multi-frame depth estimation network based on single-frame Monodepth2 architecture. ManyDepth can process a variable number of frames during testing. For the overfitting problem in dynamic and low-textured areas, ManyDepth leverages the single-frame network to supervise the training of the multi-frame network. Following ManyDepth, DynamicDepth [5] proposes a Dynamic Object Cycle Consistent scheme to deal with the dynamic area problem. An Occlusion-aware Cost Volume is also designed to handle the occlusion issue. IterDepth [6] formulates depth estimation as residual learning and refines depth prediction results through iterative residual refinement.
Similar to single-frame depth estimation, Transformer architecture is also applied in multi-frame depth estimation. DepthFormer [65] enhances feature matching between adjacent frames during the generation of the cross-attention cost volume. Apart from attention-based modules, Dyna-DepthFormer [66] was proposed to estimate the motion field of moving objects to meet the challenge of dynamic areas.
Since single-frame depth estimation mainly leverages appearance-based features, while multi-frame methods rely on geometric constraints, some works tend to combine both to improve accuracy. MOVEDepth [12] adopts single-frame depth prediction results and velocity cues prior to the search for multi-frame depths. Furthermore, MOVEDepth fuses single-frame and multi-frame depths through uncertainty learning in the cost volume. In [13], a hybrid decoder consisting of single-frame and multi-frame pathways is proposed, and these two kinds of information are then fused in a multistage manner. In addition, [11] proposes a self-supervised framework in which single-frame and multi-frame depths supervise each other to fully leverage the mutual influence between them.
Almost all of the works mentioned above focus only on multi-frame depth estimation in outdoor scenes. Few researchers have proposed approaches to indoor monocular multi-frame depth estimation. Although MAMo [57] conducts experiments on the indoor NYU Depth V2 dataset, it is a supervised method and lacks explicit geometric reasoning.
Inspired by ManyDepth, we propose a self-supervised indoor monocular multi-frame depth estimation network called SIM-MultiDepth. In terms of the loss function, the manner in which the single-frame depth supervises the multi-frame depth is well-designed.

3. Methods

In this section, we introduce the proposed self-supervised indoor monocular multi-frame depth estimation framework, SIM-MultiDepth. The target frame and its source frame are taken as inputs to the PoseCNN to estimate the relative camera poses. Based on the feature map’s differentiable warping and hypothesized depth values, a cost volume is constructed. Then, the features in the target image and the cost volume are processed by the encoder and depth decoder to generate the depth results. To address overfitting caused by the cost volume, we updated the PoseCNN parameters relying only on the single-frame network prediction results. In addition, the generated depths by the single-frame network were used to supervise the multi-frame depth learning in precisely selected areas. The contributions guarantee the performance of SIM-MultiDepth both in textured and low-textured areas.
In Section 3.1, the overall multi-frame depth estimation network architecture is explained. Section 3.2 introduces the well-designed texture-aware depth consistency loss, and the overall loss function is illustrated in Section 3.3.

3.1. Overview of SIM-MultiDepth

We aimed for it to perform geometric reasoning with adjacent frames, usually called source frames, to estimate the target frame depth. A pipeline diagram of SIM-MultiDepth is shown in Figure 2.
Multiple frames are processed by building a cost volume that fuses their features. The target frame, I t R H × W × 3 , and the source frame, I s R H × W × 3 , are input into the PoseCNN to compute the relative camera pose, T t s . The target frame feature map, F t R H / 4 × W / 4 × C f , and the source frame feature map, F s R H / 4 × W / 4 × C f , are extracted by the encoder. C f is the number of feature map channels.
We suppose that there are N d depth values, written as d i . For each d i , the synthesized feature map, F s t i , is the following:
F s t i = F s p r o j d i , T t s , K
where K denotes the camera intrinsic parameters matrix, denotes the sampling operation, p r o j ( ) describes the projection relationship between the predicted depth and the 2D homogeneous coordinates in the feature maps. In the cost volume, C V R H / 4 × W / 4 × N d , every depth value, d i , corresponds to C V i R H / 4 × W / 4 × 1 , which can be calculated by the following:
C V i = 1 C f c = 1 C f ( | F t ( c ) F s t i ( c ) | )
Briefly, the cost volume describes the probability that the pixel (u, v) corresponds to the depth value d i . Then, the cost volume concatenated with the target frame feature is taken as input into the depth decoder to compute the depth.
We can observe that the training process falls into overfitting because there are many low-textured areas in indoor scenarios. In these areas, the information in the cost volume becomes untrustworthy. Mistakes in the cost volume affect the depth prediction results for both training and testing. Different from the multi-frame network’s overreliance on geometric matching information, a single-frame network tends to leverage appearance cues to estimate depth. The single-frame network is more robust in low-textured areas.
To solve the overfitting problem of the multi-frame network, we introduce a trained single-frame depth estimation network to provide pose information. The multi-frame network does not participate in updating the parameters of the PoseCNN. In order to further improve the accuracy, the depth results of the single-frame network serve as supervision for the multi-frame depth learning in some specific areas. In Section 3.2, we show how the supervised areas are selected.

3.2. Texture-Aware Depth Consistency Loss

To improve the overall depth estimation results, an effective single-frame supervision strategy was designed. We precisely select those pixels with depth predictions that are unreliable in low-textured areas to be supervised by the single-frame depth.
Firstly, we need to determine which pixels can be considered as lacking in discriminative textures. The loss function contains a patch-based photometric consistency loss. The photometric loss involves pixels around extracted key points with significant features. The pixels are distributed in 3 × 3 patches centered at each key point, which can be written as the following:
P = x i + x k , y i + y k , x k N , 0 , N , y k N , 0 , N , 1 i n , i N
where ( x i , y i ) are the 2D coordinates of the key point, n is the number of key points, and N is set to 2. We set the binary mask, M 1 , to 1 in the following regions, except for the pixels in P :
M 1 = p P
As shown in Figure 3, the black patches in the images have distinctive features while the white areas are considered to be nondiscriminative.
Since the overall loss function has been elaborately designed for indoor scenes, not all pixels in low-textured areas will produce wrong depth results. The supervision from the single-frame depth should be more targeted. There are some unreliable prediction results that can be formulated, as follows:
M 2 = max D c v D ^ t D ^ t , D ^ t D c v D c v > 1
where M 2 is a binary mask representing which depth values are predicted to potentially be wrong; D ^ t is the depth generated by the single-frame network. D c v is calculated by the following:
D c v = arg min ( C V )
The final texture-aware supervision mask, M , is computed as follows:
M = M 1 M 2
Intuitively, M is set to 1 at the intersection of pixels producing unreliable depths and those not involved in the patch-based photometric loss.
The texture-aware depth consistency loss term is written as follows:
L c o n s i s t e n c y = D t ( p t ) D ^ t ( p t )
where p t represents the pixels, M is set to 1, and D t is the predicted depth of the multi-frame network.
ManyDepth [4] adopts a similar single-frame supervisory approach. The difference is that the multi-frame network is supervised to approach single-frame depths in all M2 areas. The disadvantage of this strategy is that some points with significant features are forced to approach wrong single-frame prediction results. These points could have achieved better accuracy through geometric inferencing of the multi-frame network. The advantage of selecting the intersection of M 1 and M 2 is that it not only ensures that almost all strong feature points participate in the geometric reasoning of the multi-frame network but also filters out points that are identified as being unreliable. The benefits of the multi-frame depth estimation network and the single-frame network combine well to achieve a better performance.

3.3. Overall Loss Functions

For the multi-frame depth estimation network SIM-MultiDepth, the final loss function is defined as follows:
L = L p h + λ 1 L s m + λ 2 L s p p + L r i g i d + λ 3 L f e a t u r e + L c o n s i s t e n c y
where L p h denotes the patch-based photometric consistency loss, L s m is the smoothness loss, and L s p p represents the planar consistency loss. These three loss terms have the same definitions as in P2Net [3]. Moreover, λ 1 is set to 0.001, λ 2 is set to 0.05, and λ 3 is set to 3 in the experiments.
The patch-based photometric consistency loss, L p h , is formulated as follows:
L L 1 = I t P i t I s P i t s 1
L S S I M = S S I M ( I t [ P i t ] , I s [ P i t s ] )
L p h = α L S S I M + ( 1 α ) L L 1
where P is as defined in Equation (3), and α is set to 0.85.
The smoothness loss, L s m , is formulated as follows:
L s m = | x d t * | e | x I t | + | y d t * | e | y I t |
where d t * = d t / d ¯ t denotes the normalized depth prediction.
The planar consistency loss, L s p p , is formulated as follows:
L s p p = m = 1 M n = 1 N | D ( p n ) D ( p n ) |
where L s p p is as proposed in P2Net [3], M denotes the number of segmented plane super-pixels, and N denotes the number of pixels in every plane super-pixel. L s p p helps improve the accuracy in the planar regions.
L r i g i d is our previously proposed optical flow consistency loss for the single-frame depth estimation network [18]. It guarantees that the optical flow generated from the depth estimation network is more accurate. The loss term is as follows:
L r i g i d = f r i g i d ( p t ) f f l o w ( p t )
where f f l o w ( p t ) stands for the optical flow generated by the introduced flow estimation network.
The multiscale feature map synthesis loss, L f e a t u r e , is another previously designed constraint for the single-frame network [18], which is calculated as follows:
L f e a t u r e = F s r c ( p t ) F ^ ( p t )
where F s r c ( p t ) denotes feature maps of the source frame computed by the flow estimation network, and F ^ ( p t ) denotes warped feature maps by the optical flow of the depth estimation network.

4. Experiments

4.1. Implementation Details

Datasets: The multi-frame depth estimation network SIM-MultiDepth was trained and evaluated on the publicly available NYU Depth V2 dataset [16]. NYU Depth V2 is made up of 582 indoor scenes, which were captured by Microsoft Kinect. We adopted the same train split as in previous works [3,67]. The training dataset contains 21,483 images recorded in 283 scenes. All training images were undistorted and sampled at intervals of every 10 frames. The images { I t 1 , I t , I t + 1 } were used for training. Our SIM-MultiDepth was evaluated on the official test split consisting of 654 labeled target images, I t , and their source images, I t 1 . We selected the 10th frame before the target image as the source image. The images { I t , I t 1 } participated in constructing the cost volume during both the training and evaluation.
The 7-Scenes dataset [17] is composed of 7 indoor scenes. The official test video sequences are provided. For the zero-shot generalization experiments, we sampled the test split at intervals of 10 frames. The final test set contains 1700 images.
In addition to 7-Scenes, we also conducted zero-shot generalization studies on the Campus Indoor dataset collected by us [18]. The Campus Indoor dataset contains 99 images of 18 indoor scenes on a campus. Monocular images were captured using a FUJIFILM X-T30 camera. For calculating depth, 13–16 points were selected from each image, forming a total of approximately 1500 points. The selected points are distributed evenly over a large depth range. As shown in Figure 4, the red stars represent the points. Their ground truth values were measured using a laser rangefinder.
Experimental setup: The SIM-MultiDepth architecture is based on ManyDepth [4]. The adopted single-frame network is our previously proposed monocular depth estimation network, F2Depth [18]. The parameters of the single-frame network were frozen. Training images were color augmented and flipped randomly. The images are of size 288 × 384 for the training. The SIM-MultiDepth was trained on a single NVIDIA GeForce RTX 3090 GPU. We trained for 23 epochs with a batch size of 12. The learning rate was 10−4 without adjustment.
Evaluation metrics: Similar to previous research [3,18,67], we used a median-scaling strategy, since there exists scale ambiguity. The evaluation metrics include mean absolute relative error ( A b s   R e l ), root mean squared error ( R M S ), mean log10 error, and three accuracies under different thresholds (1.25, 1.252, and 1.253). The metrics were calculated according to the following:
A b s   R e l = 1 N i N | D i - D ^ i | D ^ i
R M S = 1 N i N | | D i D ^ i | | 2
M e a n   log 10 = 1 N i N | log 10 D i log 10 D ^ i |
A c c u r a c i e s = %   o f   D i   s . t .   max D i D ^ i , D ^ i D i = δ   <   t h r
where D ^ denotes the ground truth, D stands for the predicted depth results, N represents the total number of pixels, and t h r is the threshold for calculating the accuracies.

4.2. Results

4.2.1. Evaluation Results on NYU Depth V2

An evaluation was conducted using the official test set of NYU Depth V2 to compare with other supervised and self-supervised methods. To our knowledge, no multi-frame methods have been evaluated using the same test set, so only single-frame methods were selected for comparison. The results are listed in Table 1. Compared with our single-frame network, F2Depth, the proposed multi-frame method, SIM-MultiDepth, achieved significant improvements in all metrics. The results of the evaluation indicate the effectiveness of SIM-MultiDepth in indoor scenarios.
Figure 5 shows the predicted depth maps of the single-frame network P2Net, F2Depth, and our multi-frame network SIM-MultiDepth. From the visualization of the results, it can be observed that SIM-MultiDepth produced more accurate and detailed depth estimation results than F2Depth and P2Net. Not only arethe prediction results for significantly textured areas, such as chairs and sofas, satisfactory, but the depth results for low-textured areas, like tables, arealso good.
For some of the limitations that may be present in indoor scenes, we provide a comprehensive analysis to illustrate the capabilities and constraints of SIM-MultiDepth.
For transparent objects such as glass, we selected a bathroom with a glass door to evaluate SIM-MultiDepth’s performance. The results of the evaluation are shown in Figure 6. Theoretically, there should be no large changes in depth on either side of the green line in the RGB image. The predictions by SIM-MultiDepth aremuch better than the ground truth. The ground truth were captured with a Microsoft Kinect device. The infrared light emitted to the glass was not reflected back to the camera correctly.
For reflective surfaces, we selected a reflective television screen as an example. The prediction results are shown in Figure 7. There is no significant change in depth in the reflective areas of the screen. These results illustrate that SIM-MultiDepth is robust to interference from reflections.
Changes in illumination are another challenge that cannot be ignored. We multiplied the two frames, { I t , I t 1 } , by the same brightness factor, b , to simulate common changes in illumination in indoor scenes; b was set to 0.5, 0.8, 1, 1.2, 1.5, and 2. The corresponding illumination changes are shown in Figure 8.
We evaluated SIM-MultiDepth on the test set with the six brightness factors. The results of the evaluation are listed in Table 2. Compared with the original test set, the results are almost unchanged when b is set to 0.8 and 1.2. This is largely because the training data augmentation randomly selects b from [0.8, 1.2]. If b falls outside of [0.8, 1.2], the prediction results will be significantly worse. To handle this issue, the value range for b can be further expanded in the augmentation of the training data. A semantic consistency loss can also be designed to meet the challenge of changes in illumination.
Additionally, experiments focusing on baseline changes of I t and I t 1 were performed. I t and I t 1 in the original test set were selected at intervals of 10 frames. Here, we selected them at intervals of 5, 20, 40, 80, and 160 frames. Correspondingly, five experiments were conducted, in all of which exactly the same results as the original test set were achieved. The results verify that SIM-MultiDepth is highly adaptable to baseline changes.

4.2.2. Zero-Shot Generalization Results on 7-Scenes

To evaluate the generalization abilities of SIM-MultiDepth, we conducted zero-shot generalization studies on the 7-Scenes dataset. A comparison of the results for SIM-MultiDepth and F2Depth is provided in Table 3.
The SIM-MultiDepth achieved a similar generalization performance as F2Depth. In some scenarios, such as Stairs and Office, SIM-MultiDepth outperformed F2Depth. In other scenes, such as with Pumpkin and Heads, F2Depth generalized better than SIM-MultiDepth. To analyze the respective applicable scenarios of F2Depth and SIM-MultiDepth, we visualized their predicted depths in different scenes, as shown in Figure 9.
From Figure 9, it can be found that the multi-frame network SIM-MultiDepth generalized better in regions with significant features or sharp color changes. A typical example is that SIM-MultiDepth could better distinguish differences in the depths between the steps on the set of stairs than F2Depth. But the single-frame network F2Depth made better predictions in low-textured areas, such as cupboard doors, walls, and floors.
To investigate the cause of the variance in the performances on the 7-Scenes dataset, we analyzed the data characteristics for the Pumpkin and Heads for which F2Depth outperformed SIM-MultiDepth. The Pumpkin scene is mainly composed of low-textured planes, as shown in Figure 9. The Heads also has a low-textured desktop as its main component, as shown in Figure 10.

4.2.3. Zero-Shot Generalization Results on Campus Indoor

In addition to the 7-Scenes dataset, we also conducted generalization studies on the Campus Indoor dataset to provide a more comprehensive analysis. The experiment results are listed in Table 4.
SIM-MultiDepth and F2Depth exhibited similar performances on the Campus Indoor dataset, similar to the 7-Scenes dataset. SIM-MultiDepth surpassed F2Depth on nine scenes, while F2Depth performed better on the remaining nine scenes. In scene Nos. 6 and 7, SIM-MultiDepth displayed a particularly superior performance, as visualized in Figure 11. The depths of the pillars were predicted much more reliably by SIM-MultiDepth. Moreover, SIM-MultiDepth predicted the changes in depth for the entire scene more accurately, as framed in Figure 11. Specifically, in No. 6, the depth to the right of the green dotted line should be larger than that of the left; SIM-MultiDepth correctly predicted this. Also, for No. 7, the depths above and below the green line should almost be the same. Although these two methods both failed to make perfect predictions, SIM-MultiDepth reduced the difference in depths between the two parts.
For scene No. 5, F2Depth was the most superior to SIM-MultiDepth. It can be seen that the overall performance of SIM-MultiDepth was good. We determined the points that caused the main gap in the performances, which are represented by the white stars in Figure 12. Compared with scene Nos. 6 and 7, the scene structure of No. 5 is relatively simple. The main parts of No. 5 are low-textured walls and floors. We noticed that F2Depth generalized better for the entire wall than SIM-MultiDepth. White stars are distributed where SIM-MultiDepth generalized unsatisfactorily, which is the upper part of the wall. Since the Campus Indoor dataset is mainly made up of images of public spaces instead of private rooms, large areas with white walls exist. SIM-MultiDepth did not perform that satisfactorily in these regions.

4.2.4. Ablation Studies on NYU Depth V2

To validate the effectiveness of the proposed supervision mask, M , in texture-aware depth consistency loss, we conducted ablation studies on the NYU Depth V2. In Table 5, we present the prediction results without the texture-aware depth consistency loss in the first row. The evaluation results corresponding with M 1 and M 2 are reported in the second and third rows. The fourth row presents the results of our well-designed mask. It can be observed that the proposed texture-aware masking had positive effects on the multi-frame depth estimations.

5. Conclusions and Discussions

In this paper, we presented a self-supervised monocular multi-frame depth estimation framework for indoor scenes, called SIM-MultiDepth. We built a cost volume to perform geometric reasoning for depth estimation. For the overfitting problem caused by the cost volume, we introduced a trained single-frame network to compute the relative poses. To improve the prediction accuracy, a single-frame network was also leveraged to supervise the multi-frame network. We design a supervision strategy to precisely supervise necessary areas. The intersections of low-textured areas and areas producing unreliable depths were selected to be supervised by the single-frame network. The texture-aware masking approach allowed for pixels with unique textures to only participate in multi-frame geometric reasoning and without being consistent with the single-frame depth. The useful cues of both the single-frame and multi-frame depth estimation were fused in our proposed SIM-MultiDepth. The contributions were validated as effective in the experiments on the NYU Depth V2 dataset. In the generalization studies, SIM-MultiDepth outperformed the single-frame network in most areas, except low-textured areas. It can be concluded that the single-frame network relying on appearance cues still has better generalization abilities in low-textured areas. Future work will focus on building more robust cost volumes to further handle correspondence ambiguity. The attention-based Transformer module will be applied to enhance feature learning. More reliable loss terms will also be exploited to address problems related to indoor multi-frame depth estimation.

Author Contributions

Conceptualization, H.Z., X.G. and B.Z.; methodology, S.S.; software, X.G.; formal analysis, X.G. and S.S.; investigation, X.G.; resources, H.Z. and X.L.; data curation, X.G.; writing—original draft preparation, X.G.; writing—review and editing, S.S., N.L. and X.L.; visualization, X.G.; supervision, H.Z. and N.L.; project administration, N.L., X.L. and B.Z.; funding acquisition, H.Z. and B.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Application Innovation Project of CASC (grant number: 6230109004); National Key Research and Development Program of China (grant number: 2023YFC3300029); Zhejiang Provincial Natural Science Foundation of China (grant number: LD24F020007); “One Thousand Plan” projects in Jiangxi Province (grant number: Jxsg2023102268).

Data Availability Statement

The Campus Indoor dataset is not publicly available because of the campus confidentiality policy and privacy issues.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Zhou, T.; Brown, M.; Snavely, N.; Lowe, D.G. Unsupervised learning of depth and ego-motion from video. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1851–1858. [Google Scholar]
  2. Godard, C.; Mac Aodha, O.; Firman, M.; Brostow, G.J. Digging into self-supervised monocular depth estimation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 3828–3838. [Google Scholar]
  3. Yu, Z.; Jin, L.; Gao, S. P2Net: Patch-match and plane-regularization for unsupervised indoor depth estimation. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Proceedings, Part XXIV. pp. 206–222. [Google Scholar]
  4. Watson, J.; Mac Aodha, O.; Prisacariu, V.; Brostow, G.; Firman, M. The temporal opportunist: Self-supervised multi-frame monocular depth. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 19–25 June 2021; pp. 1164–1174. [Google Scholar]
  5. Feng, Z.; Yang, L.; Jing, L.; Wang, H.; Tian, Y.; Li, B. Disentangling object motion and occlusion for unsupervised multi-frame monocular depth. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; pp. 228–244. [Google Scholar]
  6. Feng, C.; Chen, Z.; Zhang, C.; Hu, W.; Li, B.; Lu, F. IterDepth: Iterative residual refinement for outdoor self-supervised multi-frame monocular depth estimation. IEEE Trans. Circuits Syst. Video Technol. 2023, 34, 329–341. [Google Scholar] [CrossRef]
  7. Xu, H.; Zhou, Z.; Qiao, Y.; Kang, W.; Wu, Q. Self-supervised multi-view stereo via effective co-segmentation and data-augmentation. Proc. AAAI Conf. Artif. Intell. 2021, 35, 3030–3038. [Google Scholar] [CrossRef]
  8. Shi, B.; Wu, Z.; Mo, Z.; Duan, D.; Yeung, S.-K.; Tan, P. A benchmark dataset and evaluation for non-lambertian and uncalibrated photometric stereo. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 3707–3716. [Google Scholar]
  9. Ju, Y.; Lam, K.-M.; Xie, W.; Zhou, H.; Dong, J.; Shi, B. Deep Learning Methods for Calibrated Photometric Stereo and Beyond. IEEE Trans. Pattern Anal. Mach. Intell. 2024; early access. [Google Scholar] [CrossRef]
  10. Chen, G.; Han, K.; Shi, B.; Matsushita, Y.; Wong, K.-Y.K. Deep photometric stereo for non-lambertian surfaces. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 44, 129–142. [Google Scholar] [CrossRef]
  11. Xiang, J.; Wang, Y.; An, L.; Liu, H.; Liu, J. Exploring the mutual influence between self-supervised single-frame and multi-frame depth estimation. IEEE Robot. Autom. Lett. 2023, 8, 6547–6554. [Google Scholar] [CrossRef]
  12. Wang, X.; Zhu, Z.; Huang, G.; Chi, X.; Ye, Y.; Chen, Z.; Wang, X. Crafting monocular cues and velocity guidance for self-supervised multi-frame depth learning. Proc. AAAI Conf. Artif. Intell. 2023, 37, 2689–2697. [Google Scholar] [CrossRef]
  13. Long, Y.; Yu, H.; Liu, B. Two-stream based multi-stage hybrid decoder for self-supervised multi-frame monocular depth. IEEE Robot. Autom. Lett. 2022, 7, 12291–12298. [Google Scholar] [CrossRef]
  14. Uhrig, J.; Schneider, N.; Schneider, L.; Franke, U.; Brox, T.; Geiger, A. Sparsity invariant cnns. In Proceedings of the 2017 International Conference on 3D Vision (3DV), Qingdao, China, 10–12 October 2017; pp. 11–20. [Google Scholar]
  15. Cordts, M.; Omran, M.; Ramos, S.; Rehfeld, T.; Enzweiler, M.; Benenson, R.; Franke, U.; Roth, S.; Schiele, B. The cityscapes dataset for semantic urban scene understanding. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 3213–3223. [Google Scholar]
  16. Silberman, N.; Hoiem, D.; Kohli, P.; Fergus, R. Indoor segmentation and support inference from rgbd images. In Proceedings of Computer Vision–ECCV 2012: 12th European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2012; pp. 746–760. [Google Scholar]
  17. Shotton, J.; Glocker, B.; Zach, C.; Izadi, S.; Criminisi, A.; Fitzgibbon, A. Scene coordinate regression forests for camera relocalization in RGB-D images. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 2013; pp. 2930–2937. [Google Scholar]
  18. Guo, X.; Zhao, H.; Shao, S.; Li, X.; Zhang, B. F2Depth: Self-supervised indoor monocular depth estimation via optical flow consistency and feature map synthesis. Eng. Appl. Artif. Intell. 2024, 133, 108391. [Google Scholar] [CrossRef]
  19. Fu, H.; Gong, M.; Wang, C.; Batmanghelich, K.; Tao, D. Deep ordinal regression network for monocular depth estimation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 2002–2011. [Google Scholar]
  20. Laina, I.; Rupprecht, C.; Belagiannis, V.; Tombari, F.; Navab, N. Deeper depth prediction with fully convolutional residual networks. In Proceedings of the 2016 Fourth International Conference on 3D Vision (3DV), Stanford, CA, USA, 25–28 October 2016; pp. 239–248. [Google Scholar]
  21. Eigen, D.; Fergus, R. Predicting depth, surface normals and semantic labels with a common multi-scale convolutional architecture. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 2650–2658. [Google Scholar]
  22. Li, J.; Klein, R.; Yao, A. A two-streamed network for estimating fine-scaled depth maps from single rgb images. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 3372–3380. [Google Scholar]
  23. Zhang, S.; Yang, L.; Mi, M.B.; Zheng, X.; Yao, A. Improving deep regression with ordinal entropy. arXiv 2023, arXiv:2301.08915. [Google Scholar]
  24. Eigen, D.; Puhrsch, C.; Fergus, R. Depth map prediction from a single image using a multi-scale deep network. Adv. Neural Inf. Process. Syst. 2014, 27, 2366–2374. [Google Scholar]
  25. Liu, F.; Shen, C.; Lin, G.; Reid, I. Learning depth from single monocular images using deep convolutional neural fields. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 38, 2024–2039. [Google Scholar] [CrossRef]
  26. Shao, S.; Pei, Z.; Wu, X.; Liu, Z.; Chen, W.; Li, Z. IEBins: Iterative elastic bins for monocular depth estimation. Adv. Neural Inf. Process. Syst. 2024, 36, 53025–53037. [Google Scholar]
  27. Agarwal, A.; Arora, C. Attention attention everywhere: Monocular depth prediction with skip attention. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 2–7 January 2023; pp. 5861–5870. [Google Scholar]
  28. Agarwal, A.; Arora, C. Depthformer: Multiscale vision transformer for monocular depth estimation with global local information fusion. In Proceedings of the 2022 IEEE International Conference on Image Processing (ICIP), Bordeaux, France, 16–19 October 2022; pp. 3873–3877. [Google Scholar]
  29. Bhat, S.F.; Alhashim, I.; Wonka, P. LocalBins: Improving depth estimation by learning local distributions. In Proceedings of the Computer Vision–ECCV 2022: 17th European Conference, Tel Aviv, Israel, 23–27 October 2022; Proceedings, Part I. pp. 480–496. [Google Scholar]
  30. Bhat, S.F.; Alhashim, I.; Wonka, P. Adabins: Depth estimation using adaptive bins. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 4009–4018. [Google Scholar]
  31. Jun, J.; Lee, J.-H.; Lee, C.; Kim, C.-S. Depth map decomposition for monocular depth estimation. In Proceedings of the Computer Vision–ECCV 2022: 17th European Conference, Tel Aviv, Israel, 23–27 October 2022; Proceedings, Part II. pp. 18–34. [Google Scholar]
  32. Ning, J.; Li, C.; Zhang, Z.; Wang, C.; Geng, Z.; Dai, Q.; He, K.; Hu, H. All in tokens: Unifying output space of visual tasks via soft token. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 1–6 October 2023; pp. 19900–19910. [Google Scholar]
  33. Shao, S.; Pei, Z.; Chen, W.; Li, R.; Liu, Z.; Li, Z. URCDC-Depth: Uncertainty rectified cross-distillation with cutflip for monocular depth estimation. IEEE Trans. Multimed. 2023, 26, 3341–3353. [Google Scholar] [CrossRef]
  34. Piccinelli, L.; Sakaridis, C.; Yu, F. iDisc: Internal discretization for monocular depth estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 21477–21487. [Google Scholar]
  35. Yuan, W.; Gu, X.; Dai, Z.; Zhu, S.; Tan, P. Neural window fully-connected CRFs for monocular depth estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 3916–3925. [Google Scholar]
  36. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30, 5998–6008. [Google Scholar]
  37. Zhao, W.; Rao, Y.; Liu, Z.; Liu, B.; Zhou, J.; Lu, J. Unleashing text-to-image diffusion models for visual perception. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 1–6 October 2023; pp. 5729–5739. [Google Scholar]
  38. Ji, Y.; Chen, Z.; Xie, E.; Hong, L.; Liu, X.; Liu, Z.; Lu, T.; Li, Z.; Luo, P. DDP: Diffusion model for dense visual prediction. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 2–3 October 2023; pp. 21741–21752. [Google Scholar]
  39. Hu, J.; Ozay, M.; Zhang, Y.; Okatani, T. Revisiting single image depth estimation: Toward higher resolution maps with accurate object boundaries. In Proceedings of the 2019 IEEE Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 7–11 January 2019; pp. 1043–1051. [Google Scholar]
  40. Yin, W.; Liu, Y.; Shen, C.; Yan, Y. Enforcing geometric constraints of virtual normal for depth prediction. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 5684–5693. [Google Scholar]
  41. Liu, C.; Yang, J.; Ceylan, D.; Yumer, E.; Furukawa, Y. PlaneNet: Piece-wise planar reconstruction from a single rgb image. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 2579–2588. [Google Scholar]
  42. Yu, Z.; Zheng, J.; Lian, D.; Zhou, Z.; Gao, S. Single-image piece-wise planar 3d reconstruction via associative embedding. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 1029–1037. [Google Scholar]
  43. Patil, V.; Sakaridis, C.; Liniger, A.; Van Gool, L. P3Depth: Monocular depth estimation with a piecewise planarity prior. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 1610–1621. [Google Scholar]
  44. Shao, S.; Pei, Z.; Chen, W.; Wu, X.; Li, Z. NDDepth: Normal-distance assisted monocular depth estimation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 1–6 October 2023; pp. 7931–7940. [Google Scholar]
  45. Li, B.; Huang, Y.; Liu, Z.; Zou, D.; Yu, W. StructDepth: Leveraging the structural regularities for self-supervised indoor depth estimation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 12663–12673. [Google Scholar]
  46. Bian, J.; Li, Z.; Wang, N.; Zhan, H.; Shen, C.; Cheng, M.-M.; Reid, I. Unsupervised scale-consistent depth and ego-motion learning from monocular video. Adv. Neural Inf. Process. Syst. 2019, 32, 1–11. [Google Scholar]
  47. Bian, J.-W.; Zhan, H.; Wang, N.; Chin, T.-J.; Shen, C.; Reid, I. Auto-rectify network for unsupervised indoor depth estimation. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 9802–9813. [Google Scholar] [CrossRef]
  48. Ji, P.; Li, R.; Bhanu, B.; Xu, Y. MonoIndoor: Towards good practice of self-supervised monocular depth estimation for indoor environments. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 12787–12796. [Google Scholar]
  49. Li, R.; Ji, P.; Xu, Y.; Bhanu, B. MonoIndoor++: Towards better practice of self-supervised monocular depth estimation for indoor environments. IEEE Trans. Circuits Syst. Video Technol. 2022, 33, 830–846. [Google Scholar] [CrossRef]
  50. Ranftl, R.; Bochkovskiy, A.; Koltun, V. Vision transformers for dense prediction. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 12179–12188. [Google Scholar]
  51. Ranftl, R.; Lasinger, K.; Hafner, D.; Schindler, K.; Koltun, V. Towards robust monocular depth estimation: Mixing datasets for zero-shot cross-dataset transfer. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 44, 1623–1637. [Google Scholar] [CrossRef]
  52. Wu, C.-Y.; Wang, J.; Hall, M.; Neumann, U.; Su, S. Toward practical monocular indoor depth estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 3814–3824. [Google Scholar]
  53. Zhao, C.; Poggi, M.; Tosi, F.; Zhou, L.; Sun, Q.; Tang, Y.; Mattoccia, S. GasMono: Geometry-aided self-supervised monocular depth estimation for indoor scenes. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 1–6 October 2023; pp. 16209–16220. [Google Scholar]
  54. Schonberger, J.L.; Frahm, J.-M. Structure-from-motion revisited. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 4104–4113. [Google Scholar]
  55. Luo, X.; Huang, J.-B.; Szeliski, R.; Matzen, K.; Kopf, J. Consistent video depth estimation. ACM Trans. Graph. (ToG) 2020, 39, 71. [Google Scholar] [CrossRef]
  56. Patil, V.; Van Gansbeke, W.; Dai, D.; Van Gool, L. Don’t forget the past: Recurrent depth estimation from monocular video. IEEE Robot. Autom. Lett. 2020, 5, 6813–6820. [Google Scholar] [CrossRef]
  57. Yasarla, R.; Cai, H.; Jeong, J.; Shi, Y.; Garrepalli, R.; Porikli, F. MAMo: Leveraging memory and attention for monocular video depth estimation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 1–6 October 2023; pp. 8754–8764. [Google Scholar]
  58. Yang, J.; Alvarez, J.M.; Liu, M. Self-supervised learning of depth inference for multi-view stereo. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 7526–7534. [Google Scholar]
  59. Ding, Y.; Zhu, Q.; Liu, X.; Yuan, W.; Zhang, H.; Zhang, C. KD-MVS: Knowledge distillation based self-supervised learning for multi-view stereo. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; pp. 630–646. [Google Scholar]
  60. Liu, C.; Gu, J.; Kim, K.; Narasimhan, S.G.; Kautz, J. Neural rgb→d sensing: Depth and uncertainty from a video camera. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 10986–10995. [Google Scholar]
  61. Hou, Y.; Kannala, J.; Solin, A. Multi-view stereo by temporal nonparametric fusion. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 2651–2660. [Google Scholar]
  62. Wu, Z.; Wu, X.; Zhang, X.; Wang, S.; Ju, L. Spatial correspondence with generative adversarial network: Learning depth from monocular videos. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 7494–7504. [Google Scholar]
  63. Wimbauer, F.; Yang, N.; Von Stumberg, L.; Zeller, N.; Cremers, D. MonoRec: Semi-supervised dense reconstruction in dynamic environments from a single moving camera. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 6112–6122. [Google Scholar]
  64. Li, R.; Gong, D.; Yin, W.; Chen, H.; Zhu, Y.; Wang, K.; Chen, X.; Sun, J.; Zhang, Y. Learning to fuse monocular and multi-view cues for multi-frame depth estimation in dynamic scenes. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 21539–21548. [Google Scholar]
  65. Guizilini, V.; Ambruș, R.; Chen, D.; Zakharov, S.; Gaidon, A. Multi-frame self-supervised depth with transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 160–170. [Google Scholar]
  66. Zhang, S.; Zhao, C. Dyna-DepthFormer: Multi-frame transformer for self-supervised depth estimation in dynamic scenes. arXiv 2023, arXiv:2301.05871. [Google Scholar]
  67. Zhou, J.; Wang, Y.; Qin, K.; Zeng, W. Moving Indoor: Unsupervised video depth learning in challenging environments. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 8618–8627. [Google Scholar]
  68. Liu, M.; Salzmann, M.; He, X. Discrete-continuous depth estimation from a single image. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 716–723. [Google Scholar]
  69. Li, B.; Shen, C.; Dai, Y.; Van Den Hengel, A.; He, M. Depth and surface normal estimation from monocular images using regression on deep features and hierarchical crfs. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1119–1127. [Google Scholar]
  70. Zhao, W.; Liu, S.; Shu, Y.; Liu, Y.-J. Towards better generalization: Joint depth-pose learning without posenet. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 9151–9161. [Google Scholar]
  71. Zhang, Y.; Gong, M.; Li, J.; Zhang, M.; Jiang, F.; Zhao, H. Self-supervised monocular depth estimation with multiscale perception. IEEE Trans. Image Process. 2022, 31, 3251–3266. [Google Scholar] [CrossRef]
  72. Song, X.; Hu, H.; Liang, L.; Shi, W.; Xie, G.; Lu, X.; Hei, X. Unsupervised monocular estimation of depth and visual odometry using attention and depth-pose consistency loss. IEEE Trans. Multimed. 2023, 26, 3517–3529. [Google Scholar] [CrossRef]
Figure 1. Depth estimation in indoor scenes. The original multi-frame method produces unsatisfactory results in low-textured areas such as doors and walls. The single-frame method handles these regions better. Our SIM-MultiDepth combines the advantages of both methods and performs the best. The absolute relative error is visualized for each method, ranging from blue (Abs Rel = 0) to red (Abs Rel = 0.3). (a) RGB image; (b) depth ground truth; (c) depth predicted by the multi-frame network; (d) error map of the multi-frame depth; (e) depth predicted by the single-frame network; (f) error map of the single-frame depth; (g) depth predicted by our SIM-MultiDepth; (h) error map of our SIM-MultiDepth.
Figure 1. Depth estimation in indoor scenes. The original multi-frame method produces unsatisfactory results in low-textured areas such as doors and walls. The single-frame method handles these regions better. Our SIM-MultiDepth combines the advantages of both methods and performs the best. The absolute relative error is visualized for each method, ranging from blue (Abs Rel = 0) to red (Abs Rel = 0.3). (a) RGB image; (b) depth ground truth; (c) depth predicted by the multi-frame network; (d) error map of the multi-frame depth; (e) depth predicted by the single-frame network; (f) error map of the single-frame depth; (g) depth predicted by our SIM-MultiDepth; (h) error map of our SIM-MultiDepth.
Remotesensing 16 02221 g001aRemotesensing 16 02221 g001b
Figure 2. A pipeline diagram of SIM-MultiDepth. Multiple frames are input to encoders to generate feature maps. With the relative poses computed by the single-frame network, the source frame feature map is warped to build the cost volume. The cost volume and the target frame feature are then decoded into depths. The depths predicted by the single-frame network serve as supervision for the multi-frame depth by computing texture-aware depth consistency loss.
Figure 2. A pipeline diagram of SIM-MultiDepth. Multiple frames are input to encoders to generate feature maps. With the relative poses computed by the single-frame network, the source frame feature map is warped to build the cost volume. The cost volume and the target frame feature are then decoded into depths. The depths predicted by the single-frame network serve as supervision for the multi-frame depth by computing texture-aware depth consistency loss.
Remotesensing 16 02221 g002
Figure 3. Examples of images and extracted patches: (a) RGB image; (b) extracted patches.
Figure 3. Examples of images and extracted patches: (a) RGB image; (b) extracted patches.
Remotesensing 16 02221 g003aRemotesensing 16 02221 g003b
Figure 4. Examples of points in the Campus Indoor dataset. The red stars represent the selected points.
Figure 4. Examples of points in the Campus Indoor dataset. The red stars represent the selected points.
Remotesensing 16 02221 g004
Figure 5. Qualitive depth prediction results for the NYU Depth V2 dataset: (a) RGB image; (b) P2Net [3]; (c) F2Depth [18]; (d) SIM-MultiDepth; (e) ground truth.
Figure 5. Qualitive depth prediction results for the NYU Depth V2 dataset: (a) RGB image; (b) P2Net [3]; (c) F2Depth [18]; (d) SIM-MultiDepth; (e) ground truth.
Remotesensing 16 02221 g005aRemotesensing 16 02221 g005b
Figure 6. Qualitive depth prediction results in a scenewith a transparent glass door: (a) RGB image; (b) SIM-MultiDepth; (c) ground truth.
Figure 6. Qualitive depth prediction results in a scenewith a transparent glass door: (a) RGB image; (b) SIM-MultiDepth; (c) ground truth.
Remotesensing 16 02221 g006
Figure 7. Qualitive depth prediction results with reflective surfaces: (a) RGB image; (b) SIM-MultiDepth; (c) ground truth.
Figure 7. Qualitive depth prediction results with reflective surfaces: (a) RGB image; (b) SIM-MultiDepth; (c) ground truth.
Remotesensing 16 02221 g007
Figure 8. The changes in illumination corresponding to different brightness factor, b , values: (a) b = 0.5 ; (b) b = 0.8 ; (c) b = 1 ; (d) b = 1.2 ; (e) b = 1.5 ; (f) b = 2 .
Figure 8. The changes in illumination corresponding to different brightness factor, b , values: (a) b = 0.5 ; (b) b = 0.8 ; (c) b = 1 ; (d) b = 1.2 ; (e) b = 1.5 ; (f) b = 2 .
Remotesensing 16 02221 g008
Figure 9. Depth visualization results for the 7-Scenes dataset. The first two rows depict the scene Stairs, for which SIM-MultiDepth outperformed F2Depth [18]. The last two rows display thescene Pumpkin, for which F2Depth generalized better than SIM-MultiDepth. (a) RGB images and ground truth; (b) F2Depth generalization results and error maps; (c) SIM-MultiDepth generalization results and error maps. The absolute relative error ranges, in color, from blue (Abs Rel = 0) to red (Abs Rel = 0.2).
Figure 9. Depth visualization results for the 7-Scenes dataset. The first two rows depict the scene Stairs, for which SIM-MultiDepth outperformed F2Depth [18]. The last two rows display thescene Pumpkin, for which F2Depth generalized better than SIM-MultiDepth. (a) RGB images and ground truth; (b) F2Depth generalization results and error maps; (c) SIM-MultiDepth generalization results and error maps. The absolute relative error ranges, in color, from blue (Abs Rel = 0) to red (Abs Rel = 0.2).
Remotesensing 16 02221 g009
Figure 10. Typical examples of the scene Heads in the 7-Scenes dataset.
Figure 10. Typical examples of the scene Heads in the 7-Scenes dataset.
Remotesensing 16 02221 g010
Figure 11. Depth visualization results for scene Nos. 6 and 7 in the Campus Indoor dataset. The first row shows scene No. 6. The second row shows scene No. 7. (a) RGB images; (b) F2Depth [18] generalization results; (c) SIM-MultiDepth generalization results.
Figure 11. Depth visualization results for scene Nos. 6 and 7 in the Campus Indoor dataset. The first row shows scene No. 6. The second row shows scene No. 7. (a) RGB images; (b) F2Depth [18] generalization results; (c) SIM-MultiDepth generalization results.
Remotesensing 16 02221 g011
Figure 12. Depth visualization results of scene No. 5 in the Campus Indoor dataset: (a) RGB image; (b) F2Depth [18] generalization results; (c) SIM-MultiDepth generalization results.
Figure 12. Depth visualization results of scene No. 5 in the Campus Indoor dataset: (a) RGB image; (b) F2Depth [18] generalization results; (c) SIM-MultiDepth generalization results.
Remotesensing 16 02221 g012
Table 1. Quantitative results of SIM-MultiDepth and other existing methods on the NYU Depth V2 dataset. The best results are in bold. ↓ indicates that the lower the value, the better; ↑ indicates that the higher the value, the better.
Table 1. Quantitative results of SIM-MultiDepth and other existing methods on the NYU Depth V2 dataset. The best results are in bold. ↓ indicates that the lower the value, the better; ↑ indicates that the higher the value, the better.
MethodSupervisionREL ↓RMS ↓Log10 ↓δ < 1.25 ↑δ < 1.252δ < 1.253
Liu [68]0.3351.0600.127---
Li [69]0.2320.8210.0940.6210.8860.968
Liu [25]0.2130.7590.0870.6500.9060.976
Eigen [21]0.1580.641-0.7690.9500.988
Li [22]0.1430.6350.0630.7880.9580.991
PlaneNet [41]0.1420.5140.0600.8270.9630.990
PlaneReg [42]0.1340.5030.0570.8270.9630.990
Laina [20]0.1270.5730.0550.8110.9530.988
DORN [19]0.1150.5090.0510.8280.9650.992
VNL [40]0.1080.4160.0480.8750.9760.994
P3Depth [43]0.1040.3560.0430.8980.9810.996
Jun [31]0.1000.3620.0430.9070.9860.997
DDP [38]0.0940.3290.0400.9210.9900.998
Moving Indoor [67]×0.2080.7120.0860.6740.9000.968
TrianFlow [70]×0.1890.6860.0790.7010.9120.978
Zhang [71]×0.1770.634-0.7330.936-
Monodepth2 [2]×0.1700.6170.0720.7480.9420.986
ADPDepth [72]×0.1650.5920.0710.7530.9340.981
SC-Depth [46]×0.1590.6080.0680.7720.9390.982
P2Net [3]×0.1590.5990.0680.7720.9420.984
F2Depth [18]×0.1580.5830.0670.7790.9470.987
Ours×0.1520.5670.0650.7920.9500.988
Table 2. Evaluation results with different brightness factors. The underlined results were predicted by the original test set. ↓ indicates that the lower the value, the better; ↑ indicates that the higher the value, the better.
Table 2. Evaluation results with different brightness factors. The underlined results were predicted by the original test set. ↓ indicates that the lower the value, the better; ↑ indicates that the higher the value, the better.
bREL ↓RMS ↓Log10 ↓δ < 1.25 ↑δ < 1.252δ < 1.253
0.50.1560.5780.0660.7830.9460.987
0.80.1520.5670.0650.7920.9500.988
10.1520.5670.0650.7920.9500.988
1.20.1520.5680.0650.7910.9490.988
1.50.1530.5710.0650.7880.9480.988
20.1570.5810.0670.7800.9460.987
Table 3. Comparison results of our SIM-MultiDepth and F2Depth [18] on 7-Scenes. The best results are in bold. ↓ represents the lower the better, ↑ represents the higher the better.
Table 3. Comparison results of our SIM-MultiDepth and F2Depth [18] on 7-Scenes. The best results are in bold. ↓ represents the lower the better, ↑ represents the higher the better.
MethodsOur SIM-MultiDepthF2Depth [18]
SceneREL ↓RMS ↓Log10 ↓δ < 1.25 ↑δ < 1.252δ < 1.253REL ↓RMS ↓Log10 ↓δ < 1.25 ↑δ < 1.252δ < 1.253
Chess0.1860.4110.0810.6680.9400.9920.1860.4090.0810.6710.9360.993
Fire0.1800.3320.0780.6900.9500.9910.1760.3220.0760.7010.9510.991
Heads0.1960.2040.0820.6740.9240.9850.1850.1950.0780.7180.9300.983
Office0.1590.3630.0660.7660.9700.9960.1620.3700.0670.7620.9630.996
Pumpkin0.1360.3720.0590.8130.9780.9960.1270.3500.0560.8460.9800.995
RedKitchen0.1710.4140.0730.7240.9500.9940.1730.4160.0740.7220.9460.992
Stairs0.1470.4370.0640.7840.9220.9740.1590.4550.0680.7660.9120.972
Average0.1670.3760.0710.7340.9540.9920.1670.3750.0710.7400.9500.992
Table 4. Comparison of the results of SIM-MultiDepth and F2Depth [18] on the Campus Indoor dataset. The best results are in bold. ↓ indicates that the lower the value, the better; ↑ indicates that the higher the value, the better.
Table 4. Comparison of the results of SIM-MultiDepth and F2Depth [18] on the Campus Indoor dataset. The best results are in bold. ↓ indicates that the lower the value, the better; ↑ indicates that the higher the value, the better.
MethodsOur SIM-MultiDepthF2Depth [18]
Scene No.REL ↓RMS ↓Log10 ↓δ < 1.25 ↑δ < 1.252δ < 1.253REL ↓RMS ↓Log10 ↓δ < 1.25 ↑δ < 1.252δ < 1.253
10.1560.5900.0630.7500.97710.164 0.592 0.064 0.7660.947 1
20.1441.0830.0660.752110.1350.9490.0600.7980.981 1
30.2180.8870.0830.6850.94610.2150.889 0.084 0.644 0.9731
40.1420.5590.0600.818110.1340.588 0.0580.777 11
50.1831.2240.0780.6670.97310.1591.0910.0670.7460.946 1
60.1361.4870.0660.7740.9330.9720.154 1.625 0.073 0.757 0.919 0.932
70.2442.2240.1090.5140.8300.9730.266 2.321 0.122 0.436 0.802 0.960
80.1720.7910.0710.7450.91510.1490.6990.0630.7720.9291
90.0990.4180.0410.907110.0880.3720.0370.94611
100.2320.7610.0910.5830.93310.1930.6860.0760.7170.917 1
110.1830.6740.0760.7330.9730.9870.1680.6670.0720.7870.9731
120.1240.5120.0580.763110.143 0.536 0.066 0.736 11
130.1440.5940.0670.7900.97410.147 0.5940.0670.8030.9741
140.1410.4430.0620.8270.98710.1350.4220.0600.88011
150.2560.4330.0870.7470.8130.9470.2360.4090.0840.720 0.8670.946
160.1630.5270.0670.7870.96010.202 0.647 0.079 0.707 0.920 1
170.1270.2700.0500.9070.9470.9870.1140.2440.0470.893 0.9601
180.1440.2270.0630.8000.98710.1440.230 0.0610.787 0.9871
Average0.1650.7690.0690.7540.9560.9930.1640.7530.0690.7600.950 0.991
Table 5. Ablation studies of our masking manner on the NYU Depth V2 dataset. The best results are in bold. ↓ indicates that the lower the value, the better; ↑ indicates that the higher the value, the better.
Table 5. Ablation studies of our masking manner on the NYU Depth V2 dataset. The best results are in bold. ↓ indicates that the lower the value, the better; ↑ indicates that the higher the value, the better.
Final Mask M REL ↓RMS ↓Log10 ↓δ < 1.25 ↑δ < 1.252δ < 1.253
No consistency loss0.1580.5910.0670.7790.9430.985
M 1 0.1530.5700.0650.7910.9490.987
M 2 0.1520.5700.0650.7900.9490.987
M 1 M 2 0.1520.5670.0650.7920.9500.988
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Guo, X.; Zhao, H.; Shao, S.; Li, X.; Zhang, B.; Li, N. SIM-MultiDepth: Self-Supervised Indoor Monocular Multi-Frame Depth Estimation Based on Texture-Aware Masking. Remote Sens. 2024, 16, 2221. https://doi.org/10.3390/rs16122221

AMA Style

Guo X, Zhao H, Shao S, Li X, Zhang B, Li N. SIM-MultiDepth: Self-Supervised Indoor Monocular Multi-Frame Depth Estimation Based on Texture-Aware Masking. Remote Sensing. 2024; 16(12):2221. https://doi.org/10.3390/rs16122221

Chicago/Turabian Style

Guo, Xiaotong, Huijie Zhao, Shuwei Shao, Xudong Li, Baochang Zhang, and Na Li. 2024. "SIM-MultiDepth: Self-Supervised Indoor Monocular Multi-Frame Depth Estimation Based on Texture-Aware Masking" Remote Sensing 16, no. 12: 2221. https://doi.org/10.3390/rs16122221

APA Style

Guo, X., Zhao, H., Shao, S., Li, X., Zhang, B., & Li, N. (2024). SIM-MultiDepth: Self-Supervised Indoor Monocular Multi-Frame Depth Estimation Based on Texture-Aware Masking. Remote Sensing, 16(12), 2221. https://doi.org/10.3390/rs16122221

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop