Next Article in Journal
Integration of Inertial Sensors in a Lower Limb Robotic Exoskeleton
Next Article in Special Issue
Fair Facial Attribute Classification via Causal Graph-Based Attribute Translation
Previous Article in Journal
Remaining Useful Life Prediction Method for Bearings Based on LSTM with Uncertainty Quantification
Previous Article in Special Issue
A Review of Image Processing Techniques for Deepfakes
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Adaptive Aggregate Stereo Matching Network with Depth Map Super-Resolution

1
School of Computer Science, Yangtze University, Jingzhou 434023, China
2
Department of Creative Technologies and Product Design, National Taipei University of Business, Taipei 10051, Taiwan
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(12), 4548; https://doi.org/10.3390/s22124548
Submission received: 18 May 2022 / Revised: 13 June 2022 / Accepted: 14 June 2022 / Published: 16 June 2022
(This article belongs to the Special Issue Artificial Intelligence in Computer Vision: Methods and Applications)

Abstract

:
In order to avoid the direct depth reconstruction of the original image pair and improve the accuracy of the results, we proposed a coarse-to-fine stereo matching network combining multi-level residual optimization and depth map super-resolution (ASR-Net). First, we used the u-net feature extractor to obtain the multi-scale feature pair. Second, we reconstructed global disparity in the lowest resolution. Then, we regressed the residual disparity using the higher-resolution feature pair. Finally, the lowest-resolution depth map was refined by using the disparity residual. In addition, we introduced deformable convolution and group-wise cost volume into the network to achieve adaptive cost aggregation. Further, the network uses ABPN instead of the traditional interpolation method. The network was evaluated on three datasets: scene flow, kitti2015, and kitti2012 and the experimental results showed that the speed and accuracy of our method were excellent. On the kitti2015 dataset, the three-pixel error converged to 2.86%, and the speed was about six times and two times that of GC-net and GWC-net.

1. Introduction

Visual depth estimation has its economic advantages over the lidar method. Therefore, a vision-based depth estimation algorithm has always been a popular research field. At present, the stereo matching algorithm is widely used in AR [1], MR [2], autopilot [3], robot navigation [4], and other fields.
Human acquisition of external image information is mainly based on the receptive field mechanism. The brain can infer and estimate the depth by understanding the image features captured by both eyes. Stereo matching technology lets the computer simulate this process. The main problem solved by this technology is how to find the matching pixels in the left and right view, calculate the disparity between the matching points, and finally obtain the depth map.
For example, in binocular image pairs corrected by epipolar line: we define x and y as coordinates along the width and height of the image, d as disparity, and D as depth. According to the principle of a binocular camera, if there is a pixel point (x, y) in the left image, there must be a point (x − d, y) in the right image to match the left. When we obtain the disparity, we can obtain the actual depth (D = (fb/d)) of the pixel point in the real world, where f is the focal length and b is the baseline distance of the binocular camera. However, in the application, it is vital to estimate the depth of the 3D world quickly and accurately.
In recent years, traditional stereo matching methods have been unable to effectively deal with morbid image regions such as strong reflection and low texture. In order to avoid wrong matches, an increasing number of scholars apply deep learning to stereo matching. Due to the complexity of the stereo matching algorithm, how to design a network that takes into account the stability of speed, accuracy, and robustness is our main research problem.
This paper proposes an adaptive aggregate stereo matching network with depth map super-resolution (ASR-net) based on the idea of coarse-to-fine. We adopt some ideas in Any-net [5] and GC-net [6]. However, although Any-net is fast, its accuracy is inferior. GC-net applies 3D convolution and it has received good accuracy, but the amount of calculation is too large. Our method combines their advantages and achieves a better disparity estimation, as shown in Figure 1. According to the experimental conclusion, the advantages of our network are as follows:
(1)
Through a u-net multi-scale feature extractor, we only reconstructed the global depth map in the lowest resolution once. Then we refined the depth map step by step. Finally, the original depth map is restored. In this way, the calculation cost was significantly reduced.
(2)
Deformable convolution and Group-wise cost volume construction were introduced into our network to thoroughly learn the global depth information of feature image pairs.
(3)
Using the depth map super-resolution network ABPN [7] to replace the traditional interpolation algorithm significantly reduces the information loss in the up-sampling process.

2. Realated Work

The traditional representative algorithm of binocular stereo matching is SGM(Semi-Global Matching) [8,9], which obtains the disparity map through four steps: cost calculation, cost aggregation, disparity calculation, and disparity optimization [10]. Inspired by the SGM algorithm, Mayer et al. first proposed the end-to-end stereo matching network Disp-net [11], which integrates the four steps of the traditional algorithm into the neural network, and directly obtains the disparity map by inputting left–right image pairs. Kendall et al. Proposed GC-net based on Disp-net in 2017. This method introduces the concept of cost volume, constructs cost volume by splicing left and right feature images, simulates the cost aggregation of traditional methods, uses the 3D convolution method to optimize disparity [12], and finally obtains the normalized disparity map. The emergence of the GC-net laid the basic framework of a stereo matching network. Later, scholars proposed PSM-Net based on pyramid pooling [13] and GWC-net based on Group-Wise correlation [14]. In the last two years, with the rise of the attention module [15], the channel attention mechanism has been gradually introduced into the stereo-matching network. By using the channel attention module and spatial attention module to optimize the features independently, the feature information of the image is fully extracted [16,17,18]. These methods exchange better computation for a height–precision disparity map, resulting in many redundant calculations.
In order to reduce the computational cost, scholars refine the network differently. Zhang Peng et al. proposed a method based on semantic edge driving. They constructed a lightweight feature extraction module to reduce the redundancy of the feature extraction module to simplify the feature extraction steps better and improve the real-time and lightweight nature of the network [19]. In order to replace the complex 3D convolution calculation, Xu et al. proposed AA-net [20]. This method integrates multi-scale feature extraction, adaptive inter-layer aggregation (ISA), and cross-layer cost aggregation (CSA), completely replaces the 3D convolution in the current SOTA(state-of-the-art) model, and dramatically improves the reasoning speed. Khamis et al. proposed stereo-net [21] in 2018. This method is based on the idea of coarse-to-fine for the first time, using lowest-resolution images to predict depth maps and then using datasets for guided filtering to improve the accuracy of low-resolution disparity maps step by step. Based on Khamis’s research, Wang et al. proposed Any-net [5] to solve the problem that the computing power of the mobile terminal is too low to adapt to the stereo matching network. Based on stereo-net, this method uses residual optimization to restore the original resolution depth map from one-sixteenth of the input image resolution; the network saves the optimization results of each scale. The lower the accuracy of the depth map, the faster the generation speed. Finally, the depth map with different accuracy can be output according to the actual requirements. Last year, the image super-resolution task [22,23] also made a further breakthrough in depth estimation. Song et al. proposed a framework based on iterative residual learning [24]. In this method, channel attention, multi-stage fusion, weight sharing, and depth thinning were used to learn and generate a height-resolution depth map.
The above methods generally use the end-to-end network to obtain the exact size depth map output from the original size input and apply many 3D convolutions in the network layer, resulting in too much computation. Some lightweight networks usually use interpolation to up-sample low-resolution inputs [25,26,27], which causes the loss of depth information in resolution restoration. We combined the super-resolution task of depth map with a stereo matching network to solve the problem of depth information loss caused by traditional interpolation in the up-sampling part. Because the whole network only carries out global depth reconstruction for low-resolution feature image pairs, the amount of redundant calculation is significantly reduced and decreases the use of 3D convolution. Compared with other networks, our method considered both speed and accuracy and dramatically improved the generalization performance of the network.

3. Network Architecture

Our work focused on depth map refinement. The structure of the network refers to Any-net [5]. A lightweight u-net network is used to obtain the feature image pairs with different scales in the preliminary feature extraction process. Then, the lowest resolution features are input into the GWD-module(Group-wise cost volume with deformable convolution) for density disparity reconstruction. Then we use ABPN [7] to up-sample the depth map output from each layer. Finally, each layer’s residual module will filter the lower resolution depth map to obtain the original size depth map.

3.1. Features Extraction

In order to realize the idea from coarse to fine, we used a lightweight u-net network [28,29] as the feature extractor (see Figure 2). The network structure is shown in Figure 3:
The feature extractor took left and right camera images as the input. In the process of the first layer, an 8-layer res-block [30] with dilated convolution [31] was used to extract edge features from binocular image pair. The subsequent layer used maximum pooling and stride convolution for down sampling and then a 2D convolution filter for feature extraction. Bilinear interpolation was used for up-sampling, and skip-connection was used between layers to prevent the loss of feature information. Finally, three scale feature image pairs of 1/4, 1/8, and 1/16 were obtained as the input parameters of the subsequent optimization process.

3.2. GWD-Module

The global depth reconstruction of a low-resolution feature map is the key to the whole network. In this part, we propose a GWD-module (Group-wise cost volume with deformable convolution) (see Figure 4), which includes the construction of group-wise cost volume and deformable convolution to realize adaptive cost aggregation. Finally, we used the multi-scale 3D convolution structure 3D u-net [32,33] to regularize the cost volume. Through this method, the network parameters were increased, and a better low-resolution disparity map was obtained.

3.2.1. Group-Wise Correlation Cost Volume

In our research, the depth results of different cost volume construction methods have their advantages:
The concatenate method obtains the correlation between left and right images through parameter learning, while the inner product method and subtraction method estimate the correlation through artificially set measurement methods. As shown in Figure 5, the concatenate method contains more local details. The subtraction and inner product contain more global features. In order to build a cost volume that can learn more information, we combined the concatenate method and the inner product method. However, in the traditional inner product correlation calculation, we calculated and averaged all feature image pairs along the channel, which resulted in the loss of feature continuity in time and space. d means disparity level, x , y , means the width and height of feature. We denoted left unary features and right feature as   f l , f r , the number of channel is   N . . , . represents the inner product of two vectors, the correlation C s was obtained by k-th feature pair f l k , f r k for each disparity level d , as shown in formula (1):
C s ( d , x , y ) = 1 N f l k ( x , y ) , f r k ( x d , y )
In order to deal with this problem, a group-wise correlation construction method is proposed concerning GWC-net [14], which combines the advantages of concatenation and inner product methods. Firstly, we recorded the number of feature graph channels input in the feature extractor as N c , and divide the channels into groups N k along the channel dimension. Each group contains N c N k channels. The corresponding k-th feature group f l k , f r k consists of the   k N c N k , k N c N k + 1 , ... , k N c N k + ( N c N k 1 ) th channels of the original feature. The correlation calculation formula is as follows:
C s ( d , x , y ) = 1 N f l k ( x , y ) , f r k ( x d , y )
In this part, the disparity candidate space is D m a x . Like the input feature map, disparity candidate space must be changed to the 1/16 of the original. Finally, the shape of the cost volume is [ D m a x / 16 , H / 16 , W / 16 , N k ] (see Figure 6).

3.2.2. Adaptive Cost Aggregation Based on Deformable Convolution

The disparity search window is a regular rectangle in the traditional cost aggregation process. This kind of search method will cause the fusion of foreground and background information and then generate severe depth loss. In order to solve this problem, an adaptive cost aggregation based on deformation convolution is proposed. For a search window, we discerned the additional offset which consists of two bias x , y along the width and height directions. As shown in Figure 4, this method has been used in depth map regression of AA-Net [20] and PatchMatch-Net [34] and achieves a better effect. The implementation method is shown in the formula:
C o s t ( d , p ) = I 2 w i C o s t ( d , p + p i + ( x + y ) )
C o s t ( d , p ) denotes the aggregated cost at pixel p for disparity candidata d . I is convolution kernel size, I 2 is the number of kernel‘s element (in our paper I = 3). w i is the aggregation weight for I t h kernel‘s points. p i is the fixed offset of the standard convolution method. x , y is additional offset. The sampling points can be qualitatively concentrated in the target area to achieve adaptive aggregation by learning and fusing these two offsets.
This step constructs three-layer convolutions to process the input cost volume. The three-layer convolutions are 1 × 1, 3 × 3, and 1 × 1. Two 1 × 1 convolutions were used to adjust the number of channels. The 3 × 3 convolution adopts the deformation convolution method and dilated convolution mode. By increasing the receptive field of the search box, the model can learn more offset information (in this paper, the expansion rate was 3).

3.2.3. 3D-Unet Cost Volume Regularization

In previous work, we solved the problem of preliminary construction of cost-volume and adaptive cost aggregation. However, some noise will still be in such cost volume (such as features extracted from non-textured image areas). Although this method takes 1/16 as the minimum resolution, the lower resolution will also contain some global semantic information. Here, we constructed a multi-scale 3D-CNN model for cost volume regularization. The encoding and decoding structure is similar to u-net, which is called 3D U-net [32,33] in this paper.
Each layer uses 3D convolution, batch normalization 3D, and a Relu activation function to process the cost volume. The down-sampling process was realized by setting the convolution step size = 2, and the up-sampling uses the same 3D deconvolution layer. Skip connection was adopted between the same scale to share weight information. This regularization method makes the cost volume contain more global information. Finally, we used the softmax function to convert the correlation into probability along with the disparity candidate space of the cost volume to facilitate the subsequent error propagation.

3.3. Residual-Module

We used the residual module to optimize the low-resolution depth map. In order to simplify the operation, we did not carry out global reconstruction like the first layer. Firstly, we used ABPN [7] to up-sample the low-resolution depth map to make it consistent with the size of the feature map of the upper layer. Then we saved the left view generated by the correct feature map under the supervision of the low-resolution disparity map through the grid sample method in PyTorch and input it into the residual module together with the original left feature map. Finally, we obtained the residual. This item represents the error generated by the disparity map. The implementation process is shown in the following Figure 7.
Assuming there is P l ( x , y ) in left feature, the right feature can obtain match point P r ( x + d , y ) under the supervision of the disparity map. If there is no error normalization in the disparity map, these two points should be identical. Nevertheless, there must have been some wrong match in the disparity map. In order to construct the cost volume, the left feature with error is subtracted from the left feature obtained by this layer to obtain the difference. In the process of global reconstruction, our disparity candidate was 0–192, which was specified by the Kitti dataset. In the step-by-step optimization of disparity by the residual module, we set the disparity consideration as (−3,3), a total of seven possible candidates. The new disparity consideration represents the range space of error in the generated residual. In this way, the constructed cost volume is reduced from D = 192 to d = 7. It dramatically improves operational efficiency.
In this part, we still used deformable convolution. In the lowest-resolution feature of 3.2, deformable convolution is mainly used to distinguish the front and back scenes of the image. In the process of residual optimization, with the improvement of resolution, ill-conditioned graphic areas (low-textured, specular, reflective regions) will gradually appear. According to the empirical analysis of traditional cost aggregation, the error will focus on these areas. At this time, the residual cost volume constructed by deformation convolution can adaptively aggregate the ill-conditioned region error to improve the accuracy of the final result.

3.4. Super-Resolution Network

In the residual optimization network, the output of each scale needs to be up-sampled before error regression with a height-resolution feature map. Traditional up-sampling methods, such as bicubic interpolation, will inevitably cause information loss for depth maps with dense information. In order to optimize the original disparity map and eliminate the error caused by the traditional sampling method as much as possible, in this paper, the ABPN model was used to replace the traditional sampling method to improve the resolution of the depth map. SAB (Spatial Attention Blocks) and RBPB (Refined Back Projection Block) is proposed in this model. In the super-resolution task of the depth map, the restoration accuracy of the ABPN is much higher than that of traditional methods such as bicubic interpolation.

3.5. Loss Function

Our training model adopted smooth Huber loss (L1 loss) in the residual module and MDS module. In the disparity regression task, compared with L2 loss, in the disparity discontinuous region, the loss is proved to be robust, and L1 can significantly suppress the influence of noise. The formula is as follows:
s m o o t h L 1 ( x ) = { 0.5 x 2 , i f   | x | < 1 | x | 0.5 , o t h e r w i s e
The loss calculation for each pixel is as follows:
L ( d , d ^ ) = 1 N i = 1 N s m o o t h L 1 ( d i d i ^ )
N represents the total number of marked pixels, d i is the actual disparity value of the GT (ground-truth) image, d i ^ is the predicted disparity value. This method corresponds to four output losses on four layers. According to the parameter quantity of each layer, we set the weight of loss for each layer as: 1 , 2 , 3 ,   4 = 0.25, 0.5, 1, 1 respectively. The final loss function combines the prediction results of all layers, L i means the value of loss. The formula is as follows:
L = i = 1 l a y e r i L i

4. Experimental Evaluation

The generation result of the super-resolution network ABPN will affect the generation efficiency of the primary network. Therefore, this model divides the training into two parts, and the training process is shown in Figure 8 below:

4.1. Dataset Preprocessing and Setup

We used PyTroch + Ubuntu to complete our two training, both using Adam (β1 = 0.9, β2 = 0.999) as our optimizer. Our experimental configuration is:
Memory: 32 GB,
CPU: AMD Ryzen9 5900X 12-core [email protected],
GPU: NVIDIA RTX3090 (24 GB).
(1) scene flow [11], (2) Kitti 2012 [35], (3) Kitti 2015 [35], (4) Middlebury 2014 [36] were used in our experiment. We extracted subsets from the four datasets for ABPN expansion data. Among them, (1), (2), and (3) were used in the training of the main network. The dataset is described as follows:
Scene Flow [11]: A large-scale synthetic dataset containing three sub-sets, FlyingThings3D, Monkaa, and Driving; contains daily supplies flying along random 3D paths, animated short films, and vehicular driving images similar to the KITTI dataset. Datasets provide a complete disparity image as ground truth. There are 35,454 training images and 4370 test images in the dataset, H = 540 and W = 960.
KITTI2012 [35]: This dataset was obtained by street photography of the Kitti data acquisition platform, which includes two gray cameras, two RGB cameras, a lidar, four optical lenses, and a GPS navigation system. The data includes vehicles, road signs, and other road scenes. It contains 194 pairs of stereo images with ground truth for training and 195 pairs of stereo images without ground truth for testing, H = 376, and W = 1240.
KITTI2015 [35]: KITTI2015 is similar to KITTI2012. The dataset contains 200 pairs of stereo images with ground truth that can be used for training and 200 pairs of stereo images without ground truth for testing, H = 376, and W = 1240.
Middlebury 2014 [36]: These 33 datasets were created by Nera Nesic, Porter Westling, Xi Wang, York Kitajima, Greg Krathwohl, and Daniel Scharstein at Middlebury College during 2011 to 2013. Each dataset consists of two views taken under several different illuminations and exposures. The full size of the image is H = 1988, width = 2964.

4.2. ABPN Training

ABPN network provides a pre-training model. However, it is necessary to use depth map datasets to fine-tune the pre-training model. We selected 200 GT depth maps from each of the three subsets of the scene flow dataset and combined all GT depth maps of kitti2015, kitti2012, and Middlebury to form a training set with a size of 1000. We first converted the PFM file in the obtained dataset into PNG format through the script. Then we constructed low-resolution datasets according to the down-sampling strategy of DVS [37] and AIR [24]. Datasets were divided into various resolutions (1/4, 1/8, 1/16), corresponding to the scale of the primary network.
The learning rate interval was set from 0.01 to 0.0001. According to the attenuation rate of 0.1, it gradually decreases with training progress (every 50 epochs), and the model converges to 300 epochs. The training time of three scales on RTX 3090 GPU is 7 h, 10 h, and 14 h. After the training, we randomly selected 100 samples from four datasets in 4.1 as the test set. The comparison results between the output results of ABPN and the traditional bicubic interpolation are as follows:
As shown in Figure 9 and Table 1, the recovery effect of ABPN on multi-scale was more accurate than that of the traditional methods. Using ABPN instead of the traditional interpolation method to carry out the up-sampling step in the primary network can further avoid the loss of depth information.

4.3. Main Network Training

After the training of the ABPN network, we saved and integrated the model into callable functions. In the pre-training process of the primary network, we used all the training sets of Sceneflow (35,454) and verified them on all the training sets (4370). The original data were cut randomly before training. We set the batch size as the limit value of 8, the learning rate starts at 0.001, and 30 epochs were trained. After 20 epochs, the online learning rate decreased to 0.0001. In the fine-tuning process, the mixed datasets of kitti2012 and kitti2015 were used to train 800 cycles. After 400 epochs learning rate was initially set to 0.001, and the learning rate decreased to half of the original every 100 epochs. For all datasets, the input images were normalized with ImageNet mean and standard deviation statistics. The network pre-training time is 23 h, and the fine-tune training time is 9 h.

4.4. Ablation Experiments

In order to verify the effectiveness of various modules proposed in this method, we tested the model through ablation experiments using Sceneflow and kitti2015 validation sets. The evaluation index of this section adopts pixel error as the experimental evaluation index. We set a threshold to judge depth estimation accuracy when the absolute difference between the output and the actual disparity is bigger than the threshold. This pixel is considered the wrong pixel. Pixel error is the percentage of these error pixels in all pixels. EPE represents the mean value of all pixel errors between the predicted image and the real image. The three pixel error is the final test result, which is composed of D1-bg (background error evaluation index), D1-fg (foreground error evaluation index), and D1-all (global evaluation index). The lower their value, the better the effect.

4.4.1. Group-Wise Cost Volume

The research in Figure 5 shows that different cost volume construction methods can have a significant impact on model parameters and learning results in the stereo-net model, and different construction methods have their unique advantages. We believe that this law is also applicable to this method. In this section, ablation experiments are conducted on the cost volume construction method in the GWD-module.
We kept the other modules of the network unchanged. Focusing on the method of cost volume construction, we set two controls: Concatenate (left + right), the number of channels is 2C, and subtract (left − right) channels is C. Our construction method combines concatenate (left + right) and inner product <left, right>, and the number of channels is 3C. Analyze the original scale output image of the model. The evaluation index adopts EPE and three-pixel error, represents the whole image matching error, and NOC represents the unobstructed area error. It can be seen from Table 2 that the subcontract method was the fastest to build the runtime, while the concatenate method was slower than the others, but the accuracy is improved. Our method had the slowest speed, but a small amount of speed sacrifice resulted in a massive improvement in accuracy. The whole region error was about 0.28% lower than that of the subtract method, and in the whole region of the three-pixel error, the error was 0.51% lower than the subtract method. Figure 10 proved that our method fully combines the feature information of different dimensions and improves the model’s accuracy more effectively than the simple splicing of concatenation.

4.4.2. Deformable Convolution

In order to explore the improvement of deformable convolution on the processing effect of image ill-conditioned areas, this section compares the model output results without deformable convolution. In Table 3, “w/o” means that deformation convolution is not used, MDS means MDS module, and Res means residual module. The results showed that on the scene flow dataset, EPE error decreased by 0.32% and 1-pixel error decreased by 3.6%. In the kitti dataset, three-pixel error D1-all decreased by 0.25%.
In Figure 11A, the depth of the occluded object is restored. In the non-textured region, the depth information is smoother than that of Figure 11B. Figure 11 shows that after deformation convolution is adopted, the problems of blurring of occluded areas after reconstruction and edge loss of non-textured areas were effectively solved.

4.4.3. SR-Network

In Section 4.2, the superiority of ABPN over traditional methods for sampling on depth maps was proved. In order to further prove the significant improvement of this method by introducing ABPN, the network trained by bicubic interpolation was compared with the network trained by ABPN. It can be seen from Figure 12 that in the traditional method, the most severe information loss in the sampling process is in the rod-shaped and sharp edge areas, such as the electric pole notice board. All have recovered well.

4.5. Results of Experiments

Finally, in order to prove the advantages of our method, it was compared with other robust stereo matching networks on two datasets:
Sceneflow: Because the dataset has a large amount of data and the accuracy of GT data is very high, it was mainly compared in terms of EPE and run-time. It can be seen from Table 4 that this method takes into account both accuracy and speed, and the error reached one-half of that of the classical stereo matching network GC-net and six times the speed.
Kitti2012, Kitti2015: For these two datasets, we used different evaluation indicators for comparison. Table 5 shows that in terms of a global error on the kitti2012 dataset, although this method was 0.28% and 0.47% higher than PSM-net and GWC-net, its speed was 2–2.5 times higher than theirs. On kitti2015 dataset, the regional error is only higher than that of PSM-net and GWC-net. It can be seen that this method takes into account both accuracy and speed, which shows that the combination of coarse-to-fine training method and complex network layer can achieve better results.

5. Conclusions

A real-time height-precision stereo matching network using deformable convolution and super-resolution network was proposed in our research. In this work, the problem of information adhesion caused by regular convolution in cost aggregation was avoided, and the problem of non-texture and occlusion area matching in traditional stereo matching methods was solved. At the same time, to deal with the issue of a large amount of 3D convolution calculation and many redundant parameters, this method avoids global disparity reconstruction at height resolution but in a low-resolution feature map, which significantly reduces the amount of calculation while ensuring the accuracy. This paper combined the advantages of different cost volume construction methods and adopted GWC cost volume further to improve the global semantic information of model learning. Finally, we used the super-resolution network ABPN to replace the traditional interpolation up-sampling method to reduce the information loss in primary network optimization. Compared with other networks, our experiments prove their superiority of them. Finally, we believe that there is still promotion space in the depth image super-resolution task. In future work, we will consider further improving the up-sampling process to improve the performance of our method.

Author Contributions

Investigation, K.C.; resources, S.-L.P.; supervision, M.Z. and B.L.; funding acquisition, M.Z.; writing—original draft preparation, K.C.; writing-review and editing, M.Z., B.L., S.-L.P. and K.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Hubei Provincial Department of Education: 21D031.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Chatzopoulos, D.; Bermejo, C.; Huang, Z.; Hui, P. Mobile augmented reality survey: From where we are to where we go. IEEE Access 2017, 5, 6917–6950. [Google Scholar] [CrossRef]
  2. Costanza, E.; Kunz, A.; Fjeld, M. Mixed reality: A survey. In Human Machine Interaction; Springer: Berlin/Heidelberg, Germany, 2009; pp. 47–68. [Google Scholar]
  3. Rasouli, A.; Tsotsos, J.K. Autonomous vehicles that interact with pedestrians: A survey of theory and practice. IEEE Trans. Intell. Transp. Syst. 2019, 21, 900–918. [Google Scholar] [CrossRef] [Green Version]
  4. Desouza, G.N.; Kak, A.C. Vision for mobile robot navigation: A survey. IEEE Trans. Pattern Anal. Mach. Intell. 2002, 24, 237–267. [Google Scholar] [CrossRef] [Green Version]
  5. Wang, Y.; Lai, Z.; Huang, G.; Wang, B.H.; Van Der Maaten, L.; Campbell, M.; Weinberger, K.Q. Anytime stereo image depth estimation on mobile devices. In Proceedings of the 2019 International Conference on Robotics and Automation (ICRA), Montreal, QC, Canada, 20–24 May 2019; pp. 5893–5900. [Google Scholar]
  6. Kendall, A.; Martirosyan, H.; Dasgupta, S.; Henry, P.; Kennedy, R.; Bachrach, A.; Bry, A. End-to-end learning of geometry and context for deep stereo regression. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 66–75. [Google Scholar]
  7. Liu, Z.S.; Wang, L.W.; Li, C.T.; Siu, W.C.; Chan, Y.L. Image super-resolution via attention based back projection networks. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW), Seoul, Korea, 27–28 October 2019; pp. 3517–3525. [Google Scholar]
  8. Chen, Y.; Yang, L.-L.; Wang, Z.-P. Literature survey on stereo vision matching algorithms. J. Graph. 2020, 41, 702. [Google Scholar] [CrossRef]
  9. Hirschmuller, H. Accurate and efficient stereo processing by semi-global matching and mutual information. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), San Diego, CA, USA, 20–25 June 2005; Volume 2, pp. 807–814. [Google Scholar]
  10. Bleyer, M.; Rhemann, C.; Rother, C. Patchmatch stereo-stereo matching with slanted support windows. Bmvc 2011, 11, 1–11. [Google Scholar]
  11. Mayer, N.; Ilg, E.; Hausser, P.; Fischer, P.; Cremers, D.; Dosovitskiy, A.; Brox, T. A large dataset to train convolutional networks for disparity, optical flow, and scene flow estimation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 4040–4048. [Google Scholar]
  12. Luo, K.; Guan, T.; Ju, L.; Huang, H.; Luo, Y. P-mvsnet: Learning patch-wise matching confidence aggregation for multi-view stereo. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 10452–10461. [Google Scholar]
  13. Chang, J.R.; Chen, Y.S. Pyramid stereo matching network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 5410–5418. [Google Scholar]
  14. Guo, X.; Yang, K.; Yang, W.; Wang, X.; Li, H. Group-wise correlation stereo network. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 3273–3282. [Google Scholar]
  15. Bahdanau, D.; Cho, K.; Bengio, Y. Neural machine translation by jointly learning to align and translate. arXiv 2014, arXiv:1409.0473. [Google Scholar]
  16. Zhang, Y.; Li, K.; Li, K.; Zhong, B.; Fu, Y. Residual non-local attention networks for image restoration. arXiv 2019, arXiv:1903.10082. [Google Scholar]
  17. Zhang, X.; Wang, H.; Bian, J. Stereo Matching Network with Multi-Cost Fusion. Comput. Eng. 2022, 48, 186–193. [Google Scholar]
  18. Sang, H.-W.; Xu, H.; Xiong, W.-C.; Zuo, Y.; Zhao, Y. Pixel attention based siamese convolutionneural network for stereo matching. Comput. Eng. Sci. 2020, 42, 877–883. [Google Scholar]
  19. Zhang, P.; Wang, X.-Q.; Xiao, Y.; Duan, B.-G.; Xu, H.-H. Real-time Binocular Depth Estimation Algorithm Based on Semantic Edge Drive. Comput. Sci. 2021, 48, 216–222. [Google Scholar]
  20. Xu, H.; Zhang, J. Aanet: Adaptive aggregation network for efficient stereo matching. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 1959–1968. [Google Scholar]
  21. Khamis, S.; Fanello, S.; Rhemann, C.; Kowdle, A.; Valentin, J.; Izadi, S. Stereonet: Guided hierarchical refinement for real-time edge-aware depth prediction. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 573–590. [Google Scholar]
  22. Zhang, Y.; Li, K.; Li, K.; Wang, L.; Zhong, B.; Fu, Y. Image super-resolution using very deep residual channel attention networks. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 286–301. [Google Scholar]
  23. Tong, W.; Chen, W.; Han, W.; Li, X.; Wang, L. Channel-attention-based DenseNet network for remote sensing image scene classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 4121–4132. [Google Scholar] [CrossRef]
  24. Song, X.; Dai, Y.; Zhou, D.; Liu, L.; Li, W.; Li, H.; Yang, R. Channel attention based iterative residual learning for depth map super-resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 5631–5640. [Google Scholar]
  25. Fu, Y.-H.; Li, Z.; Zhang, Y.-P.; Pan, D.-Y. Exudate Detection for Retinal Fundus Image Based on U-net Incorporating Residual Module. J. Chin. Comput. Syst. 2021, 42, 1479–1484. [Google Scholar]
  26. Kong, F.; Picetti, F.; Lipari, V.; Bestagini, P.; Tubaro, S. Deep prior-based seismic data interpolation via multi-res u-net. In Proceedings of the SEG International Exposition and Annual Meeting, Virtual, 14 October 2020. [Google Scholar]
  27. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  28. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; Springer: Cham, Switzerland, 2015; pp. 234–241. [Google Scholar]
  29. Xiao, X.; Lian, S.; Luo, Z.; Li, S. Weighted res-unet for height-quality retina vessel segmentation. In Proceedings of the 2018 9th International Conference on Information Technology in Medicine and Education (ITME), Hangzhou, China, 19–21 October 2018; pp. 327–331. [Google Scholar]
  30. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  31. Chen, L.C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. DeepLab: Semantic Image Segmenta-tion with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 40, 834–848. [Google Scholar] [CrossRef] [Green Version]
  32. Çiçek, Ö.; Abdulkadir, A.; Lienkamp, S.S.; Brox, T.; Ronneberger, O. 3D U-Net: Learning dense volumetric segmentation from sparse annotation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Athens, Greece, 17–21 October 2016; Springer: Cham, Switzerland, 2016; pp. 424–432. [Google Scholar]
  33. Wang, P.; Qiang, Y.; Yang, X.; Hou, T. Network Model for Lung Nodule Segmentation Based on Double Attention 3D-UNet. Comput. Eng. 2021, 47, 307–313. [Google Scholar]
  34. Wang, F.; Galliani, S.; Vogel, C.; Speciale, P.; Pollefeys, M. Patchmatchnet: Learned multi-view patchmatch stereo. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 14194–14203. [Google Scholar]
  35. Geiger, A.; Lenz, P.; Stiller, C.; Urtasun, R. Vision meets robotics: The kitti dataset. Int. J. Robot. Res. 2013, 32, 1231–1237. [Google Scholar] [CrossRef] [Green Version]
  36. Scharstein, D.; Hirschmüller, H.; Kitajima, Y.; Krathwohl, G.; Nešić, N.; Wang, X.; Westling, P. Height-resolution stereo datasets with subpixel-accurate ground truth. In Proceedings of the German Conference on Pattern Recognition, Münster, Germany, 2–5 September 2014; Springer: Cham, Switzerland, 2014; pp. 31–42. [Google Scholar]
  37. Song, X.; Dai, Y.; Qin, X. Deeply supervised depth map super-resolution as novel view synthesis. IEEE Trans. Circuits Syst. Video Technol. 2018, 29, 2323–2336. [Google Scholar] [CrossRef] [Green Version]
  38. Zbontar, J.; LeCun, Y. Computing the stereo matching cost with a convolutional neural network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1592–1599. [Google Scholar]
Figure 1. Disparity prediction results in the Scene Flow dataset; (a) is origin RGB data, (b) is Ground-Truth disparity Pseudo-color data, (c) is disparity prediction by stereonet, and (d) is our result. It can be seen from the red circle region that our method can also achieve an excellent disparity estimation in the morbid graphic area.
Figure 1. Disparity prediction results in the Scene Flow dataset; (a) is origin RGB data, (b) is Ground-Truth disparity Pseudo-color data, (c) is disparity prediction by stereonet, and (d) is our result. It can be seen from the red circle region that our method can also achieve an excellent disparity estimation in the morbid graphic area.
Sensors 22 04548 g001
Figure 2. U-net feature extractor.
Figure 2. U-net feature extractor.
Sensors 22 04548 g002
Figure 3. Our network‘s structure.
Figure 3. Our network‘s structure.
Sensors 22 04548 g003
Figure 4. GWD Module.
Figure 4. GWD Module.
Sensors 22 04548 g004
Figure 5. Different results with different cost volume construction methods in Stereo-net [21] ((a) is the real left camera’s image; (b) is the concatenate method; (c) is the inner product method; (d) is the subtract method).
Figure 5. Different results with different cost volume construction methods in Stereo-net [21] ((a) is the real left camera’s image; (b) is the concatenate method; (c) is the inner product method; (d) is the subtract method).
Sensors 22 04548 g005
Figure 6. Group-wise cost volume.
Figure 6. Group-wise cost volume.
Sensors 22 04548 g006
Figure 7. Residual module.
Figure 7. Residual module.
Sensors 22 04548 g007
Figure 8. The flow chart of the experimental procedure.
Figure 8. The flow chart of the experimental procedure.
Sensors 22 04548 g008
Figure 9. Visual comparison of different SR approaches on scene flow datasets for 4× enlargement.
Figure 9. Visual comparison of different SR approaches on scene flow datasets for 4× enlargement.
Sensors 22 04548 g009
Figure 10. The rendering of different construction methods of cost volume. (A) is the left view RGB image, (B) is the GT depth map, we analyze the red frame area in the original image, and the red circle part represents the image area where depth recovery is difficult. (C1C3) is the depth map generated by concatenate, ours and subtract methods, and (D1D3) is the corresponding pseudo-color map. It can be seen that the local details of the (C1) image are more affluent than the others, and the accuracy is height. The accuracy of the (C3) image is low, but the overall image is relatively smooth. Our method combines the advantages of them.
Figure 10. The rendering of different construction methods of cost volume. (A) is the left view RGB image, (B) is the GT depth map, we analyze the red frame area in the original image, and the red circle part represents the image area where depth recovery is difficult. (C1C3) is the depth map generated by concatenate, ours and subtract methods, and (D1D3) is the corresponding pseudo-color map. It can be seen that the local details of the (C1) image are more affluent than the others, and the accuracy is height. The accuracy of the (C3) image is low, but the overall image is relatively smooth. Our method combines the advantages of them.
Sensors 22 04548 g010
Figure 11. The deformable convolution ablation experiment results on scene flow (the largest figure shows the GT depth map, black dots represent the center of the convolution kernel. (A) is the depth map generated by this method, and (B) is the depth map without deformation convolution). It can be seen from the figure that the rectangle search window has caused depth misjudgment due to the integration of front and back scene information. The deformation convolution adaptively selects the same plane depth to prevent this problem.
Figure 11. The deformable convolution ablation experiment results on scene flow (the largest figure shows the GT depth map, black dots represent the center of the convolution kernel. (A) is the depth map generated by this method, and (B) is the depth map without deformation convolution). It can be seen from the figure that the rectangle search window has caused depth misjudgment due to the integration of front and back scene information. The deformation convolution adaptively selects the same plane depth to prevent this problem.
Sensors 22 04548 g011
Figure 12. The rendering of the ABPN ablation study on Kitti2015. (A,C) show the network generation results trained with ABPN. (B,D) show the results generated by the training model with bicubic interpolation. The significant error area is marked with red dots in the figure.
Figure 12. The rendering of the ABPN ablation study on Kitti2015. (A,C) show the network generation results trained with ABPN. (B,D) show the results generated by the training model with bicubic interpolation. The significant error area is marked with red dots in the figure.
Sensors 22 04548 g012
Table 1. Quantitative evaluation of ABPN including PSNR and SSIM for scale 4×, 8×, and 16×.
Table 1. Quantitative evaluation of ABPN including PSNR and SSIM for scale 4×, 8×, and 16×.
Algorithm16×
PSNRSSIMPSNRSSIMPSNRSSIM
Bicubic23.270.60321.350.54417.760.434
ABPN27.010.80723.020.63720.270.502
Table 2. Ablation experiment of cost volume construction method on KITTI2015.
Table 2. Ablation experiment of cost volume construction method on KITTI2015.
MethodEPE (%)3-px Error (%)Run-Time (s)
ALLNocD1-bgD1-All
Concat0.970.882.653.010.12
Subtract1.101.052.993.370.08
Ours0.820.802.442.860.15
Table 3. Ablation study of Deformable convolution.
Table 3. Ablation study of Deformable convolution.
MethodScene FlowKITTI 2015
EPE>1 pxEPED1-All
w/o in MDS1.1912.10.912.98
w/o in Res1.2312.50.933.05
w/o1.3313.91.003.11
ASR-Net1.0110.30.842.86
Table 4. Comparison of SceneFlow EPE error.
Table 4. Comparison of SceneFlow EPE error.
MethodGC-Net [6]PSMNet [13]DispNetC [11]GWC-Net [14]StereoNet [19]ASR-Net
EPE2.521.091.681.121.101.01
Time (s)0.90.40.060.130.0150.15
Table 5. Performance on KITTI test set.
Table 5. Performance on KITTI test set.
MethodKITTI2012KITTI2015Time
(s)
AllNocD1-bgD1-All
GC-Net [6]2.301.772.212.870.9
MC-CNN [38]3.632.432.893.8967
PSMNet [13]1.891.491.862.320.41
GWC-Net [14]1.701.321.742.110.32
DispNetC [11]4.654.114.324.340.06
StereoNet [19]6.024.914.304.830.015
ASR-Net2.662.172.352.860.15
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Liu, B.; Chen, K.; Peng, S.-L.; Zhao, M. Adaptive Aggregate Stereo Matching Network with Depth Map Super-Resolution. Sensors 2022, 22, 4548. https://doi.org/10.3390/s22124548

AMA Style

Liu B, Chen K, Peng S-L, Zhao M. Adaptive Aggregate Stereo Matching Network with Depth Map Super-Resolution. Sensors. 2022; 22(12):4548. https://doi.org/10.3390/s22124548

Chicago/Turabian Style

Liu, Botao, Kai Chen, Sheng-Lung Peng, and Ming Zhao. 2022. "Adaptive Aggregate Stereo Matching Network with Depth Map Super-Resolution" Sensors 22, no. 12: 4548. https://doi.org/10.3390/s22124548

APA Style

Liu, B., Chen, K., Peng, S. -L., & Zhao, M. (2022). Adaptive Aggregate Stereo Matching Network with Depth Map Super-Resolution. Sensors, 22(12), 4548. https://doi.org/10.3390/s22124548

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop