Next Article in Journal
Countermovement Rebound Jump: A Comparison of Joint Work and Joint Contribution to the Countermovement and Drop Jump Tests
Next Article in Special Issue
A Regional Brightness Control Method for a Beam Projector to Avoid Human Glare
Previous Article in Journal
Corn-Based Gluten-Free Snacks Supplemented with Various Dried Fruits: Characteristics of Physical Properties and Effect of Variables
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Hourglass 3D CNN for Stereo Disparity Estimation for Mobile Robots

1
School of Mechanical Engineering, Hanoi University of Science and Technology, Hanoi 100000, Vietnam
2
Graduate School of Engineering and Science, Shibaura Institute of Technology, Tokyo 135-8548, Japan
3
Department of Information and Communications Engineering, Shibaura Institute of Technology, Tokyo 135-8548, Japan
*
Author to whom correspondence should be addressed.
Appl. Sci. 2023, 13(19), 10677; https://doi.org/10.3390/app131910677
Submission received: 24 August 2023 / Revised: 20 September 2023 / Accepted: 24 September 2023 / Published: 26 September 2023
(This article belongs to the Special Issue Multimedia Systems Studies)

Abstract

:
Stereo cameras allow mobile robots to perceive depth in their surroundings by capturing two separate images from slightly different perspectives. This is necessary for tasks such as obstacle avoidance, navigation, and spatial mapping. By utilizing a convolutional neural network (CNN), existing works in stereo cameras based on depth estimation have achieved superior results. However, the critical requirement for depth estimation for mobile robots is to have an optimal tradeoff between computational cost and accuracy. To achieve such a tradeoff, attention-aware feature aggregation (AAFS) has been proposed for real-time stereo matching on edge devices. AAFS includes multistage feature extraction, an attention module, and a 3D CNN architecture. However, its 3D CNN architecture learns contextual information ineffectively. In this paper, a deep encoder–decoder architecture is applied to an AAFS 3D CNN to improve depth estimation accuracy. Through evaluation, it is proven that the proposed 3D CNN architecture provides significantly better accuracy while keeping the inference time comparable to that of AAFS.

1. Introduction

Mobile robots have witnessed a surge in popularity and find versatile applications in numerous fields [1]. One compelling use case for mobile robots is their deployment in hazardous environments, such as automated agriculture and the handling of dangerous materials, where they can replace human workers [2]. However, to ensure optimal performance, it is imperative for mobile robots to swiftly and accurately gauge the geometric attributes of their surroundings, specifically the depth information. Depth estimation plays a pivotal role in enabling mobile robots to excel in various tasks. It empowers these robots with the capability to perform obstacle detection [3], construct detailed environmental maps [4], and facilitate object recognition [5]. One of the potential solutions for depth estimation is stereo matching [6]. Stereo matching is a computer vision technique that simulates human vision by analyzing a pair of 2D images captured from slightly different viewpoints to reconstruct 3D scenes. The primary objective of stereo matching is to establish correspondences between pixels in these input 2D images and, subsequently, to compute the corresponding depth values for each pixel. This process is executed by identifying the disparity, which denotes the horizontal displacement between correspondences in the 2D images [7]. The accurate calculation of this disparity is instrumental in calculating the depth information, thereby empowering mobile robots to navigate, interact with, and operate effectively in their surroundings.
In order to accurately determine the disparity, recent studies have applied deep learning methods and achieved promising results [8]. Particularly, these works first used a convolutional neural network (CNN) to extract features from 2D images, then concatenate them and store the disparity values between them to construct a 4D cost volume ( h e i g h t × w i d t h × d i s p a r i t y × f e a t u r e s ) . Then, the 4D cost volume is input in a 3D CNN model for regularization into a 3D cost volume ( h e i g h t × w i d t h × d i s p a r i t y ) . Finally, the predicted disparity is regressed from the cost volume via a softmax operation ( σ ) [9].
For example, GC-Net [9] proposes to learn the context of cost volume through an encoder–decoder 3D CNN architecture. PSMNet [10] utilizes a feature extractor with a spatial pyramid pooling module and regularizes the cost volume using a 3D CNN based on stacked hourglass architecture. GA-Net [11] incorporates a 3D CNN with semiglobal matching for cost filtering. These approaches have demonstrated cutting-edge performance in stereo matching. Despite the high accuracy, when applying these methods to mobile robots, which often have low computational power, the computational cost is also a critical challenge.
As reported in [12], PSMNet [10] could only run at approximately 0.16 frames per second (fps) on an NVIDIA Jetson TX2 module. Similarly, although it has been proposed specifically for mobile robots, StereoNet [13] could only provide fewer than 2 fps on the same device. Such performances are far from the requirement for real-time applications in mobile robots, which is often a minimum of 30 fps [14].
Recently, the authors of [12] proposed attention-aware feature aggregation (AAFS) to obtain a better tradeoff between computational time and accuracy for real-time stereo matching on edge devices. The authors reported that AAFS could run at a maximum frame rate of 33 fps on low-budget devices, such as an NVIDIA Jetson TX2 module. However, the accuracy of AAFS is still limited due to the fact that it cannot efficiently exploit the contextual information of stereo images. The reason is that AAFS attempts to not increase the number of feature maps in its cascaded 3D CNN to limit the computational cost. In this case, leveraging the idea of a deep convolutional encoder–decoder, which is intended for dense prediction tasks, is a potential solution. Deep encoder–decoder tasks could reduce the computational cost by compressing the input data, then decoding the compressed data back to the input data dimension [15]. For example, a stacked hourglass based on a deep encoder–decoder consists of hourglass blocks that apply two-stride 3D convolutions to reduce the cost volume size by a factor of four [16]. This allows for an increase in feature dimensions with little impact on computational resources. Then, 3D transposed convolutions are applied to decode the volume to the original dimension.
Therefore, in this paper, we propose an improvement of AAFS by utilizing a stacked hourglass (encoder–decoder) architecture in a 3D CNN [16] to efficiently exploit the contextual information. Specifically, cost volume is processed iteratively in a top-down/bottom-up manner to better utilize global contextual information. The results demonstrate that the proposed method provides significantly better accuracy while requiring similar computational costs compared to AAFS.
The remainder of this paper is structured as follows. Section 2 provides a summary of existing works. In Section 3, the details of the disparity estimation framework with the proposed enhancement by a deep encoder–decoder 3D CNN are provided. Section 4 presents an experimental evaluation and analysis of results. Lastly, conclusions and directions for future work are presented in Section 5.

2. Related Work

Zbontar et al. originally introduced a CNN-based stereo-matching technique [17] whereby the similarity metric of tiny patch pairings was learned using a convolutional neural network. GCNet [9] was one of the first methods incorporating 4D cost volume, using the soft argmin operation in the disparity regression steps to obtain the best matching disparity. PSMNet [10] introduced a spatial pyramid pooling module and 3D stacked hourglass networks and yielded promising results. The authors of [18] proposed GwcNet, which is a modified 3D stacked hourglass architecture, and a combined 3D cost volume based on group-wise correlation. GA-Net [11] includes a semiglobal aggregation layer and a local guided aggregation layer to replace several 3D convolution layers. To replace the 3D architecture, AANet [19] includes an intrascale and cross-scale cost aggregation algorithm, which can reduce inference time and maintain equivalent accuracy. On the other hand, DeepPruner [20], a coarse-to-fine approach, includes a differentiable PathMatch-based module to estimate the pruned search range of each pixel. Although 4D cost volume-based methods have achieved promising results, they operate at high computational cost and do not accommodate real-time operation on low-budget devices.
Therefore, some recent studies have focused on lightweight stereo networks based on 4D cost volumes to achieve real-time performance while maintaining competitive accuracy. These methods typically construct and aggregate 3D cost volume at low resolution to significantly reduce computational cost. For instance, StereoNet [13] is an edge-preserving refinement network that utilizes the left images as guidance to recover high-frequency details. Gu et al. [21] proposed a cascade cost volume, which consists of two stages. Cost volume at the early stage is built on a low-resolution feature map. Then, the later stage used the estimated disparity maps from the earlier stage to construct new cost volumes to apply better semantic features. This leads to a remarkable improvement in GPU memory consumption and computation time. AAFS [12] constructs a 4D cost volume by adopting a distance metric (height × width × disparity × 1). A disparity map is then computed at the lowest resolution, and disparity residuals are computed in later stages. However, its 3D CNN cannot exploit the contextual information for cost volume regularization, resulting in a limitation in estimation accuracy.

3. Methodology

In this study, we propose the integration of an hourglass architecture to the 3D CNN of the AAFS architecture [12] to better exploit the contextual information in stereo images. Such an enhancement helps achieve better disparity estimation accuracy while maintaining a competitive computational cost. As this paper is based on AAFS, we first introduce the AAFS architecture [12], then briefly describe the proposed hourglass-based 3D CNN. Figure 1 shows the architecture of AAFS [12]. The model consists of multistage feature extraction with an attention-aware aggregation module and a 3D CNN network for disparity and residual map prediction. Section 3.1 describes multiscale feature extraction based on [12]. Section 3.2 introduces correlation cost volume based on [12]. In Section 3.3, we introduced our proposed 3D CNN architecture. Section 3.4 describes disparity regression. Finally, Section 3.5 describes the loss function.

3.1. Attention-Aware Feature Extraction

The authors of [12] proposed a feature extractor with an attention-aware module to obtain multistage feature maps from stereo images. Figure 2 depicts the architecture of the feature extractor, which consists of an efficient backbone, an attention module, and three feature aggregation modules.
With the attention mechanism, the feature extractor can adaptively collect input from multiple scales into encoded features. With an efficient backbone, the blueprint separable convolution (BSConv) [22] is utilized to justify DSConv [23] by analyzing intrakernel correlations of vanilla CNNs. For the purpose of multiscale representation, BSConv [22] is employed to progressively decrease the dimensions of feature maps and generate a sequence of feature maps with varying scales. Feature extraction produces output feature maps at three levels (1/4, 1/8, and 1/16). AAFS [12] includes three aggregation modules for three levels of features. Then, an attention module is employed to recalibrate channel importance and boost feature discriminability. The cost volumes for disparity estimation are formed from the final updated feature maps for each level.

3.2. Cost Volume

For real-time constraints, AAFS [12] employs a correlation volume to create a single-channel 4D cost volume for each disparity level with dimensions of ( 1 × d i s p a r i t y × h e i g h t × w i d t h ) . Initially, AAFS [12] computes the full disparity map at a resolution of 1/16 in the first stage, followed by the computation of disparity residuals in later stages. This limits the search range for correspondences to 5, with offsets of −2, −1, 0, 1, and 2. To compute the residuals in later stages, AAFS [12] first scales up the rough disparity map to warp the input features. This is done by applying pixel-wise disparity estimation. Assuming that the current disparity estimation is accurate, updated right feature maps should line up with left feature maps. Estimation of residual disparity follows a similar process as the computation of the full disparity map but applying a constraint range of −2 to 2 for the residual disparity map. Then, the up-scaled disparity map from the earlier stage is combined with the resulting residual disparity map.

3.3. Hourglass 3D CNN Architecture

To aggregate the feature information from the disparity spatial dimensions, AAFS [12] includes a cascaded architecture for the 3D CNN to regularize the cost volume. Such a 3D CNN consists of five 3 × 3 × 3 convolutions followed by batch normalization and ReLU, except for the final layer. Since AAFS [12] aims to achieve a tradeoff between accuracy and computational cost, it attempts to not increase the number of feature maps in the cascaded 3D CNN. In our work, an hourglass 3D CNN is proposed to replace the cascaded 3D CNN to improve the tradeoff. The architecture of the proposed 3D CNN is shown in Figure 3, with a more detailed definition of each layer proposed in Table 1. Our proposed architecture allows the model to effectively learn more contextual information, leading to accuracy improvements and maintaining a low computational cost for real-time applications on mobile devices.
As shown in Table 1, F features that represent the number of additional dimensions are set to 8, 4, and 4 for stages 0, 1, and 2, respectively. In a cascaded 3D CNN, the additional dimension is kept the same, with a stride of one. Thus, to increase accuracy, the additional dimension needs to be increased, which leads to a large increase in the computational cost because the same resolution is maintained. An hourglass 3D CNN is based on an encoder–decoder architecture, which reduces the required amount of computation by compressing the spatial dimension of cost volume and learning more feature dimensions, then upsampling back to the original dimension of cost volume in the decoder. Specifically, we employ two sequential 3 × 3 × 3 convolutions with a stride of one followed by batch normalization and ReLU; the number of feature maps is F for the cost volume. Then, we compress the spatial dimension of volume by four times with two levels of the encoder. In each encoder level, two sequential 3 × 3 × 3 convolutions are applied, and the stride of the first layer is two. This results in the spatial dimension of the volume decreasing by a factor of two, whereas F is doubled. As a result, the size of the cost volume decreases by four times, and the number of feature maps is 4F. In addition, resolution loss decreases spatial accuracy and fine-grained details. To tackle this problem, we use two residual connections, as shown in Figure 3, to retain higher-frequency information. Then, we apply two 3D transposed convolutions with a stride of two; and the number of feature maps is 2F and F, respectively. Finally, 3 × 3 × 3 convolutions in series with a stride of one and a single-feature output in the last layer are applied to predict densities. This results in a final, regularized cost volume with a size of d i s p a r i t y × h e i g h t × w e i g h t .

3.4. Disparity Regression

Disparity regression [9] is utilized to estimate the disparity map and is calculated as Equation (1).
d ^ = d = 0 D m a x d × σ c d
where d ^ is the predicted disparity and is calculated as the sum of each disparity probability d. Finally, d is calculated from the cost function ( c d ) based on a softmax operation ( σ (·)).
As in [9], using the above disparity regression achieves better results in comparison with stereo-matching methods based on classification.

3.5. Loss Function

In our proposal, the loss function is expressed as:
L = i = 0 i = 2 λ i S m o o t h L 1 d i ^ d g t
where d g t is the ground truth disparity, λ i is the loss weight, and d i ^ is the predicted disparity at stage i = 0, 1, 2.

4. Experiments

4.1. Datasets

Figure 4 shows example scenes of the following datasets, consisting of left image, right image and ground-truth disparity.
  • Scene Flow [24]: A large dataset with 960 × 540 pixel resolution and 35,454 training and 4370 testing stereo images. Disparity maps are presented as the ground truth. If the disparity exceeds the restrictions established in our experiment (the maximum disparity is set at 192), some pixels with large disparities are removed from the loss computation. We use the Cleanpass of the Scene Flow dataset, which contains lighting and shading effects and no additional effects.
  • KITTI 2012 [25]: A real-world dataset containing 194 and 195 moving street view color stereo images (1240 × 376 resolution) with and without ground-truth disparities, respectively, taken from an autonomous car. In this experiment, the images with ground-truth disparities were used as training data and were split into a training set (160 images) and a validation set (34 images).
  • KITTI 2015 [26]: A newer version of KITTI 2012, which increases the number of images both with and without ground-truth disparities to 200. Similar to KITTI 2012, the images with ground-truth disparities were used as training data and were split into a training set (160 images) and a validation set (40 images).

4.2. Evaluation Settings

The proposed method was implemented on PyTorch [27]. The model was trained using an Adam optimizer with β 1 = 0.9 and β 2 = 0.999. The input images were randomly selected and randomly cropped to 512 × 256 size based on the random forest approach [28] during the training process. We set the maximum disparity (D) to 192.
First, we trained the model with the Scene-Flow dataset [24] for 10 epochs with a learning rate of 0.001. For evaluation on the Scene-Flow dataset, the trained model was directly used. For validation on KITTI datasets, the model trained with the Scene-Flow dataset was used for fine tuning on for 300 epochs. For the first 200 epochs, we set the learning rate of fine tuning to 0.001, and for the final 100 epochs, to 0.0001. For KITTI 2012 and 2015 submission, we combined both KITTI 2012 and KITTI 2015 training sets to fine tune the pretrained model on KITTI for 500 epochs and used the KITTI 2012 and 2015 validation sets for the KITTI 2012 and 2015 submission, respectively. For the first 300 epochs, we set the learning rate of fine tuning to 0.001, and for the final 200 epochs, to 0.0001. We set the batch size to 12 for Scene-Flow and to 6 for KITTI. Training was implemented on an NVIDIA GTX 1660. We used smooth L1 loss. We set the loss weights of the three output stages to 0.3, 0.5, and 1.0.
We evaluated the performance of our approach using two metrics. For the Scene-Flow dataset [24], we calculated endpoint error (EPE), which is the mean absolute error (MAE) between disparity prediction and ground-truth disparity. For the KITTI 2012 [25] and KITTI 2015 [26] datasets, we used three-pixel error (3px), which is the percentage of pixels with a disparity of less than three pixels or less than 5% of the ground-truth disparity. To evaluate the computational costs, we relied on floating-point operations (FLOPs). The processing speed was clocked using Google Colab (Nvidia Tesla T4) in milliseconds (ms).
The proposed method was evaluated against AAFS [12] and other state-of-the-art methods, namely PSMNet [10], DeepPruner-fast [20], and StereoNet [13]. The results are summarized and discussed in the next section.

4.3. Evaluation Results and Discussion

Table 2 provides the performance of the evaluated methods in terms of 3-px error (for KITTI datasets), as well as their computational cost and real-time ability (for Scene Flow dataset), while Table 3 compares the EPE (on the Scene-Flow dataset) with that of real-time models.
Overall, according to Table 2, we can observe a common tendency. The higher the accuracy, the more computational cost is required. In particular, DeepPrunner-fast [20] and PSMNet [10] produced impressively accurate results in terms of 3-px error. Those values were all less than 3%. As a result, those FLOPs and running times are the most severe. For example, DeepPruner-fast’s FLOPs is 208.52 G, whereas its running time is 210 ms. From the table, we can also find the lowest accuracy in AAFS and StereoNet. However, their FLOPs and running time are relatively small. For example, StereoNet’s FLOP is 59.82 G, and its running time is 180 ms. As shown in Table 3, the proposed method significantly reduced EPE by 0.7 on AAFS [12] and 0.4 on StereoNet-16x [13] on the Scene-Flow test set.
In general, our proposed method is remarkably faster than the other baselines because the disparity was computed at low resolution. Specifically, our proposal was 7.5 times faster than StereoNet-16x [13] and 14 times faster than DeepPruner-fast [20]. The proposed method outperforms StereoNet-16x [13] by a considerable margin, with only 1 billion FLOPs, clearly proving that the proposed method can accurately estimate disparity maps in real time in mobile robots.
Figure 5 compares the results acquired from the KITTI 2012 and 2015 leaderboard for PSMNet [10] and AAFS [12]. In KITTI 2012 (in Figure 5a), the error map scales linearly from 0 to >=5 pixel error; red color represents all occluded pixels that are outside the image limits. In KITTI 2015 (in Figure 5b), the error map depicts correct estimates; <3 px or <5% error are indicated in blue, and incorrect estimates are indicated in red, whereas dark regions denote the occluded pixels that lie outside the image limits. Through error maps, PSMNet [10] achieved the best results, and our proposed method improved the accuracy at the boundary of the object compared to AAFS [12]. Beyond the boundary areas, our proposed method and AAFS [12] have the same accuracy as PSMNet [10], which shows that our proposed method can accurately estimate disparity maps for real-world applications while requiring a low computational cost.
Benchmark Results For benchmarking, KITTI 2012/2015 test sets were used for evaluation. As shown by the online leaderboard in Table 4, the overall thee-pixel error of the proposed model was 5.64% on KITTI 2012 and 6.14% on KITTI 2015. On the other hand, PSMNet [10], DeepPruner-fast [20], GWCNet [18], and Fast-ACVNet [29] are state-of-the-art approaches that [29] can yield accurate results; however, they cannot be applied to mobile robots due to their significant latency (>140 ms). In [29], Fast-ACVNet achieved 22–23 fps with the an NVIDIA RTX 3090 GPU, which is a challenge to deploy in mobile robots for real-time applications. In terms of the Noc metric (the proportion of incorrect pixels in non-occluded areas), our proposal improved from 6.10% to 4.93% on the KITTI 2012 test set and from 7.12% to 5.78% on the KITTI 2015 test set compared to AAFS [12]. Moreover, in terms of the All metric (the proportion of incorrect pixels in total), our proposal significantly improved from 6.94% to 5.64% on the KITTI 2012 test set and from 7.54% to 6.14% on the KITTI 2015 test set. We evaluated runtime based on Google Colab (Nvidia Tesla T4), and the results showed that our method is approximately equal to AAFS [12].

5. Conclusions

Recent studies have proven that utilizing CNNs significantly improves the performance of stereo matching. However, for application like mobile robots, it is still a challenge to estimate disparity in real time. In this work, inspired by AAFS [12], we propose the application of a 3D CNN based on a stacked-hourglass architecture to efficiently aggregate the feature information with the disparity dimension. Through experiments, the proposed method has been proven to achieve better accuracy than AAFS while requiring a similar computational cost. The predicted disparity map undoubtedly proves the ability of the proposed method to accurately estimate depth for real applications.

Author Contributions

Conceptualization, T.L., C.M.T. and T.N.D.; Methodology, T.L., C.M.T. and T.N.D.; Supervision, L.T., E.K. and P.X.T.; Writing—original draft, T.L., C.M.T. and T.N.D.; Writing—review and editing, L.T., E.K. and P.X.T. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Campbell, S.; O’Mahony, N.; Carvalho, A.; Krpalkova, L.; Riordan, D.; Walsh, J. Path Planning Techniques for Mobile Robots A Review. In Proceedings of the 2020 6th International Conference on Mechatronics and Robotics Engineering (ICMRE), Barcelona, Spain, 12–15 February 2020; pp. 12–16. [Google Scholar] [CrossRef]
  2. Vasic, M.; Billard, A. Safety Issues in Human-robot Interactions. In Proceedings of the 2013 IEEE International Conference on Robotics and Automation, Karlsruhe, Germany, 6–10 May 2013; IEEE: Piscataway, NJ, USA, 2013; pp. 197–204. [Google Scholar]
  3. Wei, B.; Gao, J.; Li, K.; Fan, Y.; Gao, X.; Gao, B. Indoor Mobile Robot Obstacle Detection Based on Linear Structured Light Vision System. In Proceedings of the 2008 IEEE International Conference on Robotics and Biomimetics, Bangkok, Thailand, 22–25 February 2009; IEEE: Piscataway, NJ, USA, 2009; pp. 834–839. [Google Scholar]
  4. Amigoni, F.; Caglioti, V. An information-based exploration strategy for environment mapping with mobile robots. Robot. Auton. Syst. 2010, 58, 684–699. [Google Scholar] [CrossRef]
  5. Ekvall, S.; Jensfelt, P.; Kragic, D. Integrating Active Mobile Robot Object Recognition and Slam in Natural Environments. In Proceedings of the 2006 IEEE/RSJ International Conference on Intelligent Robots and Systems, Beijing, China, 9–15 October 2006; IEEE: Piscataway, NJ, USA, 2006; pp. 5792–5797. [Google Scholar]
  6. Murray, D.; Little, J.J. Using Real-time Stereo Vision for Mobile Robot Navigation. Auton. Robot. 2000, 8, 161–171. [Google Scholar] [CrossRef]
  7. Scharstein, D.; Szeliski, R.; Zabih, R. A Taxonomy and Evaluation of Dense Two-frame Stereo Correspondence Algorithms. In Proceedings of the IEEE Workshop on Stereo and Multi-Baseline Vision (SMBV 2001), Kauai, Hawaii, 9–10 December 2001; pp. 131–140. [Google Scholar] [CrossRef]
  8. Flynn, J.; Neulander, I.; Philbin, J.; Snavely, N. Deep Stereo: Learning to Predict New Views from the World’s Imagery. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 5515–5524. [Google Scholar] [CrossRef]
  9. Kendall, A.; Martirosyan, H.; Dasgupta, S.; Henry, P.; Kennedy, R.; Bachrach, A.; Bry, A. End-to-End Learning of Geometry and Context for Deep Stereo Regression. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 66–75. [Google Scholar] [CrossRef]
  10. Chang, J.R.; Chen, Y.S. Pyramid Stereo Matching Network. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 5410–5418. [Google Scholar] [CrossRef]
  11. Zhang, F.; Prisacariu, V.; Yang, R.; Torr, P.H. GA-Net: Guided Aggregation Net for End-to-end Stereo Matching. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 185–194. [Google Scholar]
  12. Chang, J.R.; Chang, P.C.; Chen, Y.S. Attention-Aware Feature Aggregation for Real-Time Stereo Matching on Edge Devices. In Proceedings of the Computer Vision–ACCV 2020; Springer International Publishing: Cham, Switzerland, 2021; pp. 365–380. [Google Scholar] [CrossRef]
  13. Khamis, S.; Fanello, S.; Rhemann, C.; Kowdle, A.; Valentin, J.; Izadi, S. StereoNet: Guided Hierarchical Refinement for Real-Time Edge-Aware Depth Prediction. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018. [Google Scholar]
  14. Poppinga, B.; Laue, T. JET-Net: Real-time Object Detection for Mobile Robots. In Proceedings of the RoboCup 2019: Robot World Cup XXIII 23; Springer: Berlin/Heidelberg, Germany, 2019; pp. 227–240. [Google Scholar]
  15. Badrinarayanan, V.; Kendall, A.; Cipolla, R. SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar] [CrossRef]
  16. Newell, A.; Yang, K.; Deng, J. Stacked Hourglass Networks for Human Pose Estimation. In Proceedings of the Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Proceedings, Part VIII 14. Springer: Berlin/Heidelberg, Germany, 2016; pp. 483–499. [Google Scholar]
  17. Zbontar, J.; LeCun, Y. Stereo matching by training a convolutional neural network to compare image patches. J. Mach. Learn. Research 2016, 17, 1–32. [Google Scholar]
  18. Guo, X.; Yang, K.; Yang, W.; Wang, X.; Li, H. Group-wise Correlation Stereo Network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 3273–3282. [Google Scholar]
  19. Xu, H.; Zhang, J. AANet: Adaptive Aggregation Network for Efficient Stereo Matching. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 1959–1968. [Google Scholar]
  20. Duggal, S.; Wang, S.; Ma, W.; Hu, R.; Urtasun, R. DeepPruner: Learning Efficient Stereo Matching via Differentiable PatchMatch. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; IEEE Computer Society: Los Alamitos, CA, USA, 2019; pp. 4383–4392. [Google Scholar] [CrossRef]
  21. Gu, X.; Fan, Z.; Zhu, S.; Dai, Z.; Tan, F.; Tan, P. Cascade Cost Volume for High-Resolution Multi-View Stereo and Stereo Matching. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 14–19 June 2020; pp. 2492–2501. [Google Scholar] [CrossRef]
  22. Haase, D.; Amthor, M. Rethinking Depthwise Separable Convolutions: How Intra-Kernel Correlations Lead to Improved MobileNets. In Proceedings of the The IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
  23. Chollet, F. Xception: Deep learning with depthwise separable convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1251–1258. [Google Scholar]
  24. Mayer, N.; Ilg, E.; Häusser, P.; Fischer, P.; Cremers, D.; Dosovitskiy, A.; Brox, T. A Large Dataset to Train Convolutional Networks for Disparity, Optical Flow, and Scene Flow Estimation. In Proceedings of the IEEE International Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  25. Geiger, A.; Lenz, P.; Urtasun, R. Are we ready for Autonomous Driving? The KITTI Vision Benchmark Suite. In Proceedings of the Conference on Computer Vision and Pattern Recognition (CVPR), Providence, RI, USA, 16–21 June 2012. [Google Scholar]
  26. Menze, M.; Geiger, A. Object Scene Flow for Autonomous Vehicles. In Proceedings of the Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015. [Google Scholar]
  27. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L.; et al. Pytorch: An imperative style, high-performance deep learning library. Adv. Neural Inf. Process. Syst. 2019, 32, 8024–8035. [Google Scholar]
  28. Haeusler, R.; Nair, R.; Kondermann, D. Ensemble Learning for Confidence Measures in Stereo Vision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Portland, OR, USA, 23–28 June 2013. [Google Scholar]
  29. Xu, G.; Wang, Y.; Cheng, J.; Tang, J.; Yang, X. Accurate and Efficient Stereo Matching via Attention Concatenation Volume. arXiv 2022, arXiv:2209.12699. [Google Scholar]
Figure 1. Architectural overview of AAFS.
Figure 1. Architectural overview of AAFS.
Applsci 13 10677 g001
Figure 2. Attention-aware feature extractor.
Figure 2. Attention-aware feature extractor.
Applsci 13 10677 g002
Figure 3. Proposed 3D CNN architecture.
Figure 3. Proposed 3D CNN architecture.
Applsci 13 10677 g003
Figure 4. Example scenes from datasets. From top to bottom: Monkaa, Driving, FlyingThings3D in SceneFlow, and KITTI 2012 and KITTI 2015. From left to right: left image, right image, and ground-truth disparity.
Figure 4. Example scenes from datasets. From top to bottom: Monkaa, Driving, FlyingThings3D in SceneFlow, and KITTI 2012 and KITTI 2015. From left to right: left image, right image, and ground-truth disparity.
Applsci 13 10677 g004
Figure 5. Qualitative comparisons between PSMNet, AAFS, and our proposed method. (a) Qualitative comparison of results on the KITTI 2012 test set. From top to bottom: PSMNet, AAFS, and our proposal. From left to right: left image, disparity prediction, and error map. (b) Qualitative comparisons of results on the KITTI 2015 test set. From top to bottom: PSMNet, AAFS, and our proposal. From left to right: left image, disparity prediction, and error map.
Figure 5. Qualitative comparisons between PSMNet, AAFS, and our proposed method. (a) Qualitative comparison of results on the KITTI 2012 test set. From top to bottom: PSMNet, AAFS, and our proposal. From left to right: left image, disparity prediction, and error map. (b) Qualitative comparisons of results on the KITTI 2015 test set. From top to bottom: PSMNet, AAFS, and our proposal. From left to right: left image, disparity prediction, and error map.
Applsci 13 10677 g005
Table 1. Summary of our proposed architecture. Each 3D convolutional layer represents a block of convolution, batch normalization, and ReLU (except the final layer).
Table 1. Summary of our proposed architecture. Each 3D convolutional layer represents a block of convolution, batch normalization, and ReLU (except the final layer).
Layer DescriptionOutput Tensor Dim
Cost volume 1 16 D × 1 16 H × 1 16 W × 1
13D conv, 3 × 3 × 3 , F features 1 16 D × 1 16 H × 1 16 W × F
23D conv, 3 × 3 × 3 , F features 1 16 D × 1 16 H × 1 16 W × F
33D conv, 3 × 3 × 3 , 2F features, 1 32 D × 1 32 H × 1 32 W × 2 F
stride 2
43D conv, 3 × 3 × 3 , 2F features 1 32 D × 1 32 H × 1 32 W × 2 F
53D conv, 3 × 3 × 3 , 4F features, 1 64 D × 1 64 H × 1 64 W × 4 F
stride 2
63D conv, 3 × 3 × 3 , 4F features 1 64 D × 1 64 H × 1 64 W × 4 F
73D deconv, 3 × 3 × 3 , 2F features, 1 32 D × 1 32 H × 1 32 W × 2 F
stride 2, add layer 7 and 4
83D deconv, 3 × 3 × 3 , F features, 1 16 D × 1 16 H × 1 16 W × F
stride 2, add layer 8 and 2
93D conv, 3 × 3 × 3 , F features 1 16 D × 1 16 H × 1 16 W × F
103D conv, 3 × 3 × 3 , 1 features 1 16 D × 1 16 H × 1 16 W × 1
Table 2. Comparison of our proposed method with state-of-the-art accuracy-oriented methods on the KITTI 2012/2015 validation set.
Table 2. Comparison of our proposed method with state-of-the-art accuracy-oriented methods on the KITTI 2012/2015 validation set.
MethodKITTI 2012 (%)KITTI 2015 (%)FLOPs (G)Runtime (ms)
PSMNet [10]1.391.77737.841260
DeepPruner-fast [20]-2.59208.52210
StereoNet-16x [13]-6.3259.82180
AAFS [12]6.016.760.6112
Ours5.274.420.7113
Table 3. Comparison with the Scene-Flow test set.
Table 3. Comparison with the Scene-Flow test set.
MethodOursAAFSStereoNet-16x
EPE3.23.93.6
Table 4. Comparison on the KITTI 2012 and 2015 test benchmarks.
Table 4. Comparison on the KITTI 2012 and 2015 test benchmarks.
MethodKITTI 2012KITTI 2015ParamsTime
NocAllNocAll(M)(ms)
PSMNet [10]1.491.892.142.325.221260
Fast-ACVNet [29]--1.771.913.07140
DeepPruner-fast [20]--2.352.597.47210
GwcNet [18]1.321.701.922.116.52692
DeepPruner-best [20]--1.952.157.39685
AAFS [12]6.106.947.127.540.02312
Ours4.935.645.786.140.1213
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

La, T.; Tao, L.; Minh Tran, C.; Nguyen Duc, T.; Kamioka, E.; Tan, P.X. Hourglass 3D CNN for Stereo Disparity Estimation for Mobile Robots. Appl. Sci. 2023, 13, 10677. https://doi.org/10.3390/app131910677

AMA Style

La T, Tao L, Minh Tran C, Nguyen Duc T, Kamioka E, Tan PX. Hourglass 3D CNN for Stereo Disparity Estimation for Mobile Robots. Applied Sciences. 2023; 13(19):10677. https://doi.org/10.3390/app131910677

Chicago/Turabian Style

La, Thai, Linh Tao, Chanh Minh Tran, Tho Nguyen Duc, Eiji Kamioka, and Phan Xuan Tan. 2023. "Hourglass 3D CNN for Stereo Disparity Estimation for Mobile Robots" Applied Sciences 13, no. 19: 10677. https://doi.org/10.3390/app131910677

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop