Next Article in Journal
Maintenance Strategies for Industrial Multi-Stage Machines: The Study of a Thermoforming Machine
Next Article in Special Issue
Dynamic Object Tracking on Autonomous UAV System for Surveillance Applications
Previous Article in Journal
Security-Related Hardware Cost Optimization for CAN FD-Based Automotive Cyber-Physical Systems
Previous Article in Special Issue
A Joint 2D-3D Complementary Network for Stereo Matching
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Optimizing 3D Convolution Kernels on Stereo Matching for Resource Efficient Computations

Division of Electrical Engineering and Computer Science, Kanazawa University, Kanazawa 920-1192, Japan
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(20), 6808; https://doi.org/10.3390/s21206808
Submission received: 16 September 2021 / Revised: 4 October 2021 / Accepted: 8 October 2021 / Published: 13 October 2021

Abstract

:
Despite recent stereo matching algorithms achieving significant results on public benchmarks, the problem of requiring heavy computation remains unsolved. Most works focus on designing an architecture to reduce the computational complexity, while we take aim at optimizing 3D convolution kernels on the Pyramid Stereo Matching Network (PSMNet) for solving the problem. In this paper, we design a series of comparative experiments exploring the performance of well-known convolution kernels on PSMNet. Our model saves the computational complexity from 256.66 G MAdd (Multiply-Add operations) to 69.03 G MAdd ( 198.47 G MAdd to 10.84 G MAdd for only considering 3D convolutional neural networks) without losing accuracy. On Scene Flow and KITTI 2015 datasets, our model achieves results comparable to the state-of-the-art with a low computational cost.

1. Introduction

Stereo matching plays an important role in 3D computer vision applications, such as augmented reality (AR) [1], mixed reality (MR) [2], autonomous vehicle [3] and robot navigation [4,5]. It provides accurate disparity by a pair of stereo images. We can calculate the depth value by D = f B / d , where d denotes the disparity of the pixel, f is the focal length of the camera and B is the distance between the camera centers [6]. To get a precise disparity map is one of the most important tasks in stereo vision.
Classic stereo matching algorithms contain four parts: matching cost computation, cost support aggregation, disparity computation and disparity optimization [7]. Early studies perform machine learning methods to optimize disparity by Markov random field [8], conditional random field [9] or random forest [10]. With the rise of convolutional neural networks (CNNs), CNN-based approaches have been developed progressively. MC-CNN [6] first investigates CNNs on matching corresponding points for disparity estimation. Geometry and Context Network (GC-Net) [11] makes the training process end-to-end with a differentiable ArgMin operation on disparity estimations. The Pyramid Stereo Matching Network (PSMNet) [12] introduces spatial pyramid pooling and a stacked hourglass module for an accurate disparity map. These famous studies form a CNN-based approach framework: 2D Siamese feature extraction, cost volume aggregation, cost volume regularization and disparity regression.
One major problem with current CNN-based stereo matching algorithms is the enormous computation for cost volume regularization. The cost volume aggregation stage builds correspondence between left and right feature maps, aggregating disparity as an additional dimension on left feature maps to form 4D cost volumes [11]. For the cost volume regularization stage, most CNN-based methods build 3D convolution layers and 3D deconvolution layers, composing a 3D encoder–decoder architecture for regularizing 4D cost volumes. Compared to 2D convolution kernels, the additional dimension forming 3D convolution kernels raises computation complexity exponentially, leading the cost volume regularization to contain most of the computation among the entire architecture. Therefore, we build the 3D version of resource efficient kernel-based methods which can match up with the logic of a 3D encoder–decoder. By classifying all 3D layers in PSMNet according to their functions, we replace the original layers with our lightweight 3D convolution layers and implement a series of comparative experiments. Eventually, compared to the original PSMNet, we save 34.3 % parameters and 73.1 % multiply-add operations (MAdd) ( 95.50 % paramters and 94.53 % MAdd for 3D CNNs) without losing performance. We evaluate our model on Scene Flow and KITTI 2015 public datasets and obtain competitive results with other state-of-the-art stereo matching algorithms.
Our main contributions are listed below:
  • A throughout discussion about optimizing already established 3D convolution kernels on stereo matching algorithms;
  • A network design guideline when optimizing 3D convolution kernels on stereo matching algorithms for accurate disparity estimation and less computational complexity;
  • By following the guideline above and without changing the architecture, our model performs comparable results to modern stereo matching algorithms with significantly less computational complexity.

2. Related Works

2.1. Kernel-Based Methods

Plenty of recent studies focus on building lightweight convolution kernels, which comprise the CNN-based application suitable for low resource devices, such as mobile phone and navigation robot. The origin SqueezeNet [13] achieves AlexNet [14]-level accuracy with 50 times fewer parameters. Xception [15] and MobileNetV1 [16] implement depthwise separable convolutions for reducing the model parameters. MobileNetV2 [17] proposes an inverted residual block with channel expansion for boosting the performance. ShuffleNetV1 [18] presents grouped pointwise convolution and a channel shuffle operation to save computation. ShuffleNetV2 [19] further considers the relationship between hardware and network design, improving their performance in terms of speed and accuracy.
The channel-wise attention mechanism has proven the potential for enhancing the performance with a small implementation. Squeeze and excitation networks (SE-Net) [20] firstly present an effective attention mechanism by aggregating a feature map with global average pooling along the channel and weights it on the respective channel. Selective kernel networks (SK-Net) [21] improve the performance with optimizing the channel-wise information in two different sizes of receptive fields. Efficient channel attention networks (ECA-Net) [22] propose an effective channel attention module for saving computational burden.

2.2. Stereo Matching

GC-Net [11] settles the basic framework of end-to-end deep learning-based stereo matching algorithms: 2D Siamese CNNs, cost volume aggregation, cost volume regularization and disparity regression. Zhou et al. [23] draw a review of deep learning-based methods on stereo matching algorithms, which shows that recent stereo matching algorithms follow this framework and improve network performance by modifying partially with other mechanisms. We explain the main recent works according to where they are modified in Table 1.
According to Table 1, most of the works focus on optimizing the network architecture, especially the cost volume aggregation part, which costs the greatest computational resources. Yet, 3D convolution layers and 3D transposed convolution layers, the basic elements of cost volume aggregation, have been rarely studied. For exploring the limitation of the 3D kernel-based method on stereo matching methods, inspired by [39], we conduct a complete investigation optimizing network architecture with 3D kernel-based methods. We perform PSMNet as the basic architecture, attempting an elaborate study on all stages of the cost volume normalization with 3D kernel-based methods. By analyzing the model’s complexity and accuracy on Scene Flow and the KITTI 2015 data set, we receive a model with comparable results and less computational complexity.

3. Network Architecture

We first introduce the details about our network structure, including the architecture of the prototype PSMNet [12], and a series of 3D convolution kernels as the basic components of the network design pipeline.

3.1. PSMNet

In our paper, we take the PSMNet as the baseline and explore a series of 3D kernel-based methods to find a lightweight and accuracy model. In Figure 1, we separate all 3D convolution kernels into five categories: 3D head, 3D convolution, 3D convolution with s t r i d e = 2 , 3D deconvolution and 3D output. Table 2 shows the network settings of PSMNet, we optimize the network design by experimenting with the 3D kernel-based method on different stages of the architecture.
The 4D cost volumes ( d i s p a r i t y × h e i g h t × w i d t h × c h a n n e l ) are formed by concatenating the left and the right feature maps f l , f r ( h e i g h t × w i d t h × c h a n n e l ) in Equation (1):
C ( d , x , y , c h a n n e l ) = ( C o n c a t f l ( x , y ) , f r ( x d , y ) , c h a n n e l ) .
The 3D stacked hourglass module performs cost volume regularization. The continuous disparity map is obtained by the disparity regression process in [11]. The output disparity d ^ is calculated as the summation of each disparity d weighted by corresponding probability σ ( c d ) ,
d ^ = d = 0 D m a x d × σ ( c d ) .
The σ ( . ) denotes softmax operation, and the maximum disparity D m a x is set to 192.
For generating a smooth disparity map, the PSMNet uses a smooth L1 loss function to train the whole architecture. The loss function of PSMNet is defined as:
L ( d , d ^ ) = 1 N i = 1 N s m o o t h L 1 ( d i d ^ i ) ,
where
s m o o t h L 1 ( x ) = 0.5 x 2 , i f | x | < 1 | x | 0.5 , o t h e r w i s e .
N is the amount of label pixels. d and d ^ are the ground-truth disparity and predicted disparity, respectively.

3.2. Architecture of 3D Convolution Kernels

In this section, we introduce the architecture of our 3D kernel-based methods. We build all 3D kernels based on their 2D version and fit them to the 3D part of PSMNet according to categories of layers in Figure 1.

3.2.1. 3D MobileNetV1

As shown in Figure 2a, 3D MobileNetV1 [16] decomposes a standard 3 × 3 × 3 convolution kernel into a 3 × 3 × 3 depthwise separable convolution and a 1 × 1 × 1 pointwise convolution. The 3D depthwise separable convolution exploits a series of convolutional filters according to the channel number of the input cost volume and extracts local context in a channel-wise manner. Pointwise convolution walks through the cost volumes, restoring spatial information across different channels.
By isolating local context extraction and channel interaction, MobileNetV1 decreases computational complexity and model size significantly. Unlike most recent CNN architecture, MobileNetV1 excludes ResNet-like residual connections [40] or multi-branch operations, which makes it accessible for channel-wise operations (3D Head in Table 2).

3.2.2. 3D MobileNetV2

Figure 2b shows the 3D MobileNetV2 [17] block. It follows the main idea of MobileNetV1 by building depthwise convolution layers and pointwise convolution layers for reducing computational complexity. It also proposes inverted residual blocks with linear bottlenecks and residual connections. The linear bottlenecks increase the cost volume channels with an expansion factor for solving the problem that high dimensional targeting feature expression often collapses when operating the rectified linear unit (ReLU) activation. The 3D MobileNetV2 block with s t r i d e = 1 (Figure 2b-left) comprising the inverted residual structure helps to construct a deeper model as ResNet [40], while the block with s t r i d e = 2 (Figure 2b-right) keeps excluding the residual connections for a smooth channel-wise operation.

3.2.3. 3D ShuffleNetV1

Compared to other lightweight CNNs, ShuffleNetV1 [18] uses 1 × 1 × 1 pointwise group convolutions (GConv) for computational efficiency. As shown in Figure 2c, the symbolic channel shuffle operation helps to break through the barriers of different groups to build a more robust model. Unlike MobileNetV2, ShuffleNetV1 follows the residual structure to decrease the feature map channels to make it lightweight.
The 3D ShuffleNetV1 block with s t r i d e = 1 (Figure 2c-left) builds the standard ResNet-like residual connections, while the s t r i d e = 2 version constructs the residual connections in another way. As shown in Figure 2c-right, the main branch keeps the structure unchanged, downsamples the feature maps by half with a strided depthwise convolution (DWConv). On the other hand, the shortcut branch utilizes average pooling to halve the feature maps. As the output feature channels of two branches are C, the concatenation results raise feature channels to 2 C .

3.2.4. 3D ShuffleNetV2

Compared to ShuffleNetV1, ShuffleNetV2 [19] changes the 1 × 1 × 1 pointwise group convolution into standard pointwise convolution. In Figure 2d, since the pointwise convolution is not grouped, the channel shuffle operation is placed after the two-branches concatenation to enable information communication between two branches.
The 3D ShuffleNetV2 block with s t r i d e = 1 (Figure 2d-left) shuffles the feature channels and splits all feature maps by two with a channel split operation. Half of them remain untouched with the residual connection. Another half follows a three convolutions scheme without changing the channels. The s t r i d e = 2 version (Figure 2d-right) makes use of all feature maps on each branch. Commonly, the down sampling layers contain channel increases. The main branch (right) accomplishes the channel variation on the first pointwise convolution. However, the identity branch compiles the channel change after the 3 × 3 × 3 depthwise convolution to keep the channel-wise dependency until the pointwise layer.

3.2.5. 3D ECA Blocks

For the outputting cost volume followed by 3D CNNs, we build the 3D ECA [22] blocks for optimizing channel-wise attention. Figure 3 shows that our 3D ECA blocks aggregate cost volumes ( D × H × W ) along the channel with a 3D global average pooling operation. Different from other channel-wise attention modules, ECA blocks do not perform dimensionality reductions when extracting channel-wise information with 3D global average pooling. Then 1D convolution achieves information aggregation on the nearby extracted channel information. After passing a Sigmoid activation function ( σ ), we implement the channel-wise product on the input cost volume to form a 3D channel-wise attention mechanism.

4. Computational Complexity Matrics

Before we dive into the network design pipeline, we introduce the metrics for evaluating the computational complexity as follows:
  • Parameters are the number of trainable neurons in the designed convolutional neural network;
  • Multiply-Add operations(MAdd) describe the accumulated operations when training neural networks. [41] explain the calculation of floating point operations (FLOPs). MAdd is approximately half of FLOPs;
  • Memory Access Cost (MAC) is the amount of allocating computational resource during the training process;
  • Model Size shows the storage size of all trained parameters.

5. Network Design Pipline

In this chapter, we use the 3D kernel-based methods introduced above to replace the standard 3D CNNs in PSMNet, and design a series of comparative experiments to illustrate the impact of different 3D convolution kernels on the performance. Since we focus on discussing the role of 3D convolution kernels in the stereo matching algorithm, we completely follow the original PSMNet on the network layer setting in Table 2.

5.1. Implementation

For all models, we use the same implementation for a fair comparison.

5.1.1. Dataset and Evaluation Metrics

We evaluate our models on Scene Flow [42] and KITTI 2015 [43] datasets:
  • Scene Flow is a large scale dataset with synthetic stereo images. It contains 35,454 training and 4370 testing image pairs with 940 × 540 resolutions. We report the end-point-error (EPE) for evaluations, where EPE shows the average disparity error in pixels;
  • KITTI 2015 contains real street scenes taken by driving a car. It includes 200 training image pairs with ground truth disparity maps collected by LiDAR and 200 other test image pairs without ground truth disparity. The size of the training and test images is 1240 × 376 . We repot D1-all metrics as the official leaderboard.

5.1.2. Implementation Details

We train all models with an Adam optimizer on one NVIDIA RTX 3090 GPU. During the training process, all input images are randomly cropped to 512 × 256 . We first train our models from scratch on the Scene Flow dataset for 20 epochs with a batch size of four. The learning rate is 0.0005 constantly. Then we train the models on KITTI 2015 with Scene Flow pre-trained weights for 2000 epochs. The initial learning rate is 0.0005 and is decreased by half at 400th, 600th and 800th epochs. Since the training dataset of KITTI 2015 only contains 200 input pairs, we perform the training process with 10-fold cross validation [44] for preventing overfitting. In Appendix A, we discuss the difference between normal cross validation and 10-fold cross validation during the re-implementation of the original PSMNet on KITTI 2015 dataset.

5.2. Optimize 3D Convolution Kernels in PSMNet

In Figure 1 and Table 2, we specify all 3D convolution kernels in PSMNet to five categories: 3D Head, 3D Conv, 3D Conv s t r i d e = 2 , 3D Deconv and 3D Out. We replace the traditional CNNs in PSMNet with the 3D convolution kernels in Section 3.2 according to different functions of 3D convolution kernels.

5.2.1. 3D Head

The first 3D convolution layer (3D Head) reduces the number of channels of the cost volume from 64 to 32. Among all 3D convolution kernels, MobileNetV1 builds without residual connection, which is beneficial when operating feature channels. We build 3D MobileNetV1 on the first layer.

5.2.2. 3D Convoltion Layers

For all 3D convolution layers (3D Conv and 3D Conv s t r i d e = 2 ), we used MobileNetV1, MobileNetV2, ShuffleNetV1 and ShuffleNetV2 to replace the original kernels.
In Figure 2c-right, ShuffleNetV1 with s t r i d e = 2 always blocks double the channels to 2 C . However, as shown in Table 2, 3 D S t a c k 3 _ x layer downsamples the cost volume without changing the channels. To make an impartial comparison with PSMNet, as shown in Figure 4, we build a ShuffleNetV1 block with downsampling without changing the number of output channels. In the main branch, we modify the channel of the last pointwise group convolution into C 2 . As for the identity branch, we add a pointwise group convolution followed by the 3D average pooling layer and decrease the channels to C 2 . Therefore, the output cost volumes remain with the same channel number as the PSMNet architecture.
After obtaining the results in Table 3 through comparative experiments, we found that 3D ShuffleNetV2 performs the best among all 3D convolution kernels. Then, we added the 3D ECA block on 3D ShuffleNetV2. The original paper only implements the ECA block on residual connection modules. However, the concatenation in ShuffleNetV2 and the addition operation in the residual connection reflect different logic when operating the channels of cost volumes. We discussed the most suitable position for inserting a 3D ECA block in ShuffleNetV2 in Appendix B. The 3D ECA-ShuffleNetV2 is shown in Figure 5.

5.2.3. 3D Deconvoltion Layers

Three dimensional (3D) transposed convolutions upsample the input cost volumes to twice the size. It first expands the input size by zero padding the cost volume of each channel, and then compiles the standard 3D convolution to introduce learning parameters for more refined textural information. In our implementation, we follow the ShuffleNetV2 as the 3D convolution. However, when inserting 0 values to restore the size, the pointwise convolution will destroy the learned semantic information and textural information. We add upsamping with trilinear interpolation before ShuffleNetV2 for a continuous and rough cost volume, then the ShuffleNetV2 block restores the cost volume with a series of parameterized convolution layers.
Following the architecture of PSMNet in Table 2, we build the 3 D S t a c k 5 _ x layer by simply adding an upsampling layer as we mentioned. Since the channel needs to be reduced to half, we construct the 3 D S t a c k 6 _ x layer as illustrated in Figure 6. The channel split operation divides the input channels into two branches. In the main branch (right), the first pointwise convolution decreases the channels to C 4 instead C 2 . In the identity branch (left), we include a pointwise convolution to reduce the channels to C 4 . Eventually, we get the output cost volume with twice the size and half the channels.

5.2.4. 3D Out

For the last 3D CNN layer ( O u t 2 _ x ), we keep the 3D convolution kernel unchanged for establishing the disparity and textural information.

5.2.5. Network Design Overview

As shown in Table 3, in the first stage, we discuss the influence of MobileNetV1, MobileNetV2, ShuffleNetV1 and ShuffleNetV2 as the basic 3D convolution kernels of the PSMNet on the accuracy and computational complexity. Then we follow the same strategy for designing comparative experiments. We put the MobileNetV1 on the first layer for saving computation, ECA blocks on every 3D convolution kernel for boosting the model accuracy and transposed ShuffleNetV2 for switching all 3D parts of PSMNet to a lightweight method. Eventually, the computational complexity (MAdd) of PSMNet reduced from 256.66 G to 69.03 G 3D convolution kernels implementation. Parameters and model size also decrease from 5.23 M to 3.43 M, and from 21.1 M to 14.1 M, respectively. MAC increases to build more small layers during training. We reduced the computational complexity of the model while keeping the performance almost unchanged. In Appendix C, we parameterize of all models in a layer-manner to emphasize how different 3D convolution kernels change the model computational complexity.

6. Benchmark Results

In Section 5, we built 3D convolution kernels and explored the best combination of 3D kernels with comparative experiments. Due to the number of comparative experiments being relatively large, for saving time, we only train all models to close results without convergence. For benchmarking, we train our model on two NVIDIA V100 for setting the batch size to eight. Since now we have a larger batch size, we double the learning rate for two stages of training. For Scene Flow, the learning is 0.001 constantly. For KITTI 2015, the initial learning rate is 0.001 and is decreased by half at the 400th, 600th and 800th epochs. We perform 10-fold cross validation on the first 1000 epochs. Then we train another 1000 epochs without 10-fold cross validation with a 0.000125 learning rate.
We evaluate our model on Scene Flow and KITTI 2015. In Table 4, our model achieves accurate results on these datasets with a low-cost MAdd in terms of computation. For Scene Flow, our model outperforms the original PSMNet 0.18 on end-point-errors. As for KITTI 2015, Table 5 demonstrates that our model achieves similar results to PSMNet, only taking 26.9% of the MAdd. Meanwhile, our model surpasses the PSMNet significantly in foreground pixels (D1-fg in the table). Figure 7 further visualizes the disparity estimation result on the KITTI 2015 test set.

7. Discussion

In Table 5, compared to other studies, our model contains very little computational complexity. However, the runtime is longer than that of the original PSMNet. As mentioned in [47,48], the reason may be that the cuDNN library does not fully support depthwise convolutions and pointwise convolutions. For the GPU platform of the cuDNN library, the optimization of classic convolutions on end-to-end training is better. So, it will be faster than some lightweight convolutions, although it produces more computation theoretically. We present the runtime of all models in Table 6.

8. Conclusions

In this paper, based on PSMNet as a prototype, we design a series of kernel-based methods aiming for a lightweight and accurate model without modifying the original architecture. By optimizing 3D convolution kernels with corresponding kernel-based methods, our model greatly reduces computational complexity and achieves comparable results to the modern stereo matching algorithms. In future work, as we mentioned in Section 7, we will investigate the implementation of 3D depthwise convolutions and 3D pointwise convolutions on the cuDNN library and improve our model to become faster in training and inference.

Author Contributions

Conceptualization, J.X.; methodology, J.X.; software, J.X.; validation, J.X., D.M. and S.Y.; investigation, J.X.; resources, S.Y.; data curation, J.X. and D.M.; writing—original draft preparation, J.X.; writing—review and editing, J.X., D.M. and S.Y.; visualization, J.X.; supervision, S.Y.; project administration, J.X. and S.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Scene Flow Datasets: FlyingThings3D, Driving, Monkaa in https://lmb.informatik.uni-freiburg.de/resources/datasets/SceneFlowDatasets.en.html (accessed on 16 September 2021); KITTI 2015 in http://www.cvlibs.net/datasets/kitti/eval_scene_flow.php?benchmark=stereo (accessed on 16 September 2021).

Acknowledgments

The authors would like to sincerely thank the editors and reviewers for their time and valuable suggestions on our manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. 10-Fold Cross Validation

Since KITTI 2015 has a small amount of training set, during the re-implementation of PSMNet, we examine the normal training manner and 10-fold cross validation with the same pretrained model on Scene Flow data set.
Table A1. The re-implementation of PSMNet with cross validation and 10-fold cross validation. Both follow the same implementation details in Section 5.1.2. We calculate the average 3-pixel error of all image pairs in the whole training set with the model we submitted for evaluation.
Table A1. The re-implementation of PSMNet with cross validation and 10-fold cross validation. Both follow the same implementation details in Section 5.1.2. We calculate the average 3-pixel error of all image pairs in the whole training set with the model we submitted for evaluation.
MethodAll (%)Noc (%)Loss3-Pixel Error in
D1-BgD1-FgD1-AllD1-BgD1-FgD1-AllTraining Set
Cross Validation2.034.892.511.864.532.300.2910.707
10-fold Cross Validation1.944.762.411.754.422.220.3140.747
As illustrated in Table A1, the cross validation training manner leads to a lower loss and more accurate disparity map on the training set, while the evaluation result is worse. This situation reflects it has an explicit overfitting on the training set.

Appendix B. 3D ECA-ShuffleNetV2 Blocks

In order to achieve the attention mechanism on each channel of the cost volumes, we build the 3D ECA block after concatenation of the main branch and the identity branch. Two potential 3D ECA-ShuffleNetV2 blocks are shown in Figure A1. For saving time, we only train both potential 3D ECA-ShuffleNetV2 blocks on Scene Flow data set for 10 epochs and evaluate the model by training loss.
We show the training performance in Figure A2, the ( b ) architecture slightly outperforms than ( a ) architecture. For model ( a ) , after it learning the channel-wise attention information, the channel shuffle operation disrupts the channel order of cost volumes. Based on the results, we think that the order of channels contains part of the spatial representation ability of the model.
Figure A1. We built ECA block both before and after channel shuffle operation to find a more appropriate architecture of 3D ECA-ShuffleNetV2. (a) before channel shuffle operation. (b) after channel shuffle operation.
Figure A1. We built ECA block both before and after channel shuffle operation to find a more appropriate architecture of 3D ECA-ShuffleNetV2. (a) before channel shuffle operation. (b) after channel shuffle operation.
Sensors 21 06808 g0a1
Figure A2. Training loss on Scene Flow data set with 3D ECA-ShuffeNetV2 (a) & (b).
Figure A2. Training loss on Scene Flow data set with 3D ECA-ShuffeNetV2 (a) & (b).
Sensors 21 06808 g0a2

Appendix C. Parameterize Model Details on Optimizing 3D Convolution Kernels

In order to reflect the influence of the reduction of computational complexity on building different 3D convolution kernels, we show the parameterize model details on Table A2. Since our work mainly focused on the regularization of the cost volume, we omit the model details of feature extraction to one output as 2 D C N N . When only considering the 3D CNNs in the model, we reduce the parameters and MAdd from 1.887 M to 0.085 M ( 95.50 % ) and from 198.47 G to 10.84 G ( 94.53 % ).
Table A2. Parameterize model details on all 3D convolution kernels in our network design pipeline. The numbers in bold denote the modified layers compare to former stages.
Table A2. Parameterize model details on all 3D convolution kernels in our network design pipeline. The numbers in bold denote the modified layers compare to former stages.
MethodBaslineM_V1M_V2S_V1S_V2+M_V1_Head+ECA+Transposed S_V2
LayerPara(M)MAdd(G)Para(M)MAdd(G)Para(M)MAdd(G)Para(M)MAdd(G)Para(M)MAdd(G)Para(M)MAdd(G)Para(M)MAdd(G)Para(M)MAdd(G)
2D CNNs3.34058.1913.34058.1913.34058.1913.34058.1913.34058.1913.34058.1913.34058.1913.34058.191
3DConv00.05521.7810.05521.7810.05521.7810.05521.7810.05521.7810.0041.5600.0041.5730.0041.573
3DConv10.08332.7160.0062.3780.0062.4160.0010.5660.0031.2650.0031.2650.0031.3020.0031.302
3DStack1_10.0552.7270.0030.1530.0081.1830.0010.1170.0050.6420.0050.6420.0050.6450.0050.645
3DStack2_10.1115.4420.0060.2990.0201.0190.0010.0600.0030.1560.0030.1560.0030.1590.0030.159
3DStack3_10.1110.6810.0060.0370.0200.4960.0020.0260.0080.1430.0080.1430.0080.1430.0080.143
3DStack4_10.1110.6810.0060.0370.0200.1270.0010.0070.0030.0190.0030.0190.0030.0200.0030.020
3DStack5_10.1115.4420.1115.4420.1115.4420.1115.4420.1115.4420.1115.4420.1115.4220.0030.159
3DStack6_10.05521.7680.05521.7680.05521.7680.05521.7680.05521.7680.05521.7680.05521.7680.0020.755
3DStack1_20.0552.7270.0030.1530.0081.1830.0010.1170.0050.6420.0050.6420.0050.6450.0050.645
3DStack2_20.1115.4420.0060.2990.0201.0190.0010.0600.0030.1560.0030.1560.0030.1590.0030.159
3DStack3_20.1110.6810.0060.0370.0200.4960.0020.0260.0080.1430.0080.1430.0080.1430.0080.143
3DStack4_20.1110.6810.0060.0370.0200.1270.0010.0070.0030.0190.0030.0190.0030.0200.0030.020
3DStack5_20.1115.4420.1115.4420.1115.4420.1115.4420.1115.4420.1115.4420.1115.4220.0030.159
3DStack6_20.05521.7680.05521.7680.05521.7680.05521.7680.05521.7680.05521.7680.05521.7680.0020.755
3DStack1_30.0552.7270.0030.1530.0081.1830.0010.1170.0050.6420.0050.6420.0050.6450.0050.645
3DStack2_30.1115.4420.0060.2990.0201.0190.0010.0600.0030.1560.0030.1560.0030.1590.0030.159
3DStack3_30.1110.6810.0060.0370.0200.4960.0020.0260.0080.1430.0080.1430.0080.1430.0080.143
3DStack4_30.1110.6810.0060.0370.0200.1270.0010.0070.0030.0190.0030.0190.0030.0200.0030.020
3DStack5_30.1115.4420.1115.4420.1115.4420.1115.4420.1115.4420.1115.4420.1115.4220.0030.159
3DStack6_30.05521.7680.05521.7680.05521.7680.05521.7680.05521.7680.05521.7680.05521.7680.0020.755
Out1_10.02810.910.0020.7930.0020.8050.00040.1890.0010.4220.0010.4220.0010.4340.0010.434
Out2_10.0010.3400.0010.3400.0010.3400.0010.3400.0010.3400.0010.3400.0010.3400.0010.340
Out1_20.02810.910.0020.7930.0020.8050.00040.1890.0010.4220.0010.4220.0010.4340.0010.434
Out2_20.0010.3400.0010.3400.0010.3400.0010.3400.0010.3400.0010.3400.0010.3400.0010.340
Out1_30.02810.910.0020.7930.0020.8050.00040.1890.0010.4220.0010.4220.0010.4340.0010.434
Out2_30.0010.3400.0010.3400.0010.3400.0010.3400.0010.3400.0010.3400.0010.3400.0010.340
3D CNNs1.887198.4700.631110.7660.772117.7370.571106.1940.619109.8420.56889.6210.56889.7280.08510.840
Full Model5.227256.6613.971168.9574.112175.9283.912164.3853.959168.0333.908147.8123.908147.9193.42569.031

References

  1. Zenati, N.; Zerhouni, N. Dense Stereo Matching with Application to Augmented Reality. In Proceedings of the IEEE International Conference on Signal Processing and Communications, Dubai, United Arab Emirates, 24–27 November 2007; pp. 1503–1506. [Google Scholar] [CrossRef]
  2. Noh, Z.; S, M.; Pan, Z. A Review on Augmented Reality for Virtual Heritage System. In Proceedings of the Springer International Conference on Technologies for E-Learning and Digital Entertainment, Banff, AB, Canada, 9–11 August 2009; Volume 5670. [Google Scholar] [CrossRef]
  3. Häne, C.; Sattler, T.; Pollefeys, M. Obstacle detection for self-driving cars using only monocular cameras and wheel odometry. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Hamburg, Germany, 28 September–2 October 2015; pp. 5101–5108. [Google Scholar]
  4. Nalpantidis, L.; Sirakoulis, G.C.; Gasteratos, A. Review of stereo matching algorithms for 3D vision. In Proceedings of the 16th International Symposium on Measurement and Control in Robotic, Warsaw, Poland, 21–23 June 2007. [Google Scholar]
  5. Samadi, M.; Othman, M.F. A New Fast and Robust Stereo Matching Algorithm for Robotic Systems. In Proceedings of the 9th International Conference on Computing and InformationTechnology (IC2IT2013), Bangkok, Thailand, 9–10 May 2013. [Google Scholar]
  6. Zbontar, J.; LeCun, Y. Computing the stereo matching cost with a convolutional neural network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 1592–1599. [Google Scholar]
  7. Scharstein, D.; Szeliski, R. A Taxonomy and Evaluation of Dense Two-Frame Stereo Correspondence Algorithms. Int. J. Comput. Vis. 2002, 47, 7–42. [Google Scholar] [CrossRef]
  8. Zhang, L.; Seitz, S.M. Estimating optimal parameters for MRF stereo from a single image pair. IEEE Trans. Pattern Anal. Mach. Intell. (TPAMI) 2007, 29, 331–342. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  9. Scharstein, D.; Pal, C. Learning Conditional Random Fields for Stereo. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Minneapolis, MN, USA, 17–22 June 2007; pp. 1–8. [Google Scholar]
  10. Haeusler, R.; Nair, R.; Kondermann, D. Ensemble Learning for Confidence Measures in Stereo Vision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Portland, OR, USA, 23–28 June 2013; pp. 305–312. [Google Scholar]
  11. Kendall, A.; Martirosyan, H.; Dasgupta, S.; Henry, P.; Kennedy, R.; Bachrach, A.; Bry, A. End-to-end learning of geometry and context for deep stereo regression. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 66–75. [Google Scholar]
  12. Chang, J.-R.; Chen, Y.-S. Pyramid Stereo Matching Network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018; pp. 5410–5418. [Google Scholar]
  13. Iandola, F.N.; Moskewicz, M.W.; Ashraf, K.; Han, S.; Dally, W.J.; Keutzer, K. SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and <0.5MB model size. arXiv 2016, arXiv:1602.07360. [Google Scholar]
  14. Krizhevsky, A.; Sutskever, I.; Hinton, G. ImageNet Classification with Deep Convolutional Neural Networks. In Proceedings of the Conference on Neural Information Processing Systems (NeurIPS), Lake Tahoe, NV, USA, 3–6 December 2012. [Google Scholar]
  15. Chollet, F. Xception: Deep Learning With Depthwise Separable Convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 22–25 July 2017; pp. 1251–1258. [Google Scholar]
  16. Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv 2017, arXiv:1704.04861. [Google Scholar]
  17. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.-C. Mobilenetv2: Inverted residuals and linear bottlenecks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018; pp. 4510–4520. [Google Scholar]
  18. Zhang, X.; Zhou, X.; Lin, M.; Sun, J. Shufflenet: An extremely efficient convolutional neural network for mobile devices. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018; pp. 6848–6856. [Google Scholar]
  19. Ma, N.; Zhang, X.; Zheng, H.T.; Sun, J. ShuffleNet V2: Practical Guidelines for Efficient CNN Architecture Desig. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 116–131. [Google Scholar]
  20. Hu, J.; Shen, L.; Sun, G. Squeeze-and-Excitation Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018; pp. 7132–7141. [Google Scholar]
  21. Li, X.; Wang, W.; Hu, X.; Yang, J. Selective Kernel Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 510–519. [Google Scholar]
  22. Wang, Q.; Wu, B.; Zhu, P.; Li, P.; Zuo, W.; Hu, Q. ECA-Net: Efficient Channel Attention for Deep Convolutional Neural Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 11534–11542. [Google Scholar]
  23. Zhou, K.; Meng, X.; Cheng, B. Review of Stereo Matching Algorithms Based on Deep Learning. Computational Intelligence and Neuroscience 2020, 2020, 8562323. [Google Scholar] [CrossRef] [PubMed]
  24. Guo, X.; Yang, K.; Yang, W.; Wang, X.; Li, H. Group-wise correlation stereo network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 3273–3282. [Google Scholar]
  25. Khamis, S.; Fanello, S.; Rhemann, C.; Kowdle, A.; Valentin, J.; Izadi, S. StereoNet: Guided Hierarchical Refinement for Real-Time Edge-Aware Depth Prediction. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 573–590. [Google Scholar]
  26. Zhang, F.; Prisacariu, V.; Yang, R.; Torr, P.H. GA-Net: Guided Aggregation Net for End-To-End Stereo Matching. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 185–194. [Google Scholar]
  27. Zhu, C.; Chang, Y. Hierarchical Guided-Image-Filtering for Efficient Stereo Matching. Appl. Sci. 2019, 15, 3122. [Google Scholar] [CrossRef] [Green Version]
  28. Zhang, Y.; Chen, Y.; Bai, X.; Yu, S.; Yu, K.; Li, Z.; Yang, K. Adaptive Unimodal Cost Volume Filtering for Deep Stereo Matching. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; pp. 12926–12934. [Google Scholar]
  29. Duggal, S.; Wang, S.; Ma, W.; Hu, R.; Urtasun, R. DeepPruner: Learning Efficient Stereo Matching via Differentiable PatchMatch. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Seoul, Korea, 27–28 October 2019; pp. 4384–4393. [Google Scholar]
  30. Barnes, C.; Shechtman, E.; Finkelstein, A.; Goldman, D.B. PatchMatch: A Randomized Correspondence Algorithm for Structural Image Editing. Acm Trans. Graph. (Proc. Siggraph) 2009, 28, 24. [Google Scholar] [CrossRef]
  31. Xu, H.; Zhang, J. AANet: Adaptive Aggregation Network for Efficient Stereo Matching. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 1959–1968. [Google Scholar]
  32. Gu, X.; Fan, Z.; Zhu, S.; Dai, Z.; Tan, F.; Tan, P. Cascade Cost Volume for High-Resolution Multi-View Stereo and Stereo Matching. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 2495–2504. [Google Scholar]
  33. Shen, Z.; Dai, Y.; Rao, Z. CFNet: Cascade and Fused Cost Volume for Robust Stereo Matching. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Virtual, 19–25 June 2021; pp. 13906–13915. [Google Scholar]
  34. Wei, M.; Zhu, M.; Wu, Y.; Sun, J.; Wang, J.; Liu, C. A Fast Stereo Matching Network with Multi-Cross Attention. Sensors 2021, 18, 6016. [Google Scholar] [CrossRef] [PubMed]
  35. Huang, Z.; Gu, J.; Li, J.; Yu, X. A stereo matching algorithm based on the improved PSMNet. PLoS ONE 2021, 25, e0251657. [Google Scholar] [CrossRef]
  36. Xie, S.; Girshick, R.; Dollár, P.; Tu, Z.; He, K. Aggregated Residual Transformations for Deep Neural Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 5987–5995. [Google Scholar]
  37. Cheng, L.; Papandreou, G.; Schroff, F.; Adam, H. Rethinking Atrous Convolution for Semantic Image Segmentation. arXiv 1706, arXiv:1706.05587. [Google Scholar]
  38. Jia, X.; Chen, W.; Liang, Z.; Luo, X.; Wu, M.; Li, C.; He, Y.; Tan, Y.; Huang, L. A Joint 2D-3D Complementary Network for Stereo Matching. Sensors 2021, 21, 1430. [Google Scholar] [CrossRef] [PubMed]
  39. Vasileiadis, M.; Bouganis, C.-S.; Stavropoulos, G.; Tzovaras, D. Optimising 3D-CNN Design towards Human Pose Estimation on Low Power Devices. In Proceedings of the British Machine Vision Conference (BMVC), Cardiff, UK, 9–12 September 2019. [Google Scholar]
  40. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  41. Molchanov, P.; Tyree, S.; Karras, T.; Aila, T.; Kautz, J. Pruning Convolutional Neural Networks for Resource Efficient Inference. In Proceedings of the International Conference on Learning Representations (ICLR), Toulon, France, 24–26 April 2017; pp. 1–17. [Google Scholar]
  42. Mayer, N.; Ilg, E.; Hausser, P.; Fischer, P.; Cremers, D.; Dosovitskiy, A.; Brox, T. A Large Dataset to Train Convolutional Networks for Disparity, Optical Flow, and Scene Flow Estimation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 4040–4048. [Google Scholar]
  43. Menze, M.; Geiger, A. Object scene flow for autonomous vehicles. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 3061–3070. [Google Scholar]
  44. Ng, A.Y. Preventing “overfitting” of cross-validation data. Machine Learning. In Proceedings of the International Conference on Machine Learning (ICML), Nashville, TN, USA, 8–12 July 1997; pp. 245–253. [Google Scholar]
  45. Tosi, F.; Liao, Y.; Schmitt, C.; Geiger, A. SMD-Nets: Stereo Mixture Density Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), virtual, 19–25 June 2021; pp. 8942–8952. [Google Scholar]
  46. Yee, K.; Chakrabarti, A. Fast Deep Stereo with 2D Convolutional Processing of Cost Signatures. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Snowmass Village, CO, USA, 1–5 March 2020; pp. 183–191. [Google Scholar]
  47. Qin, Z.; Zhang, Z.; Li, D.; Zhang, Y.; Peng, Y. Diagonalwise Refactorization: An Efficient Training Method for Depthwise Convolutions. In Proceedings of the 2018 International Joint Conference on Neural Networks (IJCNN), Rio de Janeiro, Brazil, 8–13 July 2018. [Google Scholar]
  48. Lu, G.; Zhang, W.; Wang, Z. Optimizing Depthwise Separable Convolution Operations on GPUs. IEEE Trans. Parallel Distrib. Syst. 2021, 33, 70–87. [Google Scholar] [CrossRef]
Figure 1. Architecture of PSMNet: We specified all 3D CNN kernels with different colors representing different interactions on the size and channel of cost volumes.
Figure 1. Architecture of PSMNet: We specified all 3D CNN kernels with different colors representing different interactions on the size and channel of cost volumes.
Sensors 21 06808 g001
Figure 2. The architecture of 3D lightweight CNN kernels. (a) 3D MobileNetV1. (b) 3D MobileNetV2. (c) 3D ShuffleNetV1. (d) 3D ShuffleNetV2. Each module has the same effect as a 3 × 3 × 3 convolution kernel.
Figure 2. The architecture of 3D lightweight CNN kernels. (a) 3D MobileNetV1. (b) 3D MobileNetV2. (c) 3D ShuffleNetV1. (d) 3D ShuffleNetV2. Each module has the same effect as a 3 × 3 × 3 convolution kernel.
Sensors 21 06808 g002
Figure 3. The architecture of 3D ECA blocks. We build the 3D ECA blocks for exploring channel-wise attention during cost volume regularization. In the original paper, the kernel size k of Conv1D is adaptively determined according to channel dimension C. In the 3D part of PSMNet, we set all kernel sizes k to 3 in terms of the relatively small feature channels.
Figure 3. The architecture of 3D ECA blocks. We build the 3D ECA blocks for exploring channel-wise attention during cost volume regularization. In the original paper, the kernel size k of Conv1D is adaptively determined according to channel dimension C. In the 3D part of PSMNet, we set all kernel sizes k to 3 in terms of the relatively small feature channels.
Sensors 21 06808 g003
Figure 4. The implementation of 3D ShuffleNetV1 with s t r i d e = 2 at 3 D S t a c k 3 _ x layer. Following the PSMNet, we downsample the cost volume by half without changing the channels.
Figure 4. The implementation of 3D ShuffleNetV1 with s t r i d e = 2 at 3 D S t a c k 3 _ x layer. Following the PSMNet, we downsample the cost volume by half without changing the channels.
Sensors 21 06808 g004
Figure 5. We build the 3D ECA blocks after the channel shuffle operations on 3D ShuffleNet V2. (a) 3D ECA-ShuffleNetV2 with s t r i d e = 1 . (b) 3D ECA-ShuffleNetV2 with s t r i d e = 2 .
Figure 5. We build the 3D ECA blocks after the channel shuffle operations on 3D ShuffleNet V2. (a) 3D ECA-ShuffleNetV2 with s t r i d e = 1 . (b) 3D ECA-ShuffleNetV2 with s t r i d e = 2 .
Sensors 21 06808 g005
Figure 6. The trilinear interpolation upsamples the input to roughly twice the size of cost volume, then the 3D ECA-ShuffleNetV2 refines the resolution of cost volume with parameterized layers. (a) 3D Transposed ECA-ShuffleNetV2 at 3 D S t a c k 5 _ x layer. (b) 3D Transposed ECA-ShuffleNetV2 at 3 D S t a c k 6 _ x layer.
Figure 6. The trilinear interpolation upsamples the input to roughly twice the size of cost volume, then the 3D ECA-ShuffleNetV2 refines the resolution of cost volume with parameterized layers. (a) 3D Transposed ECA-ShuffleNetV2 at 3 D S t a c k 5 _ x layer. (b) 3D Transposed ECA-ShuffleNetV2 at 3 D S t a c k 6 _ x layer.
Sensors 21 06808 g006
Figure 7. Visualization of prediction error on KITTI test set (red and yellow pixels denote error disparities). The red boxes denote foreground regions and green boxes denote background regions. (a) Left image. (b) Our model. (c) PSMNet. (d) AANet.
Figure 7. Visualization of prediction error on KITTI test set (red and yellow pixels denote error disparities). The red boxes denote foreground regions and green boxes denote background regions. (a) Left image. (b) Our model. (c) PSMNet. (d) AANet.
Sensors 21 06808 g007
Table 1. Recent studies on stereo matching. We clarify where their modifications target stereo matching architecture.
Table 1. Recent studies on stereo matching. We clarify where their modifications target stereo matching architecture.
NameMain FeatureTargeting Part
GwcNet [24]Performs a group-wise correlation to generateCost volume aggregation &
a multi-feature cost volume.Cost volume regularization
StereoNet [25]Utilizes more downsampling to form a low-resolution cost volume,2D Siamese CNNs &
which lead a great improvement in speed.Cost volume regularization
GANet [26]Introduces a semi-global guided aggregation layer andCost volume regularization
a local guided aggregation layer
for replacing 3D convolution layer in 3D encoder-decoder.
Zhu et al. [27]Apply edge-preserving guided-Image-filtering (GIF) atCost volume regularization
different resolutions on multi-scale stereo matching.
AcfNet [28]Includes the ground truth cost volume andCost volume regularization
confidence map for intermediate supervision.
DeepPruner [29]Aggregates a sparse cost volumeCost volume aggregation
with a differentiable PatchMatch [30] module.
AANet [31]Presents a intra-scale aggregation moduleCost volume regularization
for replacing 3D convolution layer & cross-scale aggregation for
integrating multi-scale cost volume.
CSN [32] & CFNet [33]Implements the architecture in a coarse-to-fine manner.Cost volume aggregation &
Cost volume regularization
Wei et al. [34]Improve StereoNet[25] with edge-guided refinement &2D Siamese CNNs &
multi-cross attention module on multi-level cost volumes.Cost volume regularization
Huang et al. [35]Implement ResNetXt [36] and Atrous Spatial Pyramid Pooling2D Siamese CNNs
(ASSP) [37] on 2D CNNs.
JDCNet [38]Use the 2D stereo encoder-decoder to generate a disparity range2D Siamese CNNs &
for guiding 3D aggregation network.Cost volume aggregation
Table 2. Parameters of PSMNet architecture. Batch normalization and ReLU layers are used except the summation operations. H and W denote the height and width of an input pair. D means the disparities of the stereo input images.
Table 2. Parameters of PSMNet architecture. Batch normalization and ReLU layers are used except the summation operations. H and W denote the height and width of an input pair. D means the disparities of the stereo input images.
2D Part: Feature Extraction 3D Part: Cost Volume Optimization
Layer NameSettingOutput DimensionLayer NameSettingOutput Dimension
Stereo input- H × W × 3 × 2 Concat left and right feature maps- 1 4 D × 1 4 H × 1 4 W × 64
2D CNNs 3D Stacked Hourglass module
Conv0_x 3 × 3 , 32 × 3 1 2 H × 1 2 W × 32 3DConv0(3D Head) 3 × 3 × 3 , 32 1 4 D × 1 4 H × 1 4 W × 32
Conv1_x 3 × 3 , 32 3 × 3 , 32 × 3 1 2 H × 1 2 W × 32 3DConv1(3D Conv) 3 × 3 × 3 , 64 × 3 1 4 D × 1 4 H × 1 4 W × 32
Conv2_x 3 × 3 , 64 3 × 3 , 64 × 16 1 4 H × 1 4 W × 64 3DStack1_x(3D Conv, s t r i d e = 2 ) 3 × 3 × 3 , 64 , s t r i d e = 2 1 8 D × 1 8 H × 1 8 W × 64
Conv3_x 3 × 3 , 128 3 × 3 , 128 × 3 1 4 H × 1 4 W × 128 3DStack2_x(3D Conv) 3 × 3 × 3 , 64 1 8 D × 1 8 H × 1 8 W × 64
Conv4_x 3 × 3 , 128 3 × 3 , 128 × 3 , d i l a = 2 1 4 H × 1 4 W × 128 3DStack3_x(3D Conv, s t r i d e = 2 ) 3 × 3 × 3 , 64 , s t r i d e = 2 1 16 D × 1 16 H × 1 16 W × 64
Spatial Pyramid Pooling (SPP) module 3DStack4_x(3D Conv) 3 × 3 × 3 , 64 1 16 D × 1 16 H × 1 16 W × 64
Branch_1 64 × 64 , 128, Avg_pooling 1 4 H × 1 4 W × 32 3DStack5_x(3D Deconv)ConvTranspose3d 3 × 3 × 3 , 64 1 8 D × 1 8 H × 1 8 W × 64
1 × 1 , 32, Conv
Upsample, Biliner interpolation
Branch_2 32 × 32 , 128, Avg_pooling 1 4 H × 1 4 W × 32 3DStack6_x(3D Deconv)ConvTranspose3d 3 × 3 × 3 , 32 1 4 D × 1 4 H × 1 4 W × 32
1 × 1 , 32, Conv
Upsample, Biliner interpolation
Branch_3 16 × 16 , 128, Avg_pooling 1 4 H × 1 4 W × 32 Out1_x(3D Conv) 3 × 3 × 3 , 32 1 4 D × 1 4 H × 1 4 W × 32
1 × 1 , 32, Conv
Upsample, Biliner interpolation
Branch_4 8 × 8 , 128, Avg_pooling 1 4 H × 1 4 W × 32 Out2_x(3D Out) 3 × 3 × 3 , 1 1 4 D × 1 4 H × 1 4 W × 1
1 × 1 , 32, Conv
Upsample, Biliner interpolation
Concat[Conv2_16,Conv4_3,Branch_1,Branch_2,Branch_3,Branch_4] 1 4 H × 1 4 W × 320 UpsamplingTrilinear interpolation D × H × W
Fusion 3 × 3 , 128 1 4 H × 1 4 W × 32 -Disparity regression H × W
1 × 1 , 32,
Table 3. Experiments for 3D convolution kernels on PSMNet. We calculate the MAdd with 512 × 256 resolution as input. EPE is the end-point-error on the Scene Flow dataset. D1-all denotes the evaluation of the KITTI 2015 leaderboard. MobelNetV2* denotes implementation of the kernel with e x p a n s i o n r a t i o = 2 considering the MAC of the model. At each stage, we select the best 3D convolution kernel (grey line) and optimize it on the model.
Table 3. Experiments for 3D convolution kernels on PSMNet. We calculate the MAdd with 512 × 256 resolution as input. EPE is the end-point-error on the Scene Flow dataset. D1-all denotes the evaluation of the KITTI 2015 leaderboard. MobelNetV2* denotes implementation of the kernel with e x p a n s i o n r a t i o = 2 considering the MAC of the model. At each stage, we select the best 3D convolution kernel (grey line) and optimize it on the model.
3D Conv KernelEPED1-AllParameters (Millon)MAdd (Gb)MAC (Mb)Model Size (Mb)
3D Convolution Layers
Baseline(re-implemented)1.1232.415.23256.66389521.1
MobileNetV11.2522.683.97168.96542116.2
MobileNetV2*1.2152.604.11175.93609716.8
ShuffleNetV11.2952.883.91164.39465116.0
ShuffleNetV21.2412.623.96168.03507116.2
3D Head
3D CNN1.2412.623.96168.03507116.2
MobileNetV1 (Head)1.2332.613.91147.81541116.0
3D ECA Blocks
Without ECA Blocks1.2332.613.91147.81541116.0
ECA Blocks1.1602.503.91147.92583516.0
3D Deconvoltion Layers
3D Transposed CNN1.1602.503.91147.92583516.0
Transposed ShuffleNetV21.1242.433.4369.03677114.1
Table 4. Evaluation results on the Scene Flow dataset. Our model is competitive with other top-performing models.
Table 4. Evaluation results on the Scene Flow dataset. Our model is competitive with other top-performing models.
MethodOursBaseline [16]GC-Net [11]GANet [26]DeepPruner-Best [29]DispNetC [42]StereoNet [25]JDCNet [38]
EPE0.911.092.510.840.861.681.100.83
Table 5. Evaluation results on the KITTI 2015 dataset. The first seven methods are accurate methods (included the baseline). The other five are considered fast methods. Our model not only achieves results comparable to those of some accurate methods but also requires significantly less computational complexity. We only calculated the MAdd of some representative models.
Table 5. Evaluation results on the KITTI 2015 dataset. The first seven methods are accurate methods (included the baseline). The other five are considered fast methods. Our model not only achieves results comparable to those of some accurate methods but also requires significantly less computational complexity. We only calculated the MAdd of some representative models.
MethodAll (%)Noc (%)Runtime(s)MAdd (G)
D1-BgD1-FgD1-AllD1-BgD1-FgD1-All
Baseline [12]1.864.622.321.714.312.140.41256.66
PSMNet-lite (Ours)1.914.562.351.754.062.130.6369.03
MC-CNN [6]2.898.883.892.487.643.3367-
GC-Net [11]2.216.162.872.025.582.610.9733.36
GwcNet [24]1.743.932.111.613.491.920.32247.6
DeepPruner-Fast [29]1.873.562.151.713.181.950.18-
GANet-15 [26]1.553.821.931.403.371.730.36-
CSN [32]1.594.032.001.433.551.780.6-
SMD-Net [45]1.694.012.081.543.701.890.41-
StereoNet [25]4.307.454.83---0.01547.08
DispNetC [42]4.324.414.344.113.724.050.03-
DeepPruner-Best [29]2.323.912.592.133.432.350.06-
AANet [31]1.995.392.551.804.932.320.062-
Fast DS-CS [46]2.834.313.082.533.742.730.02-
JDCNet [38]1.914.472.331.733.862.080.079-
Table 6. Inference time of all models in the network design pipeline. We use one NVIDIA RTX 3090 for all implementations. “+ MobileNetV1” replaces the 3 D C o n v 0 with the 3D MobileNetV1 module. “+ ECA block” add 3D ECA block on all 3D kernels. “+ Transposed ShuffleNetV2” replaces the 3 D S t a c k 5 _ x and 3 D S t a c k 6 _ x deconvolution layers with 3D Transposed ECA-ShuffleNetV2.
Table 6. Inference time of all models in the network design pipeline. We use one NVIDIA RTX 3090 for all implementations. “+ MobileNetV1” replaces the 3 D C o n v 0 with the 3D MobileNetV1 module. “+ ECA block” add 3D ECA block on all 3D kernels. “+ Transposed ShuffleNetV2” replaces the 3 D S t a c k 5 _ x and 3 D S t a c k 6 _ x deconvolution layers with 3D Transposed ECA-ShuffleNetV2.
MethodMobileNetV1MobileNetV2ShuffleNetV1ShuffleNetV2+ MobileNetV1+ ECA Blocks+ Transposed
HeadShuffleNetV2
Runtime (s)0.480.530.350.400.410.600.63
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Xiao, J.; Ma, D.; Yamane, S. Optimizing 3D Convolution Kernels on Stereo Matching for Resource Efficient Computations. Sensors 2021, 21, 6808. https://doi.org/10.3390/s21206808

AMA Style

Xiao J, Ma D, Yamane S. Optimizing 3D Convolution Kernels on Stereo Matching for Resource Efficient Computations. Sensors. 2021; 21(20):6808. https://doi.org/10.3390/s21206808

Chicago/Turabian Style

Xiao, Jianqiang, Dianbo Ma, and Satoshi Yamane. 2021. "Optimizing 3D Convolution Kernels on Stereo Matching for Resource Efficient Computations" Sensors 21, no. 20: 6808. https://doi.org/10.3390/s21206808

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop