Next Article in Journal
On the Synergy between Nonconvex Extensions of the Tensor Nuclear Norm for Tensor Recovery
Previous Article in Journal
Electromyogram in Cigarette Smoking Activity Recognition
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

3D Object Detection Using Frustums and Attention Modules for Images and Point Clouds

Department of Electrical Engineering, Hanyang University, Ansan 15588, Korea
*
Author to whom correspondence should be addressed.
Signals 2021, 2(1), 98-107; https://doi.org/10.3390/signals2010009
Submission received: 21 September 2020 / Revised: 23 December 2020 / Accepted: 6 January 2021 / Published: 12 February 2021

Abstract

Three-dimensional (3D) object detection is essential in autonomous driving. Three-dimensional (3D) Lidar sensor can capture three-dimensional objects, such as vehicles, cycles, pedestrians, and other objects on the road. Although Lidar can generate point clouds in 3D space, it still lacks the fine resolution of 2D information. Therefore, Lidar and camera fusion has gradually become a practical method for 3D object detection. Previous strategies focused on the extraction of voxel points and the fusion of feature maps. However, the biggest challenge is in extracting enough edge information to detect small objects. To solve this problem, we found that attention modules are beneficial in detecting small objects. In this work, we developed Frustum ConvNet and attention modules for the fusion of images from a camera and point clouds from a Lidar. Multilayer Perceptron (MLP) and tanh activation functions were used in the attention modules. Furthermore, the attention modules were designed on PointNet to perform multilayer edge detection for 3D object detection. Compared with a previous well-known method, Frustum ConvNet, our method achieved competitive results, with an improvement of 0.27%, 0.43%, and 0.36% in Average Precision (AP) for 3D object detection in easy, moderate, and hard cases, respectively, and an improvement of 0.21%, 0.27%, and 0.01% in AP for Bird’s Eye View (BEV) object detection in easy, moderate, and hard cases, respectively, on the KITTI detection benchmarks. Our method also obtained the best results in four cases in AP on the indoor SUN-RGBD dataset for 3D object detection.
Keywords: 3D vision; attention module; fusion; point cloud; vehicle detection 3D vision; attention module; fusion; point cloud; vehicle detection

Share and Cite

MDPI and ACS Style

Li, Y.; Xie, H.; Shin, H. 3D Object Detection Using Frustums and Attention Modules for Images and Point Clouds. Signals 2021, 2, 98-107. https://doi.org/10.3390/signals2010009

AMA Style

Li Y, Xie H, Shin H. 3D Object Detection Using Frustums and Attention Modules for Images and Point Clouds. Signals. 2021; 2(1):98-107. https://doi.org/10.3390/signals2010009

Chicago/Turabian Style

Li, Yiran, Han Xie, and Hyunchul Shin. 2021. "3D Object Detection Using Frustums and Attention Modules for Images and Point Clouds" Signals 2, no. 1: 98-107. https://doi.org/10.3390/signals2010009

APA Style

Li, Y., Xie, H., & Shin, H. (2021). 3D Object Detection Using Frustums and Attention Modules for Images and Point Clouds. Signals, 2(1), 98-107. https://doi.org/10.3390/signals2010009

Article Metrics

Back to TopTop