Next Article in Journal
A Narrow Optical Pulse Emitter Based on LED: NOPELED
Previous Article in Journal
Learning Assurance Analysis for Further Certification Process of Machine Learning Techniques: Case-Study Air Traffic Conflict Detection Predictor
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

A Comprehensive Review on Lane Marking Detection Using Deep Neural Networks

1
Faculty of Engineering and Technology, Multimedia University, Melaka 75450, Malaysia
2
Department of Computer Science, Wayne State University, Detroit, MI 48202, USA
3
Department of Computer Science and Engineering, Feni University, Feni 3900, Bangladesh
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(19), 7682; https://doi.org/10.3390/s22197682
Submission received: 18 August 2022 / Revised: 22 September 2022 / Accepted: 3 October 2022 / Published: 10 October 2022
(This article belongs to the Section Vehicular Sensing)

Abstract

:
Lane marking recognition is one of the most crucial features for automotive vehicles as it is one of the most fundamental requirements of all the autonomy features of Advanced Driver Assistance Systems (ADAS). Researchers have recently made promising improvements in the application of Lane Marking Detection (LMD). This research article has taken the initiative to review lane marking detection, mainly using deep learning techniques. This paper initially discusses the introduction of lane marking detection approaches using deep neural networks and conventional techniques. Lane marking detection frameworks can be categorized into single-stage and two-stage architectures. This paper elaborates on the network’s architecture and the loss function for improving the performance based on the categories. The network’s architecture is divided into object detection, classification, and segmentation, and each is discussed, including their contributions and limitations. There is also a brief indication of the simplification and optimization of the network for simplifying the architecture. Additionally, comparative performance results with a visualization of the final output of five existing techniques is elaborated. Finally, this review is concluded by pointing to particular challenges in lane marking detection, such as generalization problems and computational complexity. There is also a brief future direction for solving the issues, for instance, efficient neural network, Meta, and unsupervised learning.

1. Introduction

Autonomous driving has become a hotspot research topic as the intelligent transport system and environmental perception improves daily. LMD is one of the significant parts of the environmental perception system, where many efforts have been made in the previous decade. Nevertheless, developing an efficient lane detection framework under different environmental circumstances is a highly challenging task because it has many dependencies that may influence the framework’s final output.
Various preprocessing techniques have a significant role in lane marking detection systems, mostly dependent on heuristic features. Distinct types of filters such as Finite Impulse Response (FIR) [1], Gaussian [2], and mean and median [3] are used to remove the noise from the input dataset. Duan et al. [4] introduced threshold segmentation to deal with the variation in illumination. Additionally, PLSF [5] and Otsu [6] are also applied for the same region. There are different Regions of Interest (ROI) that are examined to avoid redundancy, such as vanishing point-based ROI [7], adaptive ROI [8], and Fixed-size ROI [9]. An essential preprocessing tool to enhance the quality of lane marks is colour conversion, such as the RGB to HSV colour model.
There are many algorithms applied to extract lane features, especially for straight lanes, for instance, Hough [10], Canny [8], Sobel [9], and FIR filter [11]. Catmull–Rom spline [12], clothoid curve [13], parabolic [14], and cubic B-spline [2] are applied for curved lanes. A few other techniques are used under complex conditions, such as image enhancement [15] and wavelet analysis [16].
DNN (Deep Neural Network) has become one of the most promising computer vision techniques since AlexNet won the ILSVRC challenge in 2012. These deep learning techniques have shown promising performances in various fields of research. Recently, various efficient deep learning approaches have been examined for lane marking detection. From the beginning, Convolution Neuron Network (CNN) [17,18] to the GAN-based method [19] and segmentation process [20] have obtained efficient results on LMD. Additionally, DAGMapper [21] and attention map [22] have been applied to understand the structural features of the lanes. Though these techniques have obtained auspicious results, LMD is still challenging for its lack of generalization capability. For instance, a trained model in a particular scenario, such as daytime, may obtain poor results in other environmental scenarios, such as nighttime.
This article provides an efficient, comprehensive review of LMD using different deep neural networks. The manuscript provides a review of the complete process of lane marking detection (LMD) using deep learning techniques, considering the sequential process. It provides a clear indication of preprocessing and post-processing approaches for lane marking detection. In addition, the manuscript provides an optimization process to improve the algorithm as well as to remove the post-processing steps. The loss function is an important part of LMD, and it is categorically discussed, broken into classification, regression, and adversarial training to make it these categories easy to understand. Deep learning algorithms for LMD are explained in three major categories: object detection, classification, and segmentation of lanes, to cover all aspects of this field with proper objectives, limitations, suggested improvements, and structures. The summary table (Table 1 and Table 2) provides the deep learning algorithms for LMD with achievements, results and constraints to indicate their goals and barriers. More importantly, a discussion shows a comparative result with outcome figures by training and testing the models with the Tusimple dataset. Finally, a future direction is provided to give a probable option for improving the LMD techniques. The remaining sections of the review article are organized as follows: Section 2 outlines distinct deep learning techniques (including preprocessing, loss function, cluttering, and preprocessing) for LMD. Section 3 describes the comparative result of some experimental results. Finally, Section 4 gives the conclusion and future thoughts on LMD techniques.

2. LMD Using DNN

The existing lane marking detection approaches can be classified into two major categories: single-stage and two-stage [23]. The initial segment of the two-stage frameworks extracts the heuristic recognition and deep learning-based lane features. In contrast, the second segment refers to the post-processing steps, which may include fitting, clustering, or interfacing. However, the single-stage lane detection approach provides final results directly from the input stage, including post-processing and cluster results. The LMD using the deep neural network has been discussed from four perspectives: preprocessing, network architecture, network loss functions, and post-processing.

2.1. Pre-Processing

ROI cropping is applied to remove the irrelevant information from the input dataset in the traditional and initial parts of the deep learning approaches. Consequently, it reduces the computational complexity and increases the running speed of the framework. As the lane markings are visualized on the lower part of the image frames, the clipped portion refers to the frames’ upper or sky part. Thus, it reduces the computational complexity by around 30% [23].
Some advanced techniques, such as meta-learning, can be examined to ameliorate the generalization of the CNN method. It can also be improved by diversifying the training dataset. The augmentation technique has a significant role in diversifying and increasing the number of data in the image dataset. In this process, data can be cropped, rotated, brightened, and mirrored to assort the training dataset shown in Figure 1 as a reference.

2.2. Network Architecture of LMD

There are many strategies to detect the LMD using a deep learning network, though these strategies can be categorized based on defining the LMD task. Therefore, these techniques can be classified as object detection, classification, and segmentation of lanes. Every feature point on lane segments is labeled, and detects the lanes as an object by the regression coordinates. In comparison, lane position is determined by combining the prior information in the classification techniques. On the contrary, background and lane pixels are labelled as distinct classes and detect the lane through semantic or instant segmentation. However, some LMD techniques are also satisfied with multiple purposes along with detecting lane marks, such as road marking detection, road type classification, and drivable area detection. Initially, architectural information can be managed from the primary convolution network, such as ResNet, VGG, and FCN.

2.2.1. The Initial Network Architecture of LMD

CNN was first introduced to extract the lane feature in LMD by Kim et al. [17]. Additionally, random sample consensus (RANSAC) was used to group the identical architecture of the lane locations. The CNN architecture, shown in Figure 2, consists of three convolution layers, two subsampling layers, and three fully connected layers (FCL). The input dataset was converted into 192 × 28 after the ROI and edge detection. The last FCL provided the predicted output of 100 × 15.
Though it has improved LMD compared to the traditional methods, it also has some research limitations. The approach requires complex data processing unit and has a complex architecture of eight layers. Therefore, other researchers have developed other improved deep neural networks to overcome the existing limitations.

2.2.2. Lane Detection Based on Object Detection

Various types of visual detection systems are available for the autonomous driving system, such as road marking detection, vehicle detection, and, most importantly, lane marking detection. Sermanet et al. [24] introduced the overfeat technique, emphasizing the importance of a multi-supervised training approach, which simultaneously improved performance due to location, detection, and classification. Two key points typically focus on object detection, such as predicting the object and position of the object on the image.
Huval et al. [25] introduced empirical evaluation of the deep learning (EELane) technique with an overfeat detector to detect the highway’s lane markings. This research aims to apply six regressions to predict the lanes. The initial four regression dimensions indicate the finishing aspects of the line under the segmented lane boundary. The reaming regression dimension conceding the camera suggests the more profound finishing points. The geometrical information from CNN has been applied for many purposes, such as edge detection and inpainting, to assist the main task. The reader can go through it for a detailed understanding [26].
Seokju et al. [27] introduced VPGNet based on VPD, also a geometric estimation method of CNN. It is a modified version of the vanishing point tracking method, composed of four segments. The Vanishing point can guide road marking recognition and lane detection, which was the main contribution of the VPGNet. VPGNet has some post-processing framework for lane regression and clustering, increasing computational complexity. The architecture of the network is shown in Figure 3.
EELane and VPGNet showed the effectiveness of the multi-branch techniques where lane detection can be guided from prior knowledge by sharing different tasks into contiguous representations. Huang et al. [28] combined the spatial and temporal data in the CNN framework to detect the lane markings by selecting the lane boundaries. Therefore, the computational time is reduced, allowing it to be run more effectively in the automated driving car under intricate weather and traffic schemes in real-time. With this aim in mind, lane location estimation is obtained by evaluating Inverse Perspective Transformation (IPM) from the overhead view of the images using spatial and temporal relevancy of lanes. The images are cropped into relevant sub-image, carrying out the local lanes’ boundary information. Sequentially, the CNN framework is applied to detect the actual location and boundary of the lanes. The final structure is optimized to reduce the computational complexity by selecting the adjacent lanes based on the lane change for searching for the lanes’ actual position. The architecture of the network is depicted in Figure 4. The study of spatial and temporal relevancy of lanes made it different from EELane and VPGNet, whereas IM’s implementation created the condition of its robustness. However, this makes up for low illumination conditions, such as at night and rainy conditions [28].

2.2.3. Lane Detection Based on the Classification

Image classification refers to the discrimination process of objects available in the input image frame. However, the location of the lane can not be tracked through this process. Therefore, some modification is required in the classification technique to track the lane’s location. Let us consider the amendment on the classification is y = f(x,pm(p)), where f(x) is the CNN mapping function, and pm(p) is the prior knowledge depending on the lane location. Gurghian et al. [18] have come up with DeepLane depending on the same idea, which network architecture is shown in Figure 5. DeepLane received the training dataset, which was created from the image frames of the downward camera. It was classified into 317 classes, among which, 316 were for the probable lane position and reaming one was for missing lanes. A softmax function was applied to the last fully connected layer to achieve the probability distribution. The lane position was estimated Ei through the following equation:
E i = a r g m a x y i ,   0 i 316 ,   where ,   y i = y o , y 1 , . , y 316
Though DeepLane has achieved a better result than a complex network [17], the prior fixing of the lane position has limited its robustness. In addition, the classification techniques do not fit with lane marking detection, as it is associated with the high-level task. As discussed earlier, the regression of the lane coordinate as an object detection process is also a better possible way to detect lane marking detection.

2.2.4. Lane Detection Based on the Segmentation

Segmentation approaches such as [29,30,31] can be the best option for lane marking detection, as mentioned by Shriyash et al. [32]. These approaches strictly emphasize per-pixel classification rather than focusing on particular shapes. Lane detection based on the segmentation framework achieved more efficient results, except for the concern of the above limitation. This problem is solved by many strategies, such as the strategy proposed by Chiu et al. [33], which referred to the lane marking detection system as an image segmentation problem. However, the conventional segmentation approaches did not last long.
  • End-to-End Segmentation Approach
Due to the previous reason, the researcher started to apply end-to-end segmentation approaches for lane marking detection. The network can carry more features according to the larger size of the convolution kernel. Zhang et al. introduced a GCN [34] algorithm to detect particular lane areas. A lane departure system based on Mask-RCNN [35] is proposed by Riera Luis et al. to detect the lane marks and an additional Kalman filter to track the lanes. Shriyash et al. [36] proposed a CNN architecture that consists of ten neuron layers to detect the lanes in real time. Different types of lanes also have a notable contribution to more comprehensive recognition detection. The modified ERFNet architecture was designed by Fabio et al. [37] to classify the road lanes and identify the drivable area.
Semantic Segmentation through DCNN may have some deficiencies, as it has no learnable pooling parameters. For instance, there is no learnable parameter in max/min pooling or un-sampling layers. Therefore, there is an extreme possibility of losing many features when attempting to recognize a large-perspective field. Kontun et al. [38] introduced dilated convolution to resolve this issue, which can be studied more in [39]. Though this framework had significant advantages, the effective design of CNN architecture emphasizing dilated convolution has become a new issue.
Chen et al. proposed a Deep Convolution Neural Network based on the lane markings detector (LMD), aiming to have the optimal CNN architecture design with dilated convolution [40]. The lane markings detector, similar to ResNet [41] and VGG [42], is used as an encoder to classify, and DeconvNet [43], U-Net [44], and FCN [45] are used as a decoder to create feature maps. Additionally, dilated convolutions were embedded in the encode–decode section of the architecture shown in Figure 6. Lo et al. [40] introduced a CNN architecture based on DDB (Digressive Dilation Block) and FSS (Feature Size Selection), considering the spatial and downsampling operation, which was also embedded with dilation convolution [46].
Long-range information in lane marking detection is another concern. Wang et al. [47] designed a non-local operation depending on a non-local framework [48]. The model could extract the long-distance or range information, as long-distance information is also one of a lane’s properties. Li et al. [49] proposed Instance batch normalization and Attention Network (IANet) to emphasize the model for considering a particular lane region. It is more appropriate for two-class segmentation scenarios, according to the experimental result.
Considering efficient classification by focusing on pixels rather than shape, Jan et al. [50] came up with an adversarial network known as generative adversarial networks (GAN). It has a generator to create the synthetic data and a discriminator to differentiate the real data from the generator’s output data. The initial concept for the GAN was to predict data closely approximate to the real data. The recent concept tells us to differentiate accurately to determine whether the input is generated or real. A reader can go through [51,52,53] for further information about the GAN. Ghafoorian et al. [19] designed Embedding loss GAN (EL-GAN) based on the GNN concept. The framework is divided into two segments, as generator and discriminator. The schematic diagram of the EL-GAN framework is shown in Figure 7. U-Net’s unique algorithm is applied for the generator to train the input, and Tiramisu DenseNet [54] is used for detecting the lane markings. This process is continued up to the level of convergence. In the case of the discriminator, DenseNet [55,56] is used with the fully connected Generative Adversarial Network classification [57].
The framework generator is trained by adversarial embedding and Adam optimizer, whereas the discriminator is trained by stochastic gradient descent and ordinary cross-entropy. Embedding loss can be considered perceptual loss [58], whereas EL-GAN combines perceptual loss and CGAN.
  • Segmentation based on multitask
Geometrical features of roads also have an important role in lane marking detection, which have better performance results than VPGNet. Zhang et al. [59] proposed Geometric Constrained Network (GLCNet), which has multitasked to interlink the lane boundary and lane segmentation sub-structure. The architecture of GLCNet [59] is shown in Figure 8, which indicates that every decode section has a link with the encode section to transfer corresponding features into two distinct tasks. Therefore, the information from the decode sections can be redounded reciprocally. This multitask strategy opened the gate for the researchers to develop a framework for the link between lane boundary and lane area. Considering the same idea as GCLNet, John et al. [60] designed PSINet for multiple detection purposes, such as road scene labels, lane marks, and free space on the road.
In addition to the geometric or special feature, temporal correlation might have a significant effect where a lane can not be detected due to the linear structure of the captured video. As Long short-term memory (LSTM) has memory capture capability, the lane can be extracted from the previous frame by this LSTM approach. Hence, Qin et al. [61] proposed a CNN-LSTM method that includes two LSTM layers between the encode–decode stage. The major achievement of this method is that it has obtained ameliorate performance results under different occlusion scenarios. The architecture of the CNN-LSTM method is depicted in Figure 9, which indicates the temporal information transfer between the encode–decode stage through LSTM.

2.2.5. Simplification of the Post-Processing Step

Without considering the optimization by the post-processing step, the described frameworks extracted lane features more efficiently. It is very challenging to differentiate the lane features from the output, excluding the post-processing approach. Effective strategies are more important than particular network architecture to discover the optimal result. This sub-section focuses on these strategies, rather than a deep neural network (DNN) architecture, on lane marking detection.
There are two types of algorithmic output possible for lane marking detection using DNN, such as lane points and lane lines. Hence, the possibilty is raised to utilize different lane features, excluding post-processing steps. There might be three possible solutions to overcome the particular constraint: semantic segmentation by labelling each line as separate classes, instance segmentation by referring to every lane as a different instance, and multi-branch CNN structure by detecting every lane line through the individual branch.
Xingang et al. [20] applied a Spatial Convolution Neural Network (SCNN) to detect the lanes under occlusion scenarios as multi-class semantic segmentation. SCNN framework is based on the LargeFOV layout [62], and the weight of the initial thirteen convolution layers is taken from VGG16 [42]. To predict the lanes precisely, it generates pixel-wise probability maps for training the network. Consequently, it applies a CNN to differentiate the lane markings on its own. Finally, the probability maps are sent to the system to predict the lane markings of different classes. The architecture of the SCCN is shown in Figure 10, where various branches were designed to predict other lane classes.
Shriyash et al. [32] proposed Coordinate Network (CooNet) as a lane point regression approach. It is a multi-branch neural network shown in Figure 11, where lanes are predicted in their perspective branches. However, this network has no clustering process as the network directly provides the lane output through the coordinate regression.
To detect multi lanes with changes from the lanes, Davy et al. [63] introduced an end-to-end lane detection approach by applying the LaneNet deep learning method based on the encoding–decoding procedure E-Net [64], as shown in Figure 12. It takes the shared encodes from the input images and finds the embedding binary segmentation for each pixel for creating the cluster together. All pixels can associate with the neighbourhood pixels. It utilized the H-Net to collect the ideal information about the perspective transformation by imposing a relevant condition on the input image. The research aimed to take the challenge on lane changes, unlike the bird’s eye view. Additionally, this approach has no limitation on the number of lanes, whereas CooNet and SCNN can only detect up to four lanes.

2.2.6. Optimization Approaches

There is always a scope to improve the existing performance in the perspective research field. Still, there is a particular opportunity to optimize the lane marking detection process, as some research limitations exist for that particular application. The new question is, how can one researcher design a framework utilizing a trained model? The answer has come from the transfer learning technique and the knowledge distillation approach.
The dataset for transfer learning can be categorized into the target and source datasets. The target dataset relates to the task directly, and the source dataset indicates an additional dataset for the task. Fine-tuning becomes the major challenge due to the presence of both datasets in transfer learning. Hinton et al. [65] proposed a solution by introducing knowledge distillation, where the teacher network is used to guide the student network, which contains small parameters. It improves the performance of the student network. A few other researchers [66,67] enriched the knowledge distillation into attention distillation. This idea significantly improved lane marking detection when Kim et al. [68] designed Transfer Learning for Ego Lane detection (TLELane). In the TLELane architecture, two transfer learning stages differentiate the general scene from the road scene and capture the target lane to the left to right ego lane from that particular general scene. The attention map extracts high contextual features from different perspective layers in the trained lane marking detection segmentation-based network. These extracted features hold information regarding the rough outline and lane location. Thus, it is a promising way to replicate attention maps for the deeper block by utilizing the initial block. Apart from attention dilation, Hou et al. [22] introduced a self-learning distillation known as the self-attention distillation framework.
There is also a way to remove the post-processing steps, including clustering. Thus, the CNN needs to carry both the predicted lane and parametric description of each lane. Ze et al. [69] designed a combinational neural network with CNN and LSTM named Real-time Lane Network (RLaneNet). LSTM can face an uncertain number of lanes and also has a decoder to retain the parameter information of each lane. According to the mathematical assumption, a lane can be drawn from three corresponding coordinate points and a quadratic function. Based on this assumption, RLaneNet predicted three corresponding coordinate points of the lane that intersect the lane line with three horizontal lines. In contrast, Differentiable Least-squares Fitting Network (DLFNet) [70] lane curvature parameters are estimated by the weight of the least squares. These weights are captured from a deep neural network, and a geometrical loss function minimizes the area between ground truth and lane. The least-squares fitting can be defined as the following equation:
NXα = NY, where X and Y are coordinate matrices, N is the weighted pixel map, and α is the best-fitting curve parameter.

2.2.7. Loss Function in LMD Networks

The measurement of loss of the deep neural network is another key factor in making predicted data consistent with the ground truth data. It also ensures the optimization of the neural network. Different loss functions such as classification, regression, and adversarial training have different tasks in the network discussed in this section. The loss function for classification techniques is discussed below.
Cross-Entropy (CE), L1, and L2 loss have been used most in the case of pixel-level and lane line classification. The equation of L1 and L2 can be derived as follows:
L 1 y ^ , x = 1 h w m r , s , t y ^ r , s , t x r , s , t
L 2 y ^ , x = 1 h w m r , s , t y ^ r , s , t x r , s , t 2
where h refers to height, w refers to weight, and m refers to the number of channels. Additionally, x and y ^ represent the corresponding input and output.
The CE loss function can adopt the interclass competition mechanism, as indicated in Equation (3).
L c e = n c n i log m i
where mi is predicted probability and n1 is the class. For C = 2, the CE loss equation becomes Equation (4).
L b c e = n = 1 c = 2 n i log m i = n 1 log m 1 1 n 1 log 1 m 1
Weighed CE loss function, as shown by Equation (5), is used when unbalance exists in the sample data. For instance, there is a big unbalance ratio between background and lane line areas. Zou et al. settled the weight for the background and lane line by 1.0.
L w c e = i c w i n i log m i
In the case of significant errors, L1 and L2 are more sentient compared to small errors. Let us consider the simplified loss function of L2 as:
J = 1 2 y i y ^ i 2   where ,   y ^ i = W x i + b
d J d W = y i y ^ i σ W x i + b x i
From the above partial derivative, if the value of σ W x i + b becomes near 0 or 1, the derivative will also become 0, indicating an initial slow divergence. However, the derivative CE, as in Equation (8), does not depend on another multiplication term to have the possibility of bringing 0. Therefore, the CE loss function is more applicable in lane marking detection applications, mostly on semantic segmentation.
d L c e d W = σ m i y i x i
However, there is a possibility of scattering the learned feature, since the CE loss function only focused on the correct label, ignoring the difference between the incorrect label. The different solutions were dependent on the perspective function. Authors have used A-Softmax [71] and L-Softmax [72] functions, considering the perspective activation function. Contrastingly, Zhang et al. [59] proposed IoU loss, considering the perspective loss function, which indicates the relationship between ground truth and predicted probability. The loss function for regression techniques is discussed below in detail.
Coordinate and grid regression is based on the distance measurement used in [31,35,73]. Coordinate regression can be defined as Equation (9):
L colour = i = 1 15 x c i x g i + i = 1 15 y c i y g i
where x g i and y g i represent the coordinate of the ground truth, and x c i and y c i represent the corresponding predicted coordinate. At the same time, the grid regression can be expressed as Equation (10):
L = λ coord n = 0 R 2 m = 0 t 1 n m o b j x n x ^ n 2 + y n y ^ n 2 + λ coord n = 0 R 2 m = 0 t 1 n m o b j w n w ^ n 2 + h n h ^ n 2
where, x i , y i are the centre coordinates, w i , h i are the weight and height of the ground truth, and x ^ i , y ^ i , w ^ i , h ^ i comprise the corresponding prediction.
The loss function for adversarial techniques is discussed below in detail.
Generative adversarial networks (GAN) have created a different computer vision task with a generator to create the synthetic data and a discriminator to differentiate the real data from the generator’s output data. GAN’s loss function is defined as Equations (11)–(13), which is a modified version of CE.
m i n c m a x B V ( B , C ) = E x P x log B x + E y P y y log 1 B C y
m a x B V ( B , C ) = E x Pdata x log B x + E y P y y log 1 B C y
m a x B V ( B , C ) = E x Pdata x log B x + E y P y y log 1 B C y
Ghafoorian et al. [19] used various types of losses in the EL-GAN method, such as Cross-Entropy Loss L c e , L 2 loss and adversarial loss Lad are indicated in Equations (14)–(16), respectively.
L f t = L j t G x ; θ g e n , y = L c e G x ; θ g e n , y L c e y ˙ , y = 1 w h i w h j c y i , j ln y ˙ i , j
L 2 y ˙ , y ; x , θ d i s c = D e y ; x , θ d i s c D e y ˙ ; x , θ d i s c 2
L a d = E x P x log 1 D G x
L2 Loss differentiates the convoluted features between the real and generated images, referred to as perceptual loss. This loss was mostly applied in higher resolution fields to retain the enriched structure-preserving [58].
This section describes different types of loss functions that utilize in lane marking detection applications. There are also many other loss functions available in lane detection applications, though they are combined or part of mentioned loss functions.

2.2.8. Post-Processing

The post-processing step is required if the result from the neural network is the predicted lane coordinates. Clustering or curve-fitting approaches can be applied to transform these points into mathematical descriptions.
DBSCAN has been used mostly to interface the predicted lane pixels with the input images. DBSCAN works more efficiently than other clustering techniques like K-means in arbitrary and noisy clusters [74]. As the lanes’ positions are close to each other and arbitrary, such as straight or curved, DBSCAN would increase efficiency in interfacing the lane pixels. The closest distance point in DBSCAN depends on the value of ε and the minimum number of points for considering the same region. If the lane point is less or equal to the mentioned eps point, it would be considered in the same lane. On the contrary, the point would be considered as in a different cluster. The process would be continued according to the predicted information until all the lanes’ points are converged.
The clustering process makes the lane coordinate into different clusters. It is also challenging to transform different clusters into a mathematical description. As mentioned in the introduction section, distinct types of curve fitting function, such as Catmull–Rom spline, cubic B-spline, and parabolic are used for curve fitting, and cubic B-spline has shown more promising results [23].

2.2.9. Summary of the LMD Network

This section contains Table 1 and Table 2, which summarize different deep learning techniques used in the lane marking detection application and compares the performance of various models. As mentioned earlier, the network summary is also categorized as a single-stage and two-stage architecture with pros and cons.
Table 1. Summary of lane detection techniques using DNN.
Table 1. Summary of lane detection techniques using DNN.
AuthorDeep Learning TechniqueCategoriesAchievementLimitation
Single stage
Li et al. [49]IANetSegmentationSuitable for two-class segmentationHigh computation due to non-local features
Gurghian et al. [18]DeepLaneClassificationFast detection with simple architectureApplication scenarios are limited
Van et al. [70]DLFNetSegmentationIt does not have a predefined conditionApplicable for the fixed number of lanes
Ze et al. [69]RLaneNetRegressionCapable of handling uncertain lane numbers without post-processingThe lane ordinate needs to be predefined.
Hou et al. [22]self-attention distillationSegmentationThe strategy is more efficientHigh computational complexity
Kim et al. [68]TLELaneSegmentationSignificant achievement on the small datasetIt can only detect the ego lane
Davy et al. [63]LanenetSegmentationCapable of handling uncertain lane numberHigh computational complexity due to the H-Net
Xingang et al. [20]SCNNSegmentationSlice convolution for long laneHigh computational complexity
Shriyash et al. [32]CooNetRegressionLess computational network as does not require clusteringApplicable for the fixed number of lanes
Two-stage
Ghafoorian et al. [19]EL-GANSegmentationCan capture lane close to the labelRequire a high number of parameters
Qin et al. [61]CNN-LSTMSegmentationUseful for the occlusion sceneComputational is complex
Zhang et al. [59]GLCNetSegmentationCapable of making efficient interlinks between subsections of the networkHigh computational complexity and difficulties in the training stage
Chen et al. [40]LMD based on VGG16SegmentationDilated convolution can expand the predicted fieldThe performance result is lower
Huang et al. [28]Spatial and temporal-based CNNObject DetectionSpatial and temporal enrich the detection areaComplex architecture
Seokju et al. [27]VPGNetObject DetectionEfficient in different environmental conditionsHigh computational complexity due to the post-processing
Huval et al. [25]EELaneObject DetectionEffective for the occlusion sceneIt contains the perpetual prediction
Kim et al. [17]RANSACClassificationOvercome the limitations of traditional approachesThe structure of the network is not accurate enough
Table 2. Summary of the performances among various deep learning techniques.
Table 2. Summary of the performances among various deep learning techniques.
AuthorsDetection Rate (%)FPR (%)FNR (%)Recall (%)Accuracy (%)Precision (%)
Jongin et al. [75]93-----
Dan et al. [76]--10.03---
Soonhong et al. [77]88.70-----
Bei et al. [73]---92.8-95.49
Xue et al. [78]-5.5----
Gurghian et al. [18]- -99.9-98.96
He et al. [79]---93.80-95.49
Kim et al. [80]98-----
Seokju et al. [27]87--88--
Zhe et al. [81]-2.794.9995.01-94.94
Umar et al. [82]99-----
Davy et al. [63]-7.82.44-96.38-
Ghafoorian et al. [19] 4.123.36-96.39-
Xingang et al. [20]-6.171.8-96.53-
Ze et al. [69]-3.9-- -
Youjin et al. [83]92.4--- -
Xiaolong et al. [84]-1.414.53--95.65
Wenjie et al. [85]-7.7-- -
Tian et al. [86]---66.4 83.5
Huang et al. [28]---96.6-97.3
Ye et al. [87]--5.17---
Chao et al. [88]---6696.2689
Philion et al. [89]-7.24.5-95.2-
Azimi et al. [90]-- -85.95-
Sun et al. [91]-2.0- 96.4-
Zhang et al. [92]95.21-----
Zou et al. [61]-4.241.8495.897.285.7
Nguyen et al. [93]----98.1-
Hou et al. [22]-6.022.05-96.64-
Fabio et al. [37]----76.53-
Lo et al. [46]------
Zang et al. [94]82.44--- -
Mamidala et al. [95]----96.1-
Liu et al. [96]----97.9-
Ko et al. [97]-2.942.63-96.7-

3. Comparative Analysis

This section describes the constructive comparison of lane marking detection using deep neural networks regarding performance parameters and output visualization. There are five [19,20,61,63,70] segmentation-based DNN which are considered for the comparative sudsy. Moreover, the Tusimple has been considered the largest dataset for lane marking detection since 2018, and is available in [98]. This dataset was also trained and tested on many lane marking detection tasks. Therefore, the Tusimple dataset is considered for performance analysis in this comparative analysis. It contains annotated image frames of different weather conditions, such as straight lane, curve lane, shadow, occluded through vehicles, low light, etc. It has around 3.6k image frames for training and around 2.7k completely unknown image frames for testing. Instead of lane markings, the annotated full lane boundary is the main notability of the Tusimple dataset. The dimension of the images is 720 × 1280. The sample original image frames of the Tusimple dataset have been depicted in Figure 13.
There are particular statistical performance measurement units for evaluating neural network results in the image processing arena. For example, accuracy defines how accurately one model can predict the particular information from the image. Since accuracy cannot only be considered a reliable performance parameter to evaluate the performance that research, the other performance parameters, such as precision, recall, and F1score can make a reliable result to evaluate the framework’s performance. The brief descriptions of the performance parameters are given as follows.
Accuracy is the ratio of the actual true prediction to the total sample data.
A c c u r a c y = A c t u a l   T r u e   P r e d i c t i o n T o t a l   S a m p l e   d a t a
The false-positive score is the ratio of wrongly predicted data to the total number of predicted data.
F P S = F a l s e   P r e i c t e d   D a t a T o t l a   P r e d i c t e d   D a t a
The false-negative score is the ratio of missed ground truth data to the total number of ground truth data.
F N S = M i s s e d   G r o u n d   t r u t h   D a t a T o t l a   G r o u n d   t r u t h   D a t a
Table 3 illustrates the performance result of different deep neural networks used for lane detection on the Tusimple dataset. All of the networks that are arranged in Table 2 have been discussed in the previous section. The CNN-LSTM technique achieves a higher performance result than other mentioned networks. However, other networks also have a significant role in the lane marking detection application, as they have developed new efficient solutions to perspective research gaps.
The final output result of these [19,20,61,63,70] five DNN are depicted in Figure 14. Figure 14 gives a generalized idea about the predicted lane markings from the mentioned networks. A curve lane from the Tusimple dataset is elected for extracting the predicted lane markings.

4. Discussion

As lane detection is the ADAS system’s preliminary requirement, it is evident that researchers must develop an advanced model for lane marking detection. This research article discusses a complete overview of lane marking detection systems using deep learning techniques. The main contributions of the article can be categorized into four perspective views. Firstly, it describes different deep learning techniques according to their category in lane marking detection applications so that the researchers can find a specific path to implement neural network techniques in this particular application. Secondly, it describes different loss functions in order to help find a way to improve the performance. Thirdly, it also elaborates on the process of simplification and optimization of the network for simplifying the network architecture. Finally, comparative performance results with visualization of the final output of five existing techniques on the Tusimple dataset are elaborated, which will provide a track to the reader regarding the performance and optimization of the proposed perspective models.
Many vision-based computer-aided features are incorporated into the modern vehicle due to the improvement of the GPU and computational power of the hardware. Though previous researchers have done a tremendous task on lane marking detection, there are still many challenges to address. The first one would be the generalization problem, and better performance can be obtained by transplanting into CNN from the proposed method of [20,22]. Additionally, the supervisor learning process has a deficiency in the appropriate adjustment of the dataset’s different situations. As the neural network utilizes many parameters, real-time application and mobility are also quite challenging for this application.
Some other clues can be followed for promising results focusing on these challenges. As semantic segmentation has computation complexity regarding embedded liability and more efficient accuracy, CNN approaches might be investigated. Again, the supervised learning process can be transformed into a semi-supervised learning approach, as supervisor learning expects a vast amount of annotated data and computational time. Accurate and optimized lane marking detection systems can be designed under different critical situations through meta-learning. Significant feature extractors and detectors may arise from the existing segmentation architecture.

Author Contributions

A.A.M. has completed all the experiments, design, and manuscript writing. E.P.P. and J.H. have supervised this research. A.T. did model tuning, manuscript writing, and editing for this research work. B.J. also worked on tuning the model and writing and editing the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

The research described in this paper was supported by Multimedia University (MMU) Mini Fund (Grant No. MMUI/180170) and the Malaysian Ministry of Higher Education (MOHE) for Fundamental Research Grant Scheme (FRGS/1/2018/TK03/MMU/02/1).

Institutional Review Board Statement

Not Applicable.

Informed Consent Statement

Not Applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wang, X.; Wang, Y.; Wen, C. Robust lane detection based on gradient-pairs constraint. In Proceedings of the 30th Chinese Control Conference, Yantai, China, 22–24 July 2011; pp. 3181–3185. [Google Scholar]
  2. Hsiao, P.-Y.; Yeh, C.-W.; Huang, S.-S.; Fu, L.-C. A portable vision-based real-time lane departure warning system: Day and night. IEEE Trans. Veh. Technol. 2009, 58, 2089–2094. [Google Scholar] [CrossRef]
  3. Wang, J.-G.; Lin, C.-J.; Chen, S.-M. Applying fuzzy method to vision-based lane detection and departure warning system. Expert Syst. Appl. 2010, 37, 113–126. [Google Scholar] [CrossRef]
  4. Duan, J.; Zhang, Y.; Zheng, B. Lane line recognition algorithm based on threshold segmentation and continuity of lane line. In Proceedings of the 2016 2nd IEEE International Conference on Computer and Communications, ICCC 2016, Chengdu, China, 14–17 May 2017; pp. 680–684. [Google Scholar] [CrossRef]
  5. Gaikwad, V.; Lokhande, S. Lane Departure Identification for Advanced Driver Assistance. IEEE Trans. Intell. Transp. Syst. 2015, 16, 910–918. [Google Scholar] [CrossRef]
  6. Chai, Y.; Wei, S.J.; Li, X.C. The Multi-scale Hough transform lane detection method based on the algorithm of Otsu and Canny. Adv. Mater. Res. 2014, 1042, 126–130. [Google Scholar] [CrossRef]
  7. Ding, D.; Lee, C.; Lee, K.-Y. An adaptive road ROI determination algorithm for lane detection. In Proceedings of the 2013 IEEE International Conference of IEEE Region 10 (TENCON 2013), Xi’an, China, 22–25 October 2013. [Google Scholar] [CrossRef]
  8. Wu, P.-C.; Chang, C.-Y.; Lin, C.H. Lane-mark extraction for automobiles under complex conditions. Pattern Recognit. 2014, 47, 2756–2767. [Google Scholar] [CrossRef]
  9. Mu, C.; Ma, X. Lane Detection Based on Object Segmentation and Piecewise Fitting. TELKOMNIKA Indones. J. Electr. Eng. 2014, 12, 3491–3500. [Google Scholar] [CrossRef]
  10. Niu, J.; Lu, J.; Xu, M.; Lv, P.; Zhao, X. Robust Lane Detection using Two-stage Feature Extraction with Curve Fitting. Pattern Recognit. 2016, 59, 225–233. [Google Scholar] [CrossRef]
  11. Aung, T. Video Based Lane Departure Warning System using Hough Transform. Int. Inst. Eng. 2014. [Google Scholar] [CrossRef]
  12. Wang, Y.; Shen, D.; Teoh, E.K. Lane detection using spline model. Pattern Recognit. Lett. 2000, 21, 677–689. [Google Scholar] [CrossRef]
  13. Xiong, H.; Huang, L.; Yu, M.; Liu, L.; Zhu, F.; Shao, L. On the number of linear regions of convolutional neural networks. In Proceedings of the 37th International Conference on Machine Learning (ICML), online, 13–18 July 2020; pp. 10445–10454. [Google Scholar]
  14. McCall, J.; Trivedi, M. Video-based lane estimation and tracking for driver assistance: Survey, system, and evaluation. IEEE Trans. Intell. Transp. Syst. 2006, 7, 20–37. [Google Scholar] [CrossRef] [Green Version]
  15. Li, Y.; Chen, L.; Huang, H.; Li, X.; Xu, W.; Zheng, L.; Huang, J. Nighttime lane markings recognition based on Canny detection and Hough transform. In Proceedings of the 2016 IEEE International Conference on Real-Time Computing and Robotics, RCAR, Angkor Wat, Cambodia, 6–10 June 2016; pp. 411–415. [Google Scholar] [CrossRef]
  16. Mingfang, D.; Junzheng, W.; Nan, L.; Duoyang, L. Shadow Lane Robust Detection by Image Signal Local Reconstruction. Int. J. Signal Process. Image Process. Pattern Recognit. 2016, 9, 89–102. [Google Scholar] [CrossRef]
  17. Kim, J.; Lee, M. Robust lane detection based on convolutional neural network and random sample consensus. In Proceedings of the International Conference on Neural Information Processing, Kuching, Malaysia, 3–6 November 2014; Volume 8834, pp. 454–461. [Google Scholar] [CrossRef]
  18. Gurghian, A.; Koduri, T.; Bailur, S.V.; Carey, K.J.; Murali, V.N. DeepLanes: End-To-End Lane Position Estimation Using Deep Neural Networks. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 38–45. [Google Scholar] [CrossRef]
  19. Ghafoorian, M.; Nugteren, C.; Baka, N.; Booij, O.; Hofmann, M. EL-GAN: Embedding loss driven generative adversarial networks for lane detection. In Lecture Notes in Computer Science; Springer: Cham, Switzerland, 2019; Volume 11129, pp. 256–272. [Google Scholar] [CrossRef] [Green Version]
  20. Pan, X.; Shi, J.; Luo, P.; Wang, X.; Tang, X. Spatial as deep: Spatial CNN for traffic scene understanding. In Proceedings of the 32nd AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; pp. 7276–7283. [Google Scholar]
  21. Homayounfar, N.; Liang, J.; Ma, W.-C.; Fan, J.; Wu, X.; Urtasun, R. DAGMapper: Learning to map by discovering lane topology. In Proceedings of the IEEE International Conference on Computer Vision, Seoul, Korea, 27–28 October 2019; pp. 2911–2920. [Google Scholar] [CrossRef]
  22. Hou, Y.; Ma, Z.; Liu, C.; Loy, C.C. Learning lightweight lane detection CNNS by self attention distillation. In Proceedings of the IEEE International Conference on Computer Vision, Seoul, Korea, 27 October–3 November 2019; Volume 2019, pp. 1013–1021. [Google Scholar] [CrossRef] [Green Version]
  23. Tang, J.; Li, S.; Liu, P. A review of lane detection methods based on deep learning. Pattern Recognit. 2021, 111, 107623. [Google Scholar] [CrossRef]
  24. Sermanet, P.; Eigen, D.; Zhang, X.; Mathieu, M.; Fergus, R.; LeCun, Y. Overfeat: Integrated recognition, localization and detection using convolutional networks. In Proceedings of the 2nd International Conference Learning Representations ICLR 2014, Banff, AB, Canada, 14–16 April 2014. [Google Scholar]
  25. Huval, B.; Wang, T.; Tandon, S.; Kiske, J.; Song, W.; Pazhayampallil, J.; Andriluka, M.; Rajpurkar, P.; Migimatsu, T.; Chen-Yue, R. An Empirical Evaluation of Deep Learning on Highway Driving. arXiv 2015, arXiv:1504.01716. [Google Scholar]
  26. Elharrouss, O.; Almaadeed, N.; Al-Maadeed, S.; Akbari, Y. Image Inpainting: A Review. Neural Process. Lett. 2020, 51, 2007–2028. [Google Scholar] [CrossRef] [Green Version]
  27. Lee, S.; Kim, J.; Yoon, J.S.; Shin, S.; Bailo, O.; Kim, N.; Lee, T.-H.; Hong, H.S.; Han, S.-H.; Kweon, I.S. VPGNet: Vanishing Point Guided Network for Lane and Road Marking Detection and Recognition. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 1965–1973. [Google Scholar] [CrossRef] [Green Version]
  28. Huang, Y.; Chen, S.; Chen, Y.; Jian, Z.; Zheng, N. Spatial-temproal based lane detection using deep learning. In Artificial Intelligence Applications and Innovations; Springer International Publishing: Cham, Switzerland, 2018; Volume 519. [Google Scholar] [CrossRef]
  29. Al Mamun, A.; Em, P.P.; Hossen, J. Lane marking detection using simple encode decode deep learning technique: SegNet. Int. J. Electr. Comput. Eng. 2021, 11, 3032–3039. [Google Scholar] [CrossRef]
  30. Al Mamun, A.; Ping, E.P.; Hossen, J. An efficient encode-decode deep learning network for lane markings instant segmentation. Int. J. Electr. Comput. Eng. 2021, 11, 4982–4990. [Google Scholar] [CrossRef]
  31. Al Mamun, A.; Em, P.P.; Hossen, M.J.; Tahabilder, A.; Jahan, B. Efficient lane marking detection using deep learning technique with differential and cross-entropy loss. Int. J. Electr. Comput. Eng. 2022, 12, 4206–4216. [Google Scholar] [CrossRef]
  32. Chougule, S.; Koznek, N.; Ismail, A.; Adam, G.; Narayan, V.; Schulze, M. Reliable multilane detection and classification by utilizing CNN as a regression network. In Lecture Notes in Computer Science; Springer: Cham, Switzerland, 2019; Volume 11133, pp. 740–752. [Google Scholar] [CrossRef]
  33. Chiu, K.Y.; Lin, S.F. Lane detection using color-based segmentation. In Proceedings of the IEEE Intelligent Vehicles Symposium, Las Vegas, NV, USA, 6–8 June 2005; Volume 2005, pp. 706–711. [Google Scholar] [CrossRef]
  34. Peng, C.; Zhang, X.; Yu, G.; Luo, G.; Sun, J. Large kernel matters—Improve semantic segmentation by global convolutional network. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 1743–1751. [Google Scholar] [CrossRef] [Green Version]
  35. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask R-CNN. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 386–397. [Google Scholar] [CrossRef]
  36. Chougule, S.; Ismail, A.; Soni, A.; Kozonek, N.; Narayan, V.; Schulze, M. An efficient encoder-decoder CNN architecture for reliable multilane detection in real time. In Proceedings of the IEEE Intelligent Vehicles Symposium, Changshu, China, 26–30 June 2018; pp. 1444–1451. [Google Scholar] [CrossRef]
  37. Pizzati, F.; Garcia, F. Enhanced free space detection in multiple lanes based on single CNN with scene identification. In Proceedings of the 2019 IEEE Intelligent Vehicles Symposium (IV), Paris, France, 9–12 June 2019; pp. 2536–2541. [Google Scholar] [CrossRef] [Green Version]
  38. Yu, F.; Koltun, V. Multi-scale context aggregation by dilated convolutions. arXiv 2016, arXiv:1511.07122. [Google Scholar]
  39. Wang, P.; Chen, P.; Yuan, Y.; Liu, D.; Huang, Z.; Hou, X.; Cottrell, G. Understanding Convolution for Semantic Segmentation. In Proceedings of the 2018 IEEE Winter Conference on Applications of Computer Vision (WACV), Lake Tahoe, NV, USA, 12–15 March 2018; pp. 1451–1460. [Google Scholar] [CrossRef] [Green Version]
  40. Chen, P.-R.; Lo, S.-Y.; Hang, H.-M.; Chan, S.-W.; Lin, J.-J. Efficient Road Lane Marking Detection with Deep Learning. In Proceedings of the 2018 IEEE 23rd International Conference on Digital Signal Processing (DSP), Shanghai, China, 19–21 November 2018; pp. 1–5. [Google Scholar] [CrossRef] [Green Version]
  41. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar] [CrossRef] [Green Version]
  42. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2015, arXiv:1409.1556. [Google Scholar]
  43. Noh, H.; Hong, S.; Han, B. Learning deconvolution network for semantic segmentation. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1520–1528. [Google Scholar] [CrossRef] [Green Version]
  44. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Lecture Notes in Computer Science; Springer: Cham, Switzerland, 2015; Volume 9351, pp. 234–241. [Google Scholar] [CrossRef] [Green Version]
  45. Shelhamer, E.; Long, J.; Darrell, T. Fully Convolutional Networks for Semantic Segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 640–651. [Google Scholar] [CrossRef] [PubMed]
  46. Lo, S.-Y.; Hang, H.-M.; Chan, S.-W.; Lin, J.-J. Multi-Class Lane Semantic Segmentation using Efficient Convolutional Networks. In Proceedings of the 2019 IEEE 21st International Workshop on Multimedia Signal, Kuala Lumpur, Malaysia, 27–29 September 2019. [Google Scholar] [CrossRef] [Green Version]
  47. Wang, X.; Girshick, R.; Gupta, A.; He, K. Non-local Neural Networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 7794–7803. [Google Scholar] [CrossRef]
  48. Buades, A.; Coll, B.; Morel, J.-M. A non-local algorithm for image denoising. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), San Diego, CA, USA, 20–25 June 2005; Volume 2, pp. 60–65. [Google Scholar] [CrossRef]
  49. Li, W.; Qu, F.; Liu, J.; Sun, F.; Wang, Y. A lane detection network based on IBN and attention. Multimed. Tools Appl. 2020, 79, 16473–16486. [Google Scholar] [CrossRef]
  50. Goodfellow, I.J.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. arXiv 2014, arXiv:1406.2661. [Google Scholar] [CrossRef] [Green Version]
  51. Gulrajani, I.; Ahmed, F.; Arjovsky, M.; Dumoulin, V.; Courville, A. Improved Training of Wasserstein GANs. 2017. Available online: https://github.com/igul222/improved (accessed on 30 November 2020).
  52. Mao, X.; Li, Q.; Xie, H.; Lau, R.Y.K.; Wang, Z.; Smolley, S.P. Least Squares Generative Adversarial Networks. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2813–2821. [Google Scholar] [CrossRef] [Green Version]
  53. Arjovsky, M.; Chintala, S.; Bottou, L. Wasserstein generative adversarial networks. arXiv 2017, arXiv:1701.07875. [Google Scholar]
  54. Dosovitskiy, A.; Ros, G.; Codevilla, F.; Lopez, A.; Koltun, V. CARLA: An Open Urban Driving Simulator. arXiv 2017, arXiv:1711.03938. [Google Scholar]
  55. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2261–2269. [Google Scholar] [CrossRef] [Green Version]
  56. Jegou, S.; Drozdzal, M.; Vazquez, D.; Romero, A.; Bengio, Y. The One Hundred Layers Tiramisu: Fully Convolutional DenseNets for Semantic Segmentation. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2017; pp. 1175–1183. [Google Scholar] [CrossRef] [Green Version]
  57. Li, C.; Wand, M. Precomputed real-time texture synthesis with markovian generative adversarial networks. In Lecture Notes in Computer Science; Springer: Cham, Switzerland, 2016; Volume 9907, pp. 702–716. [Google Scholar] [CrossRef] [Green Version]
  58. Johnson, J.; Alahi, A.; Fei-Fei, L. Perceptual losses for real-time style transfer and super-resolution. In Lecture Notes Computer Science; Springer: Cham, Switzerland, 2016; Volume 9906, pp. 694–711. [Google Scholar] [CrossRef] [Green Version]
  59. Zhang, J.; Xu, Y.; Ni, B.; Duan, Z. Geometric Constrained Joint Lane Segmentation and Lane Boundary Detection. In Lecture Notes in Computer Science; Springer: Cham, Switzerland, 2018; Volume 11205, pp. 502–518. [Google Scholar] [CrossRef]
  60. John, V.; Karunakaran, N.M.; Guo, C.; Kidono, K.; Mita, S. Free Space, Visible and Missing Lane Marker Estimation using the PsiNet and Extra Trees Regression. In Proceedings of the 2018 24th International Conference on Pattern Recognition (ICPR), Beijing, China, 20–24 August 2018; pp. 189–194. [Google Scholar] [CrossRef]
  61. Zou, Q.; Jiang, H.; Dai, Q.; Yue, Y.; Chen, L.; Wang, Q. Robust lane detection from continuous driving scenes using deep neural networks. IEEE Trans. Veh. Technol. 2020, 69, 41–54. [Google Scholar] [CrossRef] [Green Version]
  62. Chen, L.-C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 40, 834–848. [Google Scholar] [CrossRef] [Green Version]
  63. Neven, D.; De Brabandere, B.; Georgoulis, S.; Proesmans, M.; Van Gool, L. Towards End-to-End Lane Detection: An Instance Segmentation Approach. In Proceedings of the IEEE Intelligent Vehicles Symposium, Rio de Janeiro, Brazil, 8–13 July 2018; Volume 2018, pp. 286–291. [Google Scholar] [CrossRef] [Green Version]
  64. Paszke, A.; Chaurasia, A.; Kim, S.; Culurciello, E. ENet: A Deep Neural Network Architecture for Real-Time Semantic Segmentation. azXiv 2016, arXiv:1606.02147. [Google Scholar]
  65. Hinton, G.; Vinyals, O.; Dean, J. Distilling the Knowledge in a Neural Network. arXiv 2015, arXiv:1503.02531. [Google Scholar]
  66. Zagoruyko, S.; Komodakis, N. Paying more attention to attention: Improving the performance of convolutional neural networks via attention transfer. arXiv 2017, arXiv:1612.03928. Available online: https://github.com/szagoruyko/attention-transfer (accessed on 30 November 2020).
  67. Wang, F.; Jiang, M.; Qian, C.; Yang, S.; Li, C.; Zhang, H.; Wang, X.; Tang, X. Residual attention network for image classification. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 6450–6458. [Google Scholar] [CrossRef] [Green Version]
  68. Kim, J.; Park, C. End-To-End Ego Lane Estimation Based on Sequential Transfer Learning for Self-Driving Cars. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2017; Volume 2017, pp. 1194–1202. [Google Scholar] [CrossRef]
  69. Wang, Z.; Ren, W.; Qiu, Q. LaneNet: Real-Time Lane Detection Networks for Autonomous Driving. arXiv 2018, arXiv:1807.01726. [Google Scholar]
  70. Van Gansbeke, W.; De Brabandere, B.; Neven, D.; Proesmans, M.; Van Gool, L. End-to-end lane detection through differentiable least-squares fitting. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW), Seoul, Korea, 27–28 October 2019; pp. 905–913. [Google Scholar] [CrossRef] [Green Version]
  71. Liu, W.; Wen, Y.; Yu, Z.; Li, M.; Raj, B.; Song, L. SphereFace: Deep hypersphere embedding for face recognition. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 June 2017; pp. 6738–6746. [Google Scholar] [CrossRef] [Green Version]
  72. Liu, W.; Wen, Y.; Yu, Z.; Yang, M. Large-Margin Softmax Loss for Convolutional Neural Networks. In Proceedings of the 33rd International Conference on International Conference on Machine Learning, New York, NY, USA, 19–24 June 2016. [Google Scholar]
  73. He, B.; Ai, R.; Yan, Y.; Lang, X. Accurate and robust lane detection based on Dual-View Convolutional Neutral Network. In Proceedings of the IEEE Intelligent Vehicles Symposium, Gothenburg, Sweden, 19–22 June 2016; pp. 1041–1046. [Google Scholar] [CrossRef]
  74. Dang, S.; Ahmad, P.H. Performance Evaluation of Clustering Algorithm Using Different Datasets. J. Inf. Eng. Appl. 2015, 5, 39–46. Available online: www.ijarcsms.com (accessed on 10 September 2020).
  75. Son, J.; Yoo, H.; Kim, S.; Sohn, K. Real-time illumination invariant lane detection for lane departure warning system. Expert Syst. Appl. 2015, 42, 1816–1824. [Google Scholar] [CrossRef]
  76. Levi, D.; Garnett, N.; Fetaya, E. StixelNet: A Deep Convolutional Network for Obstacle Detection and Road Segmentation. In Proceedings of the British Machine Vision Conference (BMVC); BMVA Press: Swansea, UK, 2015; pp. 109.1–109.12. [Google Scholar] [CrossRef]
  77. Jung, S.; Youn, J.; Sull, S. Efficient lane detection based on spatiotemporal images. IEEE Trans. Intell. Transp. Syst. 2016, 17, 289–295. [Google Scholar] [CrossRef]
  78. Li, X.; Wu, Q.; Kou, Y.; Hou, L.; Yang, H. Lane detection based on spiking neural network and hough transform. In Proceedings of the 2015 8th International Congress on Image and Signal Processing (CISP), Shenyang, China, 14–16 October 2015; pp. 626–630. [Google Scholar] [CrossRef]
  79. He, B.; Ai, R.; Yan, Y.; Lang, X. Lane marking detection based on convolution neural network from point clouds. In Proceedings of the 2016 IEEE 19th International Conference on Intelligent Transportation Systems (ITSC), Rio de Janeiro, Brazil, 1–4 November 2016; pp. 2475–2480. [Google Scholar] [CrossRef]
  80. Kim, J.; Kim, J.; Jang, G.-J.; Lee, M. Fast learning method for convolutional neural networks using extreme learning machine and its application to lane detection. Neural Netw. 2017, 87, 109–121. [Google Scholar] [CrossRef]
  81. Chen, Z.; Chen, Z. RBNet: A deep neural network for unified road and road boundary detection. In Lecture Notes Computer Science; Springer: Cham, Switzerland, 2017; Volume 10634, pp. 677–687. [Google Scholar] [CrossRef]
  82. Ozgunalp, U.; Fan, R.; Ai, X.; Dahnoun, N. Multiple Lane Detection Algorithm Based on Novel Dense Vanishing Point Estimation. IEEE Trans. Intell. Transp. Syst. 2017, 18, 621–632. [Google Scholar] [CrossRef] [Green Version]
  83. Youjin, T.; Wei, C.; Xingguang, L.; Lei, C. A robust lane detection method based on vanishing point estimation. Procedia Comput. Sci. 2018, 131, 354–360. [Google Scholar] [CrossRef]
  84. Liu, X.; Deng, Z.; Yang, G. Drivable road detection based on dilated fpn with feature aggregation. In Proceedings of the 2017 IEEE 29th International Conference on Tools with Artificial Intelligence (ICTAI), Boston, MA, USA, 6–8 November 2017; pp. 1128–1134. [Google Scholar] [CrossRef]
  85. Song, W.; Yang, Y.; Fu, M.; Li, Y.; Wang, M. Lane Detection and Classification for Forward Collision Warning System Based on Stereo Vision. IEEE Sensor J. 2018, 18, 5151–5163. [Google Scholar] [CrossRef]
  86. Tian, Y.; Gelernter, J.; Wang, X.; Chen, W.; Gao, J.; Zhang, Y.; Li, X. Lane marking detection via deep convolutional neural network. Neurocomputing 2018, 280, 46–55. [Google Scholar] [CrossRef]
  87. Ye, Y.Y.; Hao, X.L.; Chen, H.J. Lane detection method based on lane structural analysis and CNNs. IET Intell. Transp. Syst. 2018, 12, 513–520. [Google Scholar] [CrossRef]
  88. Chao, F.; Yu-Pei, S.; Ya-Jie, J. Multi-Lane Detection Based on Deep Convolutional Neural Network. IEEE Access 2019, 7, 150833–150841. [Google Scholar] [CrossRef]
  89. Philion, J. FastDraw: Addressing the long tail of lane detection by adapting a sequential prediction network. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 11574–11583. [Google Scholar] [CrossRef] [Green Version]
  90. Azimi, S.M.; Fischer, P.; Korner, M.; Reinartz, P. Aerial LaneNet: Lane-Marking Semantic Segmentation in Aerial Imagery Using Wavelet-Enhanced Cost-Sensitive Symmetric Fully Convolutional Neural Networks. IEEE Trans. Geosci. Remote Sens. 2019, 57, 2920–2938. [Google Scholar] [CrossRef] [Green Version]
  91. Sun, Y.; Li, J.; Sun, Z.P. Multi-Lane Detection Using CNNs and A Novel Region-grow Algorithm. J. Phys. Conf. Ser. 2019, 1187, 032018. [Google Scholar] [CrossRef]
  92. Zhang, W.; Liu, H.; Wu, X.; Xiao, L.; Qian, Y.; Fang, Z. Lane marking detection and classification with combined deep neural network for driver assistance. Proc. Inst. Mech. Eng. Part D J. Automob. Eng. 2019, 233, 1259–1268. [Google Scholar] [CrossRef]
  93. Nguyen, T.-P.; Tran, V.-H.; Huang, C.-C. Lane Detection and Tracking Based on Fully Convolutional Networks and Probabilistic Graphical Models. In Proceedings of the 2018 IEEE International Conference on Systems, Man, and Cybernetics (SMC), Miyazaki, Japan, 7–10 October 2018; pp. 1282–1287. [Google Scholar] [CrossRef]
  94. Zang, J.; Zhou, W.; Zhang, G.; Duan, Z. Traffic Lane Detection using Fully Convolutional Neural Network. In Proceedings of the 2018 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC), Honolulu, HI, USA, 12–15 November 2018; pp. 305–311. [Google Scholar] [CrossRef]
  95. Mamidala, R.S.; Uthkota, U.; Shankar, M.B.; Antony, A.J.; Narasimhadhan, A.V. Dynamic Approach for Lane Detection using Google Street View and CNN. In Proceedings of the TENCON 2019—2019 IEEE Region 10 Conference (TENCON), Kochi, India, 17–20 October 2019; pp. 2454–2459. [Google Scholar] [CrossRef] [Green Version]
  96. Liu, B.; Liu, H.; Yuan, J. Lane Line Detection based on Mask R-CNN. In Proceedings of the 3rd International Conference on Mechatronics Engineering and Information Technology (ICMEIT 2019), Dalian, China, 29–30 April 2019; Volume 87, pp. 696–699. [Google Scholar] [CrossRef]
  97. Ko, Y.; Lee, Y.; Azam, S.; Munir, F.; Jeon, M.; Pedrycz, W. Key Points Estimation and Point Instance Segmentation Approach for Lane Detection. IEEE Trans. Intell. Transp. Syst. 2022, 23, 8949–8958. [Google Scholar] [CrossRef]
  98. Tusimple Benchmark Ground Truth Issue #3 TuSimple/tusimple-benchmark. Available online: https://github.com/TuSimple/tusimple-benchmark/issues/3 (accessed on 2 December 2020).
Figure 1. Different pre-processing technique (a) original, (b) cropped, (c) brighten (d) mirrored, (e) rotating and (f) perspective.
Figure 1. Different pre-processing technique (a) original, (b) cropped, (c) brighten (d) mirrored, (e) rotating and (f) perspective.
Sensors 22 07682 g001
Figure 2. The architecture of CNN based lane marking detection technique.
Figure 2. The architecture of CNN based lane marking detection technique.
Sensors 22 07682 g002
Figure 3. Schematic diagram of VPGNet.
Figure 3. Schematic diagram of VPGNet.
Sensors 22 07682 g003
Figure 4. Schematic diagram of spatial and temporal based LMD technique.
Figure 4. Schematic diagram of spatial and temporal based LMD technique.
Sensors 22 07682 g004
Figure 5. Schematic diagram of DeepLane.
Figure 5. Schematic diagram of DeepLane.
Sensors 22 07682 g005
Figure 6. Schematic diagram of Deep Convolution Neural Network based on the lane markings detector (LMD).
Figure 6. Schematic diagram of Deep Convolution Neural Network based on the lane markings detector (LMD).
Sensors 22 07682 g006
Figure 7. Schematic diagram of EL-GAN.
Figure 7. Schematic diagram of EL-GAN.
Sensors 22 07682 g007
Figure 8. Schematic diagram of GLCNet.
Figure 8. Schematic diagram of GLCNet.
Sensors 22 07682 g008
Figure 9. Schematic diagram of CNN-LSTM.
Figure 9. Schematic diagram of CNN-LSTM.
Sensors 22 07682 g009
Figure 10. Schematic diagram of SCCN.
Figure 10. Schematic diagram of SCCN.
Sensors 22 07682 g010
Figure 11. Schematic diagram of CooNet.
Figure 11. Schematic diagram of CooNet.
Sensors 22 07682 g011
Figure 12. Schematic diagram of Lanenet.
Figure 12. Schematic diagram of Lanenet.
Sensors 22 07682 g012
Figure 13. Sample image frames of the Tusimple dataset.
Figure 13. Sample image frames of the Tusimple dataset.
Sensors 22 07682 g013
Figure 14. Predicted lane marking using DNN (a) Input (b) Lanenet (c) SCNN (d) CNN-LSTM (e) ERFNet-DLSF and (f) El-GAN.
Figure 14. Predicted lane marking using DNN (a) Input (b) Lanenet (c) SCNN (d) CNN-LSTM (e) ERFNet-DLSF and (f) El-GAN.
Sensors 22 07682 g014
Table 3. Summary of lane detection techniques using DNN.
Table 3. Summary of lane detection techniques using DNN.
AuthorsDNN MethodFPSFNSAccuracy
Davy et al. [63]Lanenet7.82.4496.38
Ghafoorian et al. [19]EL-GAN4.123.3696.39
Xingang et al. [20]SCNN6.171.896.53
Qin et al. [61]CNN-LSTM0.014160.018697.30
Hou et al. [22]Self-attention distillation6.022.0596.64
Van et al. [70]ERFNet-DLSF0.10640.098393.38
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Mamun, A.A.; Ping, E.P.; Hossen, J.; Tahabilder, A.; Jahan, B. A Comprehensive Review on Lane Marking Detection Using Deep Neural Networks. Sensors 2022, 22, 7682. https://doi.org/10.3390/s22197682

AMA Style

Mamun AA, Ping EP, Hossen J, Tahabilder A, Jahan B. A Comprehensive Review on Lane Marking Detection Using Deep Neural Networks. Sensors. 2022; 22(19):7682. https://doi.org/10.3390/s22197682

Chicago/Turabian Style

Mamun, Abdullah Al, Em Poh Ping, Jakir Hossen, Anik Tahabilder, and Busrat Jahan. 2022. "A Comprehensive Review on Lane Marking Detection Using Deep Neural Networks" Sensors 22, no. 19: 7682. https://doi.org/10.3390/s22197682

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop