Next Article in Journal
Classification Predictive Model for Air Leak Detection in Endoworm Enteroscopy System
Previous Article in Journal
Performance Analysis of Wearable Dual-Band Patch Antenna Based on EBG and SRR Surfaces
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Foggy Lane Dataset Synthesized from Monocular Images for Lane Detection Algorithms

1
China-UK Low Carbon College, Shanghai Jiao Tong University, Shanghai 201306, China
2
Midea Group, Shanghai 201702, China
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(14), 5210; https://doi.org/10.3390/s22145210
Submission received: 19 April 2022 / Revised: 7 July 2022 / Accepted: 7 July 2022 / Published: 12 July 2022
(This article belongs to the Section Vehicular Sensing)

Abstract

:
Accurate lane detection is an essential function of dynamic traffic perception. Though deep learning (DL) based methods have been widely applied to lane detection tasks, such models rarely achieve sufficient accuracy in low-light weather conditions. To improve the model accuracy in foggy conditions, a new approach was proposed based on monocular depth prediction and an atmospheric scattering model to generate fog artificially. We applied our method to the existing CULane dataset collected in clear weather and generated 107,451 labeled foggy lane images under three different fog densities. The original and generated datasets were then used to train state-of-the-art (SOTA) lane detection networks. The experiments demonstrate that the synthetic dataset can significantly increase the lane detection accuracy of DL-based models in both artificially generated foggy lane images and real foggy scenes. Specifically, the lane detection model performance (F1-measure) was increased from 11.09 to 70.41 under the heaviest foggy conditions. Additionally, this data augmentation method was further applied to another dataset, VIL-100, to test the adaptability of this approach. Similarly, it was found that even when the camera position or level of brightness was changed from one dataset to another, the foggy data augmentation approach is still valid to improve model performance under foggy conditions without degrading accuracy on other weather conditions. Finally, this approach also sheds light on practical applications for other complex scenes such as nighttime and rainy days.

1. Introduction

With the rapid development of autonomous driving and assisted driving technologies, the accurate perception of dynamic traffic elements has become an essential prerequisite for reliable active safety strategies [1,2]. As a vital function of dynamic traffic perception, lane detection has gained increasing attention in recent years. Meanwhile, vision-based lane detection technology has seen significant progress, and deep neural networks, such as the mainstream technology for computer vision, have been widely utilized in lane detection [3,4,5,6,7,8,9,10]. However, the application scenarios of most deep learning (DL)-based lane detection are still limited to ideal weather conditions, e.g., clear daytime. However, little research [11,12,13,14,15,16] focuses on low-light weather conditions, such as foggy and rainy days, which are significant for increasing the adaptability of perception technology for autonomous driving vehicles. In fact, lane detection can be more challenging in complicated weather conditions. As the meteorological and lighting conditions deviate from the ideal case, both the clarity and contrast of the images decrease. Hence, problems such as color distortion and loss of fine features occur in the image, which bring about more difficulties in extracting lane features, thus affecting the accuracy of the computer vision based model [17,18]. Therefore, this paper aims to improve the accuracy of DL-based lane detection model in the complex foggy weather condition.
Over the past years, a great deal of attention has been paid to image dehazing algorithms to increase scene visibility, from traditional image processing methods [19,20,21,22,23,24,25] to convolutional neural network (CNN) based methods [26,27,28,29,30,31,32]. Though many algorithms achieve the goal of dehazing, in practice, introducing the defogging algorithm in the lane detection pipeline is computationally expensive. This means it is costly and infeasible to deploy on resource-constrained devices or in real-time applications. Hence, improving the accuracy of DL-based lane detection model without introducing an extra computation budget is crucial. Dataset synthesis is one of the most prevailing methods to help improve the performance of deep learning models. For a data-driven algorithm, the number and diversity of the training data directly influence the DL-based lane detection model’s performance. Therefore, to improve the accuracy of lane detection under complicated weather conditions, the scale of the dataset in the corresponding scenes needs to be significantly increased [33].
Several open-source annotated datasets used for traffic lane detection in autonomous driving include Caltech Lanes [34], KITTI [35], CityScapes [36], TuSimple [37], CULane [6], ApolloScape [38], BDD100K [39], LLAMAS [40], A2D2 [41], VIL-100 [42] et al. Caltech Lanes only consists of 1224 images, while the lane markings in the images of KITTI and CityScapes are not annotated; thus, these datasets are usually used for model testing and evaluation. TuSimple and LLAMAS mainly focus on highways in sunny scenes, deficient of lane marking images under complex weather conditions. Though ApolloScape and BDD100K provide large scale images under various scenarios for driving, they lack the annotation of lane markings on an instance level. It is important to label the multiple lanes in each image of these datasets as different categories in autonomous driving, especially for trajectory planning, autonomous navigation, and lane change tasks. CULane, A2D2, and VIL-100 meet the mentioned requirements and already contain several complex weather scenes. However, the foggy weather is missing in CULane and A2D2, leading to the low accuracy of the lane detection model under foggy conditions. Additionally, the scales of A2D2 and VIL-100 are relatively small, with the latter only containing 300 images in hazy weather.
A feasible method to enlarge the dataset under hazy weather conditions is artificially generating lane images under foggy conditions using images under clear weather. Thus, the trained model can improve its performance in all scenes. The method of augmenting existing datasets by synthesized images has been used in several scenes, such as rainy [11] and nighttime [12] conditions. Despite some previous studies in synthesizing foggy images for autonomous driving, including FRIDA [13], FRIDA2 [14], Foggy CityScapes [15], and Multifog KITTI [16], there are few studies focusing on augmenting datasets using artificially generated foggy images and its effectiveness for lane detection task. Tarel et al. constructed synthetic outdoor foggy datasets FRIDA (Foggy Road Image Database) and FRIDA2 based on numerical images using SiVICTM software. The former contains 72 foggy images with 18 original images, and the latter contains 330 images in 66 diverse scenes. However, the number of images in these two datasets is too small, and the images are also at low resolutions with poor visual effects (as shown in Figure 1), hence insufficient to support lane detection model training for higher accuracy, especially in the foggy scene. Sakaridis et al. established Foggy Cityscapes, and Mai et al. established Multifog KITTI by adding synthetic fog to images from Cityscapes and KITTI, respectively. However, the scales of these two datasets are relatively small and the annotations of the lane markings are missing. Moreover, these foggy image synthesis methods rely on image perception and depth information using complex sensors like Lidar, RGBD camera, or binocular camera, which would introduce additional costs in practice. Additionally, Figure 1d,e also shows that the quality of the generated foggy images in Foggy CityScapes suffers from incorrect depth (Figure 1d), especially in the foreground.
In this work, we propose a new dataset augmentation method, using a self-supervised monocular depth prediction method to extract the depth information from monocular lane images and then synthesize foggy images based on the atmospheric scattering model. Our proposed framework estimates the depth map only from the original image, which means the method could be easily applied to other datasets without depth information collected by extra sensors. We established the FoggyCULane dataset by enlarging the CULane dataset with the synthesis of foggy images using our method to improve the accuracy of the lane detection model in foggy scenes. Networks with outstanding lane detection performance in recent years are trained with our FoggyCULane and CULane to verify the feasibility and effectiveness of our method in different complex scenes, especially in foggy scenes. The main contributions of this paper are:
  • A new dataset augmentation and synthesis method was proposed for lane detection in foggy conditions, which highly improved the accuracy of the lane detection model under foggy weather without introducing extra computational cost or complex framework for the algorithm.
  • We established a new dataset, FoggyCULane, which contains 107,451 frames of labelled foggy lanes. This would help the community and researchers to develop and validate their own data-driven lane detection or dehaze algorithms.

2. Background

Many studies have analyzed the reasons for the loss of sharpness and contrast in foggy images through modeling. In 1924, Koschieder et al. introduced the standard optical model [43] for daytime fog, which has been used extensively in image dehazing. McCartney E J et al. [44] further reported in 1976 that the absorption and scattering of light by suspended particles (including water droplets, dust, and aerosols) in the atmosphere causes the attenuation of light transmitting between the target and the camera. Besides, the scattering of suspended particles also generates background light, which leads to a decrease in contrast and saturation of images under foggy conditions. In 1999, Nayar et al. [45] established a mathematical model to describe the atmospheric scattering process clearly. The model assumes that under a strong scattering medium (e.g., foggy scenes), firstly, the light reflected by the target is absorbed and scattered by suspended particles in the atmosphere, resulting in the attenuation of the light reflected by the target, thus decreasing the brightness and contrast of the imaging. Secondly, ambient light, such as sunlight, is scattered by the scattering medium in the atmosphere to form air light. Sometimes the intensity of this scattered light is greater than that of the target light, thus causing the image to blur.
However, the reconstruction of foggy images using the above model requires depth information from the original image. Since most current traffic lane datasets do not provide the required depth information, including the most widely used CULane dataset, the extraction of depth information needs to be performed as the first step of foggy image synthesis.
Depending on the number of viewing angles, the depth estimation of images can be divided into binocular depth estimation and monocular depth estimation. However, the binocular algorithms require multiple images of the same scene with different angles, from which the corresponding targets are matched and reconstructed in 3D.
Therefore, monocular depth prediction has been extensively studied in the field of computer vision in recent years. Based on the basic assumption that there are mapping relationships between RGB images and depth maps, data-driven deep learning methods have been proposed and applied to monocular depth estimation problems. Among them, deep learning techniques represented by Convolution Neural Network (CNN) have made substantial progress and gradually become the mainstream method for monocular image depth estimation [46,47,48,49,50].
Supervised by images with depth annotations, Eigen et al. [51] first trained a CNN-based depth estimation model and realized monocular depth estimation using the deep learning method. They proposed to estimate the depth of a monocular image using two scales of neural networks: a coarse-scale network to predict the global depth of the image and a fine-scale network to optimize the local details.
In fact, acquiring large amounts of diverse images with exact depth information is challenging, and supervised learning cannot adapt to changing environments. Supervised learning methods require that each RGB image has its corresponding depth label, and the depth label annotation usually requires depth cameras or light detection and ranging (LIDAR), which are limited by detection range and cost, respectively. In addition, depth labels acquired from LIDAR are usually sparse points that are difficult to match with the original pixel-level image. Therefore, unsupervised or self-supervised depth estimation methods that do not require depth labels have been a research trend in recent years. One of the most prominent methods is the Monodepth2 network proposed by Godard et al. [52].

3. Methods

In this paper, we proposed a dataset enhancement method for lane detection in foggy conditions based on the original CULane dataset, using a self-supervised monocular depth prediction method to extract the depth information from monocular lane images and synthesize foggy images based on the atmospheric scattering model. The main framework of our method is shown in Figure 2.

3.1. The Standard Optical Model of Foggy Images

From Koschmieder’s law [43], the light intensity of the target observed by the camera mainly contains two parts: the light reflected by the object that reaches the detection system after attenuation by the transmission medium and the background light formed by particle scattering. The model of imaging in foggy (atmospheric scattering) conditions is defined by Koschmieder’s law as follows:
I ( x ) = J ( x ) t ( x ) + A ( 1 t ( x ) )
where x indicates one certain pixel of the image, I ( x ) refers to the observed foggy image of the object at the pixel x , and J ( x ) refers to the original clear image of the object. A is the global skylight representing the ambient light in the atmosphere and is generally assumed to be a constant. Additionally, t ( x ) determines the transmission of the light from the object in the atmosphere. Assuming the medium is homogeneous, the transmission t ( x ) can be further determined as:
t ( x ) = exp ( β l ( x ) )
where l ( x ) is the distance between the object and observer, β is the extinction coefficient, and larger β represents the thicker fog.
It can be seen that image degradation and information loss increase with depth l ( x ) . Therefore, to generate synthetic fog images from the original clear scene, the key is to obtain the value of l ( x ) in (2), that is, the depth information of the clear images needs to be extracted.

3.2. Monocular Depth Estimation

With the rapid development in deep learning, deep neural networks have shown excellent performance in recovering the pixel-level depth map from a single image [46,47,48,49,50]. In this paper, the Monodepth2 network (trained on the KITTI dataset) proposed by Godard et al. [52] is used for depth extraction of clear images in the CULane dataset, which is, in turn, employed to generate foggy images.
Monodepth2 is a self-supervised monocular depth estimation network that combines depth estimation and pose estimation to achieve pixel-level SOTA depth predictions. The recovery of the depth information was achieved based on the photo-consistency assumption, i.e., points in the same space should also have the same luminosity in the projection of different views. Based on minimum reprojection loss, the full-resolution multi-scale sampling method, and auto-masking loss, Monodepth2 can achieve outstanding depth estimation while trained with either monocular or binocular images as input data.
The loss functions of Monodepth2 are defined as follows:
{ I t t = I t   p r o j ( D t , T t t , K ) p e ( I a , I b ) = α 2 ( 1 SSIM ( I a , I b ) ) + ( 1 α ) I a I b L p = m i n t   p e ( I t , I t t ) L s = | x d t * | e | x I t | + | y d t * | e | y I t | , d t * = d t / d t ¯ L = μ L p + λ L s
where I t t is defined as the relative pose for each source view I t with respect to the target image I t ’s pose; p r o j (   ) projects the predicted depth map D t in resulting 2D coordinates with I t and 〈 〉 is the sampling operator; K represents pre-computed intrinsics for simplicity; p e is the photometric error function constructed by L1-norm and Structural Similarity(SSIM); d t * is the mean-normalized inverse depth; L s is edge-aware smoothness and L p is the per-pixel photometric loss; μ and λ are coefficients corresponding to auto-masking stationary pixels and multi-scale estimation respectively; L is the final loss function.
Considering that our purpose is to generate foggy images whose density varies with distance, it is unnecessary to obtain the real and absolute depth information. Instead, the trend of depth variation or the relative variation of depth meets the requirement. Therefore, Monodepth2 is sufficient for this current study.

3.3. Foggy Image Generation

To adapt to the relative depth, the actual transmission value t ( x ) is reformed as:
t ( x ) = 1 exp ( β ( 1 l ( x ) ) )
Here, l ( x ) is the depth matrix in the range [ 0 ,   1 ] , which indicates the normalized depth estimated by Monodepth2. Substituting (2) with (4) we can acquire the observed foggy image of the object I(x). The synthetic monocular depth map and foggy images acquired using the above method are shown in Figure 3. The acquired depth map can clearly represent the depth of each part of the original image and the trend of depth change. The generated foggy images look like real foggy pictures from the driver’s perspective.

3.4. Establishment of FoggyCULane Dataset

Figure 4 presents the overall structure of the original CULane dataset, which involves 133,235 pictures of traffic lanes collected by cameras mounted on different vehicles, among which 88,880 for the training set, 9675 for the validation set, and 34,680 for the test set. The test set is divided into the normal scene and eight challenging categories, including crowded, night, no line, shadow, arrow, dazzle light, curve, and crossroad.
The CULane dataset focuses on the current lane and up to 4 adjacent lane markings, which are the most concerned lanes during driving. The traffic lanes are manually annotated with cubic splines for each image, and the labels are recorded as a set of coordinate points. The lanes are still annotated based on semantic information for cases where the lane markings are obscured or not visible by vehicles.
Through the foggy image generation method, we expand the training set, validation set, and test set of CULane, respectively. Considering the diversity and complexity of fog density in the real foggy scene, we attempt to simulate fog scenes more realistically and adapt the dataset to multiple density scenes. Through experimentation, we found that the values of β should be limited to a reasonable range, or the artificially generated fog would be almost invisible or too thick. Thus, β is finally set to 2, 3, and 4, respectively, to generate three levels of foggy scenes with increasing fog density, as shown in Figure 5. Meanwhile, because we use the original images in the CULane dataset, the lane annotations are still applicable to the synthetic images.
We selected 16,532 images from the CULane train set as the original inputs for foggy image generation and processed them separately with different fog densities ( β = 2, 3, 4). Together with the original lane images, the total number of the training set for our FoggyCULane was 138,476. Similarly, while keeping the original test set images, we selected all the images in the normal scene to synthesize the foggy images and form the new test set with 63,510 images. As for the validation set, all the 9675 images were used for foggy image generation, resulting in 38,700 images in the new validation set. Table 1 presents the file structure of our synthesized foggy images in FoggyCULane based on the original CULane dataset, while Table 2 shows the comparison between test sets in CULane and FoggyCULane.

4. Experiments

4.1. Dataset

There were 186,786 images in total in our FoggyCULane dataset. To verify the effectiveness of the synthetic dataset in improving the performance of lane detection networks, we used the Spatial CNN (SCNN) network as the baseline method to investigate the datasets with different densities of fog. Other state-of-art methods were also employed to evaluate the improvement of the FoggyCULane dataset. Additionally, 300 hazy images with annotated lane makings in the VIL-100 dataset were employed to evaluate the effectiveness of our proposed framework. Despite of the hazy images in VIL-100, we also collected 182 traffic lane images under real thick foggy weather to test whether the trained SCNN network is adequate for more challenging real foggy scenes. Moreover, the fog synthetization method was also applied to another dataset, i.e., VIL-100, to test its adaptability.

4.2. Evaluation Metrics

To quantitatively evaluate the lane detection results, we refer to the methods in [6], where both the real lane annotation and the predicted lane marking are considered a line area with a width of 30 pixels, and the Intersection over Union (IoU) between them is calculated. In this paper, the IoU threshold is set to 0.5. When the IoU between the predicted lane marking and the real lane is not less than 0.5, the prediction is considered as True Positive (TP), and the opposite is considered as False Positive (FP); accordingly, True Negative (TN) and False Negative (FP) can also be defined.
As mentioned above, the expanded test set of FoggyCULane had 12 different complex scenes, including the newly added three different density foggy scenes. Here, the F1-measure is employed to be the index for quantitative evaluation and is defined as follows:
F 1   score = 2 × Precision × Recall Precision + Recall ,   Precision = T P T P + F P R e c a l l = T P T P + F N

4.3. Implementation Details

SCNN is a spatial convolutional neural network creatively proposed by Xingang Pan et al. [6], which is reported to have the highest lane detection accuracy in the CULane dataset. Here, the model was trained with Stochastic Gradient Descent (SGD) as the optimizer with a base learning rate of 0.01 and momentum of 0.9. Weight decay is set to be 10 4 . The training and testing were undertaken on 8 NVIDIA GTX 2080Ti GPUs and an Intel Xeon E5-2682 v4 CPU. Before training, the images were resized to 800 × 288 for the limited memory of the GPU. All the models were downloaded from public source code with default hyper-parameters and pre-trained as previous works did.

4.4. Effect of FoggyCULane Dataset on SCNN

Table 3 shows the results of five epochs of training of the SCNN network on different datasets, including 12 different scenes. The last three rows are the newly added three foggy scenes with increasing fog densities. The data in each column is acquired from five epochs of neural network trained with different datasets, including the original CULane dataset, the FoggyCULane dataset with three foggy densities of values 2, 3, and 4, and the FoggyCULane dataset with a mixture of the three foggy densities. The values in the table (except the Crossroad row) represent the F1-measure values in percentage for each scene, while the Crossroad row shows FP values.
As shown in Table 3, the models trained with the FoggyCULane dataset practically obtained better lane detection performance than those trained with the original CULane dataset under the three foggy conditions, especially with the highest density ( β   = 4). Figure 6 presents the one of the lane detection results of SCNN trained on different datasets.
Under the heaviest foggy weather condition, the lane detection performance (F1-measure) increased from 11.09 to 89.21 by FoggyCULane. Particularly, the model trained by the FoggyCULane dataset with the mixing of the three densities of foggy weather images achieved the best performance (86.65, 81.53, and 70.41 in F1-measure) in all foggy scenes. This improvement is due to the addition of foggy images to the training set, allowing the neural network to extract features of lane markings under foggy weather for learning. These synthesis haze images contain enough foggy information to satisfy the need for deep learning. Therefore, this study can significantly improve the lane detection performance of neural networks in foggy scenes while not negatively affecting lane detection in other complex scenes. By comparing three different density FoggyCULane datasets ( β = 2, 3, 4), it can be observed that:
  • The dataset with mixed foggy densities has better performance than the dataset with a single foggy density. On the one hand, there are more haze images in the dataset with mixed fog densities, and the neural network can be more exposed to the foggy scene when training. Therefore, it is more sensitive to lane markings under foggy weather. On the other hand, the dataset with mixed fog densities contains fog images of three densities, making the network learn and extract features for lane detection in foggy condition comprehensively during model training.
  • The model trained in the dataset with the corresponding fog density value achieves the best lane detection performance in each foggy scene. This indicates that the features extracted by the network vary with fog densities. Therefore, the dataset with mixed fog densities should be applied to obtain a better performance in practice.

4.5. Effect of FoggyCULane Dataset on Other State-of-Art Methods

To verify the effectiveness of the FoggyCULane dataset on other detection network, four other state-of-art lane detection networks (ENet-SAD [7], ERFNet [8], LaneATT [9], and LaneNet [10]) were also trained with the original CULane dataset and FoggyCULane dataset (including all the densities of fog) respectively, then tested under normal and foggy scenes to compare their performance.
As shown in Table 4, all the models trained by the FoggyCULane dataset have improved performance under foggy conditions compared with the models trained on the original CULane. The test results of these networks are similar to the results of the SCNN network, which also demonstrates the generality and applicability of our method for different lane detection networks.

4.6. Ablation Study

4.6.1. Lane Detection Results in Real Foggy Scene

VIL-100 dataset was collected for lane detection by Zhang et al. [42], which includes 10,000 frames in different real traffic scenarios including normal, crowded, curved road, damaged road, shadows, road markings, dazzle light, haze, night and crossroad, and multiple scenarios may simultaneously occur in a single frame. The VIL-100 dataset contains three hazy scenes (here named as real foggy scene 1, real foggy scene 2, and real foggy scene 3) with 100 images in each scene, and the representative images in each scene are shown in Figure 7. To further evaluate the effectiveness of our proposed framework in improving the performance of the lane detection model in real foggy scenes, 300 real foggy images with annotated lane markings from the VIL-100 dataset were adopted. SCNN model trained on CULane and FoggyCULane ( β = 2 ,   3 ,   4 ) were tested on these foggy images, respectively. Figure 7 also compares the lane detection results from the two models, and Table 5 presents the values of F1-mearsure to evaluate the performance of the models in these foggy images. As shown in Table 5, the performance of SCNN trained on FoggyCUlane was generally improved in three scenes, especially in foggy scene 1 and foggy scene 3, in which the fog is thicker, and the fog is further overlaid with night darkness. The result further indicates that our proposed framework could enhance the ability of lane detection models in real foggy scenes.
Since the fog in the selected images of VIL-100 are commonly thin, to further evaluate whether the trained lane detection model is adequate for real foggy scenes, 182 lane line images containing more challenging real foggy scenes were downloaded and compiled from the Internet. We used the SCNN models trained on the original CULane dataset and the FoggyCULane ( β = 2 ,   3 ,   4 ) dataset to verify the model’s performance on real-world foggy images. These challenging real foggy images have several characteristics:
1.
The fog density in the image varies among images, and the fog density is also not uniform in the same image.
2.
The angle and orientation of the images vary greatly from one another, including images taken from the perspective of roadside pedestrians, road surveillance cameras, and in-vehicle recorders.
3.
Vehicles and pedestrians in the images occlude the lane marks to varying degrees, and the number of lane marks in each image is not the same.
Since these real fog lane images collected from the Internet do not have lane labels, the same strategy as the Crossroad scene in the CULane test set was used here to evaluate the lane detection results using FP values.
Comparing the FP values of the above two models, we found that the performance of the SCNN model trained with the FoggyCULane dataset is significantly better than the other on real foggy days. The FP value increases from 33 to 118, as shown in Table 6, meaning the percentage of images with recognized lane markings increased from 18.13% to 64.84%.
Several representative lane detection results in foggy weather are shown in Figure 8. It can be seen from Figure 8a that for multiple and obstructed lane markings, the model still detects the lane marking blocked by the vehicle based on semantic information. Figure 8b shows the image taken from the perspective of the pedestrian on the roadside. Unlike other images, the lane marking is not located in the central area of the image, with a different angle and direction. Figure 8 shows the case of low light and high fog density, yet still, one of the lanes is detected in this complex scenario. Overall, the SCNN model trained with the FoggyCULane dataset with three different densities of foggy scenes can achieve lane detection for more complex, realistic foggy scenes.

4.6.2. Application of Proposed Framework on VIL-100 Dataset

In this section, we applied our data augmentation framework to the VIL-100 dataset and generated FoggyVIL-100 to evaluate the feasibility of the current method further. We synthesized 5400 foggy images from the normal scenes in the VIL-100 dataset. The generated images contained three levels of simulated fog ( β = 2 ,   3 ,   4 ), with 1800 images at each fog intensity level. Figure 9 shows the results of our generated foggy images on VIL-100 and the corresponding clear images and depth maps. The generated foggy images of all three foggy densities were mixed with the original images to establish a new dataset named FoggyVIL-100 to further test the effectiveness and feasibility of our proposed framework. Table 7 presents the file structure of VIL-100 and FoggyVIL-100, while Table 8 presents the comparison of scenarios in test sets in VIL-100 and FoggyVIL-100. Note that 300 hazy images in the original VIL-100 (real foggy scene 1, real foggy scene 2, and real foggy scene 3) were picked out to evaluate the performance of the lane detection model trained on different datasets.
The SCNN model was trained on the original VIL-100 dataset and FoggyVIL-100 dataset with default hyperparameters, respectively. Table 9 shows the lane detection results (F1-measure) of the SCNN model on the FoggyVIL-100 test set. As shown in Table 9, compared with the SCNN trained on original VIL-100, that trained on FoggyVIL-100 improved F1-measure in all scenes, especially in generated foggy scenes, which were improved from 72.84 to 81.57, 54.00 to 79.22, and 13.57 to 66.19 under three fog densities, respectively. Meanwhile, the model performance in other weather conditions was not degraded. In fact, it mostly improved, which may be the result of the increased scale and enriched features of the trained set, especially for crowed, dazzl light, and night scenes, in which the light conditions are commonly low.
Table 9 shows that the performance of the SCNN model trained on FoggyCULane was improved in the FoggyVIL-100 test set. This suggests that even when the camera position or level of brightness was changed from one dataset to another, the foggy data augmentation approach is still valid to improve model performance.
Moreover, the models trained on VIL-100 and FoggyVIL-100 were also evaluated in the FoggyCULane test set, with the results depicted in Table 10. Table 10 also shows that the performance of the model was improved by training on the augmented dataset, especially for the heaviest simulated foggy scene (from 11.10 to 36.51). Table 11 and Table 12 present the evaluation results in real foggy scenes collected from the original VIL-100dataset and the internet, respectively. As is shown in Table 11, the augmented dataset improved the performance of the SCNN model in three real foggy scenes by 12.41%, 3.35%, and 15.07%, respectively, and the FP values in real foggy images collected from the internet were also improved from 24 to 77 as shown in Table 12. The results in this section further proved the feasibility and effectiveness of our method.

5. Conclusions

In this paper, we focus on improving the lane detection performance of neural networks in foggy conditions. We proposed a data augmentation method to synthesize foggy images from the clear weather condition. A total of 107,451 foggy images are synthesized based on the CULane dataset using our method; thus, the scale of the original CULane dataset is expanded 1.8 times to solve the problem of insufficient lane dataset under foggy weather. Therefore, a new lane detection dataset, FoggyCULane containing three different densities of foggy weather scenarios, was built.
The results of our work indicate that artificially synthesizing images to expand the dataset can significantly improve the performance of lane detection in the corresponding complex scene. Accuracy in the F1-measure of the SCNN model in different densities of foggy weather was improved from 74.65, 51.41, and 11.09 to 86.65, 81.53, and 70.41, respectively. The improvement is especially notable in the higher fog density condition; meanwhile, the expansion of the dataset does not negatively affect the performance of other complex scenes. Aside from the SCNN model, we also evaluated our method on other SOTA networks. All the models trained on the FoggyCULane dataset had improved performance under foggy conditions, demonstrating the generality and applicability of the method.
The performance of our method under real foggy scenes also showed significant improvement, with the F1-measure increasing from 59.13, 57.01, and 40.15 to 66.82, 59.44, and 46.08 in three real foggy scenes, respectively. Besides, this data augmentation method was further applied to another dataset, VIL-100, to test the adaptability of this approach. Similarly, it was found that even when the camera position or level of brightness was changed from one dataset to another, the foggy data augmentation approach is still valid for improving model performance under foggy conditions without degrading the accuracy on other weather conditions. Therefore, the method proposed in this paper can effectively improve the lane detection performance in foggy scenes for both highways and more complex urban roads without introducing extra computation on resource-constrained devices.
In summary, a data augmentation method is proposed in this work to improve the performance of DL-based lane detection algorithms under foggy weather. This work is of great practical significance for the following reasons. Firstly, the foggy image synthesis in our method avoids introducing extra sensors to acquire depth information, which means the method could be easily deployed in vision-based lane detection datasets. Additionally, despite the fact that this work mainly focuses on the lane detection task under foggy weather, it is expected that the proposed method could be further extended to improve object recognition and semantic segmentation tasks under other complex scenes for future study.

Author Contributions

Conceptualization, X.N. and Z.X.; methodology, X.N., Z.X. and W.Z.; software, X.N. and Z.X.; validation, X.N. and W.Z.; formal analysis, X.N. and Z.X.; investigation, X.N. and Z.X.; resources, X.N. and Z.X.; data curation, X.N. and Z.X.; writing—original draft preparation, X.N.; writing—review and editing, X.N., X.D., N.L. and Y.C.; visualization, X.N.; supervision, X.D.; project administration, X.D.; funding acquisition, X.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research was financially supported by National Natural Science Foundation of China under Grant No. 52006137, Shanghai Sailing Program, under Grant No. 19YF1423400, as well as China Postdoctoral Science Funding under Grant No. 2016M600313.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data available on request from the authors.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Urmson, C.; Anhalt, J.; Bagnell, D.; Baker, C.; Bittner, R.; Clark, M.N.; Dolan, J.; Duggins, D.; Galatali, T.; Geyer, C.; et al. Autonomous driving in urban environments: Boss and the Urban Challenge. J. Field Robot. 2008, 25, 425–466. [Google Scholar] [CrossRef] [Green Version]
  2. Khan, M.A.; Sayed, H.E.; Malik, S.; Zia, T.; Khan, J.; Alkaabi, N.; Ignatious, H. Level-5 Autonomous Driving—Are We There Yet? A Review of Research Literature. ACM Comput. Surv. 2022, 55, 1–38. [Google Scholar] [CrossRef]
  3. Huval, B.; Wang, T.; Tandon, S.; Kiske, J.; Song, W.; Pazhayampallil, J.; Andriluka, M.; Rajpurkar, P.; Migimatsu, T.; Cheng-Yue, R. An empirical evaluation of deep learning on highway driving. arXiv 2015, arXiv:1504.01716. [Google Scholar]
  4. Kim, J.; Lee, M. Robust lane detection based on convolutional neural network and random sample consensus. In Neural Information Processing. ICONIP 2014; Springer: Cham, Switzerland, 2014; pp. 454–461. [Google Scholar]
  5. Li, J.; Mei, X.; Prokhorov, D.; Tao, D. Deep neural network for structural prediction and lane detection in traffic scene. IEEE Trans. Neural Netw. Learn. Syst. 2016, 28, 690–703. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Pan, X.; Shi, J.; Luo, P.; Wang, X.; Tang, X. Spatial as deep: Spatial CNN for traffic scene understanding. In Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018. [Google Scholar]
  7. Hou, Y.; Ma, Z.; Liu, C.; Loy, C.C. Learning lightweight lane detection cnns by self attention distillation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 1013–1021. [Google Scholar]
  8. Romera, E.; Alvarez, J.M.; Bergasa, L.M.; Arroyo, R. Erfnet: Efficient residual factorized convnet for real-time semantic segmentation. IEEE Trans. Intell. Transp. Syst. 2017, 19, 263–272. [Google Scholar] [CrossRef]
  9. Tabelini, L.; Berriel, R.; Paixão, T.M.; Badue, C.; De Souza, A.F.; Olivera-Santos, T. Keep your eyes on the lane: Attention-guided lane detection. arXiv 2020, arXiv:2010.12035. [Google Scholar]
  10. Wang, Z.; Ren, W.; Qiu, Q. Lanenet: Real-time lane detection networks for autonomous driving. arXiv 2018, arXiv:1807.01726. [Google Scholar]
  11. Hu, X.; Fu, C.-W.; Zhu, L.; Heng, P.-A. Depth-attentional features for single-image rain removal. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 8022–8031. [Google Scholar]
  12. Liu, T.; Chen, Z.; Yang, Y.; Wu, Z.; Li, H. Lane detection in low-light conditions using an efficient data enhancement: Light conditions style transfer. In Proceedings of the 2020 IEEE Intelligent Vehicles Symposium (IV), Las Vegas, NV, USA, 19 October–13 November 2020; pp. 1394–1399. [Google Scholar]
  13. Tarel, J.-P.; Hautiere, N.; Cord, A.; Gruyer, D.; Halmaoui, H. Improved visibility of road scene images under heterogeneous fog. In Proceedings of the 2010 IEEE Intelligent Vehicles Symposium, La Jolla, CA, USA, 21–24 June 2010; pp. 478–485. [Google Scholar]
  14. Tarel, J.-P.; Hautiere, N.; Caraffa, L.; Cord, A.; Halmaoui, H.; Gruyer, D. Vision enhancement in homogeneous and heterogeneous fog. IEEE Intell. Transp. Syst. Mag. 2012, 4, 6–20. [Google Scholar] [CrossRef] [Green Version]
  15. Sakaridis, C.; Dai, D.; Van Gool, L. Semantic foggy scene understanding with synthetic data. Int. J. Comput. Vis. 2018, 126, 973–992. [Google Scholar] [CrossRef] [Green Version]
  16. Mai, N.A.M.; Duthon, P.; Khoudour, L.; Crouzil, A.; Velastin, S.A. 3D Object Detection with SLS-Fusion Network in Foggy Weather Conditions. Sensors 2021, 21, 6711. [Google Scholar] [CrossRef]
  17. He, Y.; Liu, Z. A feature fusion method to improve the driving obstacle detection under foggy weather. IEEE Trans. Transp. Electrif. 2021, 7, 2505–2515. [Google Scholar] [CrossRef]
  18. Yan, M.; Chen, J.; Xu, J.; Xiang, L. Visibility detection of single image in foggy days based on Fourier transform and convolutional neural network. In Proceedings of the 2nd International Conference on Applied Mathematics, Modelling, and Intelligent Computing (CAMMIC 2022), Kunming, China, 25–27 March 2022; pp. 1440–1445. [Google Scholar]
  19. Schechner, Y.Y.; Narasimhan, S.G.; Nayar, S.K. Instant dehazing of images using polarization. In Proceedings of the 2001 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, CVPR 2001, Kauai, HI, USA, 8–14 December 2001; p. I. [Google Scholar]
  20. Fattal, R. Single image dehazing. ACM Trans. Graph. 2008, 27, 1–9. [Google Scholar] [CrossRef]
  21. Fattal, R. Dehazing using color-lines. ACM Trans. Graph. 2014, 34, 1–14. [Google Scholar] [CrossRef]
  22. Berman, D.; Avidan, S. Non-local image dehazing. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1674–1682. [Google Scholar]
  23. He, K.; Sun, J.; Tang, X. Single image haze removal using dark channel prior. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 33, 2341–2353. [Google Scholar]
  24. Ancuti, C.O.; Ancuti, C. Single image dehazing by multi-scale fusion. IEEE Trans. Image Process. 2013, 22, 3271–3282. [Google Scholar] [CrossRef] [PubMed]
  25. Zhu, Q.; Mai, J.; Shao, L. A fast single image haze removal algorithm using color attenuation prior. IEEE Trans. Image Process. 2015, 24, 3522–3533. [Google Scholar] [CrossRef] [Green Version]
  26. Ren, W.; Liu, S.; Zhang, H.; Pan, J.; Cao, X.; Yang, M.-H. Single image dehazing via multi-scale convolutional neural networks. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; pp. 154–169. [Google Scholar]
  27. Li, B.; Peng, X.; Wang, Z.; Xu, J.; Feng, D. Aod-net: All-in-one dehazing network. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 4770–4778. [Google Scholar]
  28. Cai, B.; Xu, X.; Jia, K.; Qing, C.; Tao, D. Dehazenet: An end-to-end system for single image haze removal. IEEE Trans. Image Process. 2016, 25, 5187–5198. [Google Scholar] [CrossRef] [Green Version]
  29. Li, R.; Zhang, X.; You, S.; Li, Y. Learning to Dehaze from Realistic Scene with A Fast Physics-based Dehazing Network. arXiv 2020, arXiv:2004.08554. [Google Scholar]
  30. Anvari, Z.; Athitsos, V. Dehaze-GLCGAN: Unpaired single image de-hazing via adversarial training. arXiv 2020, arXiv:2008.06632. [Google Scholar]
  31. Wang, Y.; Yan, X.; Wang, F.L.; Xie, H.; Yang, W.; Wei, M.; Qin, J. UCL-Dehaze: Towards Real-world Image Dehazing via Unsupervised Contrastive Learning. arXiv 2020, arXiv:2205.01871. [Google Scholar]
  32. Kumar, B.; Gupta, H.; Sinha, A.; Vyas, O. Lane Detection for Autonomous Vehicle in Hazy Environment with Optimized Deep Learning Techniques. In Proceedings of the International Conference on Advanced Network Technologies and Intelligent Computing, Varanasi, India, 17–18 December 2021; pp. 596–608. [Google Scholar]
  33. Bijelic, M.; Gruber, T.; Mannan, F.; Kraus, F.; Ritter, W.; Dietmayer, K.; Heide, F. Seeing through fog without seeing fog: Deep multimodal sensor fusion in unseen adverse weather. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 11682–11692. [Google Scholar]
  34. Aly, M. Real time detection of lane markers in urban streets. In Proceedings of the 2008 IEEE Intelligent Vehicles Symposium, Eindhoven, The Netherlands, 4–6 June 2008; pp. 7–12. [Google Scholar]
  35. Geiger, A.; Lenz, P.; Urtasun, R. Are we ready for autonomous driving? the kitti vision benchmark suite. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 3354–3361. [Google Scholar]
  36. Cordts, M.; Omran, M.; Ramos, S.; Rehfeld, T.; Enzweiler, M.; Benenson, R.; Franke, U.; Roth, S.; Schiele, B. The cityscapes dataset for semantic urban scene understanding. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 3213–3223. [Google Scholar]
  37. TuSimple. TuSimple Benchmark. Available online: https://github.com/TuSimple/tusimple-benchmark (accessed on 11 July 2017).
  38. Huang, X.; Cheng, X.; Geng, Q.; Cao, B.; Zhou, D.; Wang, P.; Lin, Y.; Yang, R. The apolloscape dataset for autonomous driving. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Salt Lake City, UT, USA, 18–23 June 2018; pp. 954–960. [Google Scholar]
  39. Yu, F.; Chen, H.; Wang, X.; Xian, W.; Chen, Y.; Liu, F.; Madhavan, V.; Darrell, T. Bdd100k: A diverse driving dataset for heterogeneous multitask learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 2636–2645. [Google Scholar]
  40. Behrendt, K.; Soussan, R. Unsupervised labeled lane markers using maps. In Proceedings of the IEEE/CVF International Conference on Computer Vision Workshops, Seoul, Korea, 27–28 October 2019; pp. 832–839. [Google Scholar]
  41. Geyer, J.; Kassahun, Y.; Mahmudi, M.; Ricou, X.; Durgesh, R.; Chung, A.S.; Hauswald, L.; Pham, V.H.; Mühlegg, M.; Dorn, S. A2d2: Audi autonomous driving dataset. arXiv 2020, arXiv:2004.06320. [Google Scholar]
  42. Zhang, Y.; Zhu, L.; Feng, W.; Fu, H.; Wang, M.; Li, Q.; Li, C.; Wang, S. VIL-100: A New Dataset and A Baseline Model for Video Instance Lane Detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 11–17 October 2021; pp. 15681–15690. [Google Scholar]
  43. Harald, K. Theorie der Horizontalen Sichtweite: Kontrast und Sichtweite; Keim Nemnich: Munich, Germany, 1924; Volume 12. [Google Scholar]
  44. McCartney, E.J. Optics of the Atmosphere: Scattering by Molecules and Particles; John Wiley and Sons: New York, NY, USA, 1976. [Google Scholar]
  45. Nayar, S.K.; Narasimhan, S.G. Vision in bad weather. In Proceedings of the Seventh IEEE International Conference on Computer Vision, Kerkyra, Greece, 20–27 September 1999; pp. 820–827. [Google Scholar]
  46. Facil, J.M.; Ummenhofer, B.; Zhou, H.; Montesano, L.; Brox, T.; Civera, J. CAM-Convs: Camera-aware multi-scale convolutions for single-view depth. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 11826–11835. [Google Scholar]
  47. Garg, R.; Bg, V.K.; Carneiro, G.; Reid, I. Unsupervised cnn for single view depth estimation: Geometry to the rescue. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; pp. 740–756. [Google Scholar]
  48. Wang, R.; Pizer, S.M.; Frahm, J.-M. Recurrent neural network for (un-) supervised learning of monocular video visual odometry and depth. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 5555–5564. [Google Scholar]
  49. Chakravarty, P.; Narayanan, P.; Roussel, T. GEN-SLAM: Generative modeling for monocular simultaneous localization and mapping. In Proceedings of the 2019 International Conference on Robotics and Automation (ICRA), Montreal, QC, Canada, 20–24 May 2019; pp. 147–153. [Google Scholar]
  50. Aleotti, F.; Tosi, F.; Poggi, M.; Mattoccia, S. Generative adversarial networks for unsupervised monocular depth prediction. In Proceedings of the European Conference on Computer Vision (ECCV) Workshops, Munich, Germany, 8–14 September 2018; pp. 337–354. [Google Scholar]
  51. Eigen, D.; Puhrsch, C.; Fergus, R. Depth map prediction from a single image using a multi-scale deep network. arXiv 2014, arXiv:1406.228327. [Google Scholar]
  52. Godard, C.; Mac Aodha, O.; Firman, M.; Brostow, G.J. Digging into self-supervised monocular depth estimation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 3828–3838. [Google Scholar]
Figure 1. Typical samples from (a) FRIDA [13] and (b) FRIDA2 [14], (c) is one original clear image from Cityscapes [36], (d) displays its depth, and (e) is the simulated foggy image from Foggy Cityscapes [15]. (fh) show the clear image, depth map, and generated foggy image from our proposed dataset, respectively.
Figure 1. Typical samples from (a) FRIDA [13] and (b) FRIDA2 [14], (c) is one original clear image from Cityscapes [36], (d) displays its depth, and (e) is the simulated foggy image from Foggy Cityscapes [15]. (fh) show the clear image, depth map, and generated foggy image from our proposed dataset, respectively.
Sensors 22 05210 g001
Figure 2. The pipeline to improve the performance of lane detection model in foggy scene with our method.
Figure 2. The pipeline to improve the performance of lane detection model in foggy scene with our method.
Sensors 22 05210 g002
Figure 3. Examples of depth estimation and foggy image synthesis, (a) original images from CULane, (b) depth maps predicted by Monodepth2, and (c) synthetic foggy images.
Figure 3. Examples of depth estimation and foggy image synthesis, (a) original images from CULane, (b) depth maps predicted by Monodepth2, and (c) synthetic foggy images.
Sensors 22 05210 g003
Figure 4. The overall structure of CULane dataset, (a) the number of images in each set, (b) proportion of each scene.
Figure 4. The overall structure of CULane dataset, (a) the number of images in each set, (b) proportion of each scene.
Sensors 22 05210 g004
Figure 5. Example of synthetic foggy images under three different densities ( β = 2 ,   3 ,   4 ), together with their clear image, with the thickness of fog increasing with β .
Figure 5. Example of synthetic foggy images under three different densities ( β = 2 ,   3 ,   4 ), together with their clear image, with the thickness of fog increasing with β .
Sensors 22 05210 g005
Figure 6. The foggy lane detection results of SCNN trained on different datasets show the real lane markings in red and the predicted lane markings in green.
Figure 6. The foggy lane detection results of SCNN trained on different datasets show the real lane markings in red and the predicted lane markings in green.
Sensors 22 05210 g006
Figure 7. The lane detection results of SCNN tested in real foggy scenes from VIL-100 dataset. Red lines denote the real lane markings and green lines denote the predicted lane markings.
Figure 7. The lane detection results of SCNN tested in real foggy scenes from VIL-100 dataset. Red lines denote the real lane markings and green lines denote the predicted lane markings.
Sensors 22 05210 g007
Figure 8. The lane detection results of SCNN tested in real foggy images acuiqred online without labelling: the first row shows the results of SCNN trained on CULane and the second row shows the results of SCNN trained on FoggyCULane. Green lines denote detected lane markings by the model.
Figure 8. The lane detection results of SCNN tested in real foggy images acuiqred online without labelling: the first row shows the results of SCNN trained on CULane and the second row shows the results of SCNN trained on FoggyCULane. Green lines denote detected lane markings by the model.
Sensors 22 05210 g008
Figure 9. Examples of depth estimation and generated foggy images ( β = 3 ) from VIL-100, (a) original images, (b) predicted depth maps, and (c) synthesized foggy images.
Figure 9. Examples of depth estimation and generated foggy images ( β = 3 ) from VIL-100, (a) original images, (b) predicted depth maps, and (c) synthesized foggy images.
Sensors 22 05210 g009
Table 1. The overall file structure of the Foggy CULane dataset compared with the original CULane dataset, β is set to 2, 3, and 4 for simulated fog, and β = 1 represents the original clear image.
Table 1. The overall file structure of the Foggy CULane dataset compared with the original CULane dataset, β is set to 2, 3, and 4 for simulated fog, and β = 1 represents the original clear image.
β Train Set ImagesValidation Set ImagesTest Set ImagesTotal Images
CULane188,880967534,680133,235
Foggy CULane188,880967534,680133,235
216,5329675961035,817
316,5329675961035,817
416,5329675961035,817
Table 2. Comparison between test sets of CULane and FoggyCULane   ( β = 2 ,   3 ,   4 ) .
Table 2. Comparison between test sets of CULane and FoggyCULane   ( β = 2 ,   3 ,   4 ) .
ScenariosCULane
Test Set
FoggyCULane
Test Set
Normal96109610
Crowded81158115
Night70407040
No line40584058
Shadow936936
Arrow900900
Dazzle light486486
Curve415415
Crossroad31203120
Foggy_beta209610
Foggy_beta309610
Foggy_beta409610
Table 3. Evaluation results (F1-measure) of SCNN tested in different scenes, with the model trained with different datasets.
Table 3. Evaluation results (F1-measure) of SCNN tested in different scenes, with the model trained with different datasets.
CategorySCNN
Trained on
Origin CULane
SCNN Trained on
FoggyCULane
with   β   =   2
SCNN Trained on
FoggyCULane
with   β   =   3
SCNN Trained on
FoggyCULane
with   β   =   4
SCNN Trained on
FoggyCULane
with   β   =   2 ,   3 ,   4
Normal89.2189.8289.8189.2090.09
Crowded68.1168.3068.6367.9369.53
Night64.8565.4964.0064.2465.08
No line43.2543.2143.2243.0242.80
Shadow59.2758.3161.4056.5763.77
Arrow83.2984.5184.3083.8283.66
Dazzle light59.9365.6464.0758.7461.16
Curve62.6565.6166.5162.5162.09
Crossroad (FP)27042679315523863004
Foggy_beta274.6586.6385.5883.7486.65
Foggy_beta351.4179.5380.0777.5681.53
Foggy_beta411.0960.3365.2765.9170.41
Table 4. Evaluation results (F1-measure) of other lane detection models, trained with CULane and FoggyCULane (β = 2, 3, 4), respectively.
Table 4. Evaluation results (F1-measure) of other lane detection models, trained with CULane and FoggyCULane (β = 2, 3, 4), respectively.
CategoryENet-SADERFNetLaneATTLaneNet
CULaneFoggy
CULane
CULaneFoggy
CULane
CULaneFoggy
CULane
CULaneFoggy
CULane
Normal88.5289.6487.4688.5785.0687.0388.1388.95
Foggy_beta271.8685.2169.5684.3469.2084.9965.2483.51
Foggy_beta350.6280.3549.7379.4247.1478.2141.4777.32
Foggy_beta412.1371.3210.2365.2810.9366.459.8259.48
Table 5. Lane detection results (F1-measure averaged from 100 images) of SCNN on real foggy scenes from VIL-100 dataset, with the models trained on CULane and FoggyCULane, respectively.
Table 5. Lane detection results (F1-measure averaged from 100 images) of SCNN on real foggy scenes from VIL-100 dataset, with the models trained on CULane and FoggyCULane, respectively.
SCNN Trained on CULaneSCNN Trained on FoggyCULane
Real foggy scene 159.1366.82
Real foggy scene 257.0159.44
Real foggy scene 340.1546.08
Table 6. Lane detection results (FP) of SCNN in real foggy scene.
Table 6. Lane detection results (FP) of SCNN in real foggy scene.
SCNN Trained on CULaneSCNN Trained on FoggyCULane
FP33118
Table 7. The overall file structure of the FoggyVIL-100 dataset compared with the original VIL-100 dataset, β is set to 2, 3, and 4 for simulated fog, and β = 1 represents the original clear image.
Table 7. The overall file structure of the FoggyVIL-100 dataset compared with the original VIL-100 dataset, β is set to 2, 3, and 4 for simulated fog, and β = 1 represents the original clear image.
β Train Set
Images
Test Set
Images
Total
Images
VIL-100 β = 1 8000200010,000
Foggy
VIL-100
β = 1 8000200010,000
β = 2 14004001800
β = 3 14004001800
β = 4 14004001800
Table 8. Comparison between test sets of VIL-100 and FoggyVIL-100   ( β = 2 ,   3 ,   4 ) .
Table 8. Comparison between test sets of VIL-100 and FoggyVIL-100   ( β = 2 ,   3 ,   4 ) .
ScenariosVIL-100
Test Set
FoggyVIL-100
Test Set
Normal400400
Crowded700700
Curved road700700
Damaged road100100
Shadows200200
Road markings400400
Dazzle light200200
Night100100
Crossroad100100
Foggy_beta20400
Foggy_beta30400
Foggy_beta40400
Table 9. Lane detection results (F1-measure) of SCNN tested on different scenes of the FoggyVIL-100 test set, with the model trained with different datasets.
Table 9. Lane detection results (F1-measure) of SCNN tested on different scenes of the FoggyVIL-100 test set, with the model trained with different datasets.
CategorySCNN Trained
on Origin
VIL-100
SCNN Trained on FoggyVIL-100
Beta = 2, 3, 4
SCNN Trained
on Origin
CULane
SCNN Trained on FoggyCULane
Beta = 2, 3, 4
Normal84.3187.4178.9482.03
Crowded72.6374.4759.8961.22
Curved road65.0965.1361.3962.03
Damaged road40.6341.8641.9842.21
Shadows43.0552.3549.5654.32
Road markings76.2177.4871.9272.01
Dazzle light55.0556.3156.2557.64
Night56.4556.7058.7158.90
Crossroad60.6860.4160.2963.76
Foggy_beta272.8481.5762.8173.25
Foggy_beta354.0079.2250.0369.89
Foggy_beta413.5766.1910.4255.14
Table 10. Lane detection results (F1-measure) of SCNN tested on simulated foggy scenes from FoggyCULane dataset, with the model trained on VIL-100 dataset.
Table 10. Lane detection results (F1-measure) of SCNN tested on simulated foggy scenes from FoggyCULane dataset, with the model trained on VIL-100 dataset.
CategorySCNN
Trained on Origin
VIL-100
SCNN Trained
on FoggyVIL-100
Beta = 2, 3, 4
Foggy_beta252.3557.63
Foggy_beta348.5851.26
Foggy_beta411.1036.51
Table 11. Lane detection results (F1-measure averaged from 100 images) of SCNN on real foggy scenes from VIL-100 dataset, with the models trained on VIL-100 and FoggyVIL-100, respectively.
Table 11. Lane detection results (F1-measure averaged from 100 images) of SCNN on real foggy scenes from VIL-100 dataset, with the models trained on VIL-100 and FoggyVIL-100, respectively.
SCNN Trained on VIL-100SCNN Trained on FoggyVIL-100
Real foggy scene 160.0267.45
Real foggy scene 253.8155.68
Real foggy scene 338.7444.59
Table 12. Lane detection results of SCNN on real foggy scenes collected from internet.
Table 12. Lane detection results of SCNN on real foggy scenes collected from internet.
SCNN Trained on VIL-100SCNN Trained on FoggyVIL-100
FP2477
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Nie, X.; Xu, Z.; Zhang, W.; Dong, X.; Liu, N.; Chen, Y. Foggy Lane Dataset Synthesized from Monocular Images for Lane Detection Algorithms. Sensors 2022, 22, 5210. https://doi.org/10.3390/s22145210

AMA Style

Nie X, Xu Z, Zhang W, Dong X, Liu N, Chen Y. Foggy Lane Dataset Synthesized from Monocular Images for Lane Detection Algorithms. Sensors. 2022; 22(14):5210. https://doi.org/10.3390/s22145210

Chicago/Turabian Style

Nie, Xiangyu, Zhejun Xu, Wei Zhang, Xue Dong, Ning Liu, and Yuanfeng Chen. 2022. "Foggy Lane Dataset Synthesized from Monocular Images for Lane Detection Algorithms" Sensors 22, no. 14: 5210. https://doi.org/10.3390/s22145210

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop