Next Article in Journal
Depth Estimation for Integral Imaging Microscopy Using a 3D–2D CNN with a Weighted Median Filter
Previous Article in Journal
Convolutional Neural Networks and Heuristic Methods for Crowd Counting: A Systematic Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

GAN-Based LiDAR Translation between Sunny and Adverse Weather for Autonomous Driving and Driving Simulation

1
Emerging Design and Informatics Course, Graduate School of Interdisciplinary Information Studies, The University of Tokyo, 4 Chome-6-1 Komaba, Meguro City, Tokyo 153-0041, Japan
2
Mitsubishi Heavy Industries Machinery Systems Ltd., 1-1, Wadasaki-cho 1-chome, Hyogo-ku, Kobe 652-8585, Japan
3
Mitsubishi Heavy Industries Ltd., 1-1, Wadasaki-cho 1-chome, Hyogo-ku, Kobe 652-8585, Japan
4
The Institute of Industrial Science (IIS), The University of Tokyo, 4 Chome-6-1 Komaba, Meguro City, Tokyo 153-0041, Japan
*
Authors to whom correspondence should be addressed.
Sensors 2022, 22(14), 5287; https://doi.org/10.3390/s22145287
Submission received: 19 May 2022 / Revised: 9 June 2022 / Accepted: 28 June 2022 / Published: 15 July 2022
(This article belongs to the Section Vehicular Sensing)

Abstract

:
Autonomous driving requires robust and highly accurate perception technologies. Various deep learning algorithms based on only image processing satisfy this requirement, but few such algorithms are based on LiDAR. However, images are only one part of the perceptible sensors in an autonomous driving vehicle; LiDAR is also essential for the recognition of driving environments. The main reason why there exist few deep learning algorithms based on LiDAR is a lack of data. Recent translation technology using generative adversarial networks (GANs) has been proposed to deal with this problem. However, these technologies focus on only image-to-image translation, although a lack of data occurs more often with LiDAR than with images. LiDAR translation technology is required not only for data augmentation, but also for driving simulation, which allows algorithms to practice driving as if they were commanding a real vehicle, before doing so in the real world. In other words, driving simulation is a key technology for evaluating and verifying algorithms which are practically applied to vehicles. In this paper, we propose a GAN-based LiDAR translation algorithm for autonomous driving and driving simulation. It is the first LiDAR translation approach that can deal with various types of weather that are based on an empirical approach. We tested the proposed method on the JARI data set, which was collected under various adverse weather scenarios with diverse precipitation and visible distance settings. The proposed method was also applied to the real-world Spain data set. Our experimental results demonstrate that the proposed method can generate realistic LiDAR data under adverse weather conditions.

1. Introduction

For autonomous driving, the most important element is the perception of the surrounding environment through various sensors. Representative perception tasks include object detection [1,2], segmentation [3,4,5] and depth estimation [6,7,8]. For image processing, it was recently shown that deep learning approaches can provide high performance and accuracy in perception tasks. However, the performance of perception tasks using other sensors generally does not follow that obtained when using a camera. The most common reason for this is difficulties in data collection. Most deep learning-based methods require large amounts of data and annotations. However, other sensors (e.g., LiDAR and radar), are relatively expensive and, in order to collect relevant data, a massive amount of human resources and time is required. These reasons have led to a lack of data for these sensors. Furthermore, it is difficult to collect data for all traffic scenes. Traffic scenes are diverse, depending on the traffic volume, location, time of day, weather, and so on. As such, a human cannot collect and annotate data for all relevant situations.
With the advent of generative adversarial networks (GANs) [9], research in the translation field has been expanded in various ways. Translation methods can be mainly divided into two types: supervised and unsupervised methods. Although unsupervised translation methods can achieve powerful performance, their naturalness is less than that of supervised methods. General GAN-based methods focus on image translation; however, LiDAR also provides significant sensor data for autonomous driving. There exist some translation methods considering LiDAR characteristics. However, their main goal is commonly data augmentation, such as translation between simulated LiDAR and real LiDAR [10]. Thus, those methods cannot handle the translation of LiDAR under various weather conditions and are hard to apply to driving simulators.
Driving simulators [11,12,13] have recently been on the rise in the autonomous driving field. They allow people to practice driving as if they were commanding a real vehicle. They have emerged as a potential solution to the need for more data while avoiding the time, cost, and safety issues of current methods. However, the simulated data are unnatural, compared to the actual data, and have insufficient performance for actual driving, especially in terms of LiDAR data. As LiDAR is expensive and the collection of LiDAR data requires a massive amount of human resources and time, the existing LiDAR data are insufficient for learning. Furthermore, LiDAR data are easily influenced by adverse weather, such as rain and fog. However, to the best of our knowledge, no empirical method that can reflect the changes in various weather conditions has been introduced.
In this paper, we propose a GAN-based LiDAR translation method that can deal with weather changes. Figure 1 shows the purpose of this study. In Figure 1, we only show the translation from sunny to rainy or foggy. However, we also consider the opposite translation, because there are places where only data with severe weather, such as snow or rain, exist. In those places, the translation from adverse weather to sunny is necessary for data augmentation and verifying algorithms. For the first step, we need to transform LiDAR data as input to the proposed network. To find the optimal data format which gives the best performance, we show the results of various trial-and-error experiments depending on various formats. In addition, we modify the network architecture to handle the LiDAR characteristics, which are different from those of images. The proposed method also uses the JARI data set, which was collected under varying adverse weather conditions, such as precipitation and visible distance. Thus, we can parametrically evaluate the proposed method and demonstrate its performance, in terms of translating sunny to adverse weather.
The contributions of the proposed method are as follows:
  • We present the first LiDAR-to-LiDAR translation method based on an empirical method that deals with various adverse weather conditions.
  • For the proposed method, we suggest the optimal format for LiDAR translation through several trial-and-error experiments and comparing the results on a case-by-case basis.
  • The proposed method deals with translations between sunny and adverse weather, such as rain and fog, with high performance.
  • The proposed method is the first LiDAR-to-LiDAR translation which was evaluated parametrically using the JARI data set, which was collected under varying adverse weather conditions, such as precipitation and visible distance.
The remainder of this paper is organized into the following sections. Section 2 introduces the related work in the literature, such as LiDAR translation to handle various weather conditions, LiDAR translation for data augmentation, and that focused on various translation methods. Next, Section 3 presents our proposed method. Then, Section 4 provides the experimental results. Finally, we give our conclusions and discuss directions for future work in Section 5.

2. Related Work

2.1. LiDAR Translation to Handle Various Weather Conditions

LiDAR is easily affected by adverse weather conditions, such as rain and fog. To determine the influence of weather conditions, many researchers have carried out various studies. In the representative research introduced by Isaac et al. [14] in 2001, the influences of fog and haze on wireless communications were investigated. Furthermore, Rasshofer et al. [15], in 2011, investigated the influence of weather phenomena on automotive LiDAR systems. In recent years, there have been many other works which mentioned the degradation of LiDAR data under different adverse weather conditions [16,17,18,19,20,21,22,23]. Artificial fog simulations are mostly limited to image-based methods. Sakaridis et al. [24] created a foggy version of Cityscapes [25] for semantic segmentation. Furthermore, Hahner et al. [26] leveraged the depth information given in the original data set to improve the performance in the foggy version of the purely synthetic Synscapes [27] data set. Sakaridis et al. also released the ACDC data set [28], which provides semantic pixel-level annotations under adverse conditions. Bijelic et al. [29] recently proposed a first-order approximation for simulating fog in an automotive LiDAR setting. However, their method has a limitation: the simulation only aims to reproduce measurements in a 30 m long fog chamber. The authors in [30,31,32] created LiDAR simulators to deal with adverse weather conditions based on physical models; however, their methods are limited, in that it is impossible to cover all parametric variations in the real world, due to characteristics of the physical models.

2.2. LiDAR Translation for Data Augmentation

There exist some generative networks for LiDAR translation for data augmentation in various fields, such as autonomous driving and aerial vehicles [33]. The representative study for LiDAR data augmentation is [10], in which real LiDAR data were generated from simulated LiDAR data by employing CycleGAN [34] to learn the mapping. They considered various LiDAR inputs, such as Bird-eye View (2D BEV) and Polar-Grid Map (2D PGM), to translate between the CARLA [11] and KITTI data sets [35]. However, it cannot formulate the optimal LiDAR input format and also cannot handle weather changes, such as rain and fog. Additionally, there is the LiDAR translation research between LiDAR and radar for data augmentation [36]. PU-GAN [37] learns a rich variety of point distributions from the latent space and up-samples points by constructing an up–down–up expansion unit in the generator. Additionally, a self-attention unit was formulated to enhance feature integration. Its main goal is the enhancement of LiDAR data resolution, such that it also cannot deal with translation between sunny and adverse weather.

2.3. Generative Adversarial Networks

The first neural network to generate new data, proposed by Goodfellow et al., was the generative adversarial network (GAN) [9]. Numerous studies have researched derivatives of the GAN. One of the representative generative networks is the WGAN [38], which deals with the data imbalance problem caused by oversampling majority classes. It solves this problem by computing the Wasserstein distance between the distribution of the generated samples and the training data. LSGAN [39] is the first generative network applying MSE loss as the loss function, which improves learning stability. InfoGAN [40] learns interpretable representations by adding latent variables to the input. Conditional GAN (cGAN) [41] explicitly generates the output to be generated by utilizing the label information of the training data. DCGAN [42] utilizes convolutional layers to enhance the expressiveness of the generator and discriminator, in contrast to the original GAN, which was composed of MLPs. SAGAN [43] applies the self-attention mechanism to the structure of the GAN to reflect global features. Pix2Pix [44] utilizes cGAN [41] to learn a mapping from input to output. Similar methods with supervised data for the output have been used for many tasks; however, they cannot deal with the problem of no supervised data and require paired data in a similar environment. CycleGAN [34], DiscoGAN [45], and DualGAN [46] perform domain translation by applying cycle consistency. They do not require paired data. In UNIT [47] and CoGAN [48], two domains are assumed to share a common latent space. Not only these methods, but also several other studies have been proposed to deal with multi-domain translation. Although these methods have achieved powerful performance, most of the generative networks focus only on image translation.
In this work, we present the first LiDAR translation method that can deal with translation between sunny and adverse weather based on an empirical approach, not a physical approach.

3. Proposed Method

In this section, we introduce the proposed method. The proposed method is based on CycleGAN [34]. Our final purpose is to generate realistic LiDAR data under various weather conditions. We first determined the most optimal LiDAR data format for training. As the original CycleGAN is intended for image translation, considering the LiDAR data format is a necessary step. There are many elements in LiDAR data. In the proposed method, we utilize LiDAR distance and intensity data for translation. In contrast to the RGB values in a typical image, distance and intensity data of LiDAR are not limited data. Furthermore, the scale of the introduced LiDAR data format is totally different to that of images. To deal with these issues, we designed a network based on the existing CycleGAN structure.

3.1. Setup

An overview of the proposed model is presented in Figure 2. We denote the real inputs (e.g., sunny, rainy, and foggy) in the source and target domains as X a and X b , respectively. Let G A B = G A B _ E · G A B _ D and G B A = G B A _ E · G B A _ D be the source-to-target generator and the target-to-source generator, respectively. In G A B and G B A , G A B _ E and G B A _ E are the encoder parts for extracting features from each input, while G A B _ D and G B A _ D indicate the decoders to generate each output. The two discriminators D A B and D B A determine whether a generated image is true or synthetic in the source and target domain, respectively.

3.2. Architecture

As the original CycleGAN [34] adopted the ResNet architecture [49], the network has nine residual blocks. While the network with nine residual blocks showed brilliant performance for image translation, it is not suitable for LiDAR translation. Through several experiments, the most optimal architecture for LiDAR translation was found to include a generator with four residual blocks and a discriminator with one residual block. In Section 4, we show the results of several experiments, depending on the number of residual blocks. Additionally, the original CycleGAN adopted the tanh activation function to handle limited RGB values (0–256). As the distance and intensity values in LiDAR data are infinite values, the proposed method utilized the rectified linear unit (ReLU) activation function, instead of the tanh activation function.
The ResNet-based generator in [34] contains two downsizing processes through convolutions. While downsizing has the merit of obtaining global features, it may lose local information. To overcome this problem, the proposed method adopts skip connections [50] in the two downsizing processes. Skip connections can maintain local region information by combining information from lower layers and deeper layers.

3.3. Training

Training of the proposed method was highly similar to that of the original CycleGAN [34]. We trained two generators G A B , G B A and two discriminators D A B , D B A for LiDAR translation between sunny and adverse weather (e.g., rainy or foggy). In the proposed method, the generators G A B , G B A and discriminators D A B , D B A are trained using the following multiple losses.
In general, a standard GAN [9] contains two networks: a generator and a discriminator. These two networks are trained using the Min-Max game framework. The discriminator distinguishes whether the output, which is produced by the generator, is real or synthetic. In this work, we utilize the standard adversarial loss:
L a d v 1 = E x b X b [ log ( D A B ( x b ) ) ] + E x a X a [ log ( 1 D A B ( G ( x a ) ) ) ] ,
L a d v 2 = E x a X a [ log ( D B A ( x a ) ) ] + E x b X b [ log ( 1 D B A ( F ( x b ) ) ) ] .
In unpaired translation, mode collapse is a critical problem, which occurs when a generator cannot generate data samples as diverse as the training data distribution. To deal with this issue, the cycle consistency loss is used, which implicitly enforces diversity. G B A ( G A B ( x a ) )     x a and G A B ( G B A ( x b ) )     x b are obtained by calculating the L 1 loss between the original and reconstructed data. This loss is defined as
L c y c = E x a X a [ | | G B A ( G A B ( x a ) ) x a | | 1 ] + E x b X b [ | | G A B ( G B A ( x b ) ) x b | | 1 ] .
The identity loss is generally utilized to ensure that output is unchanged when data from the target domain are input to the generator:
L i d = E x a X a [ | | G B A ( x a ) x a | | 1 ] + E x b X b [ | | G A B ( x b ) x b | | 1 ] .
The final loss is given as follows:
L t o t a l = L a d v 1 + L a d v 2 + λ c y c L c y c + λ i d L i d ,
where λ c y c and λ i d are hyperparameters.

4. Experimental Results

For the following experiments, a 2D LiDAR representation was utilized. It is the Polar-Grid map (i.e., 2D PGM), which represents the LiDAR 3D point cloud as a 2D grid, as shown in Figure 3 [10]. The figures show the 2D PGMs for both distance and intensity in the JARI environment where the data are collected. The representation is obtained by encoding both channels and the ray step angle of the LiDAR sensor. The 3D point cloud can be reconstructed from the PGM 2D representation, given enough horizontal angular resolution. There also exists another 2D LiDAR representation, called the Bird-eye View (2D BEV), where the 3D point cloud is projected from above. As the 2D BEV loses height information and the 3D point cloud cannot be reconstructed from it, we do not utilize it in the proposed method.
We experimented with three settings. In the first experiment, we compared the results depending on various LiDAR input formats and architectures, to formulate the most optimal training configuration. Secondly, we experimented with the proposed method depending on skip-connections [50] to show the effectiveness of the proposed method. Finally, we experimented with the proposed method on data collected in the JARI laboratory. The JARI laboratory can control the various weather conditions, such as rainy and foggy, such that we could objectively carry out our experiment.

4.1. Data Sets

We mainly utilized the JARI data set, as well as the Spain data set. The JARI data are collected in the JARI driving laboratory, where it can control degrees of various weathers. For the JARI data set, sunny data are divided into dry and wet data, indicating the surface condition of the road. Rain data of the JARI data set are composed of the three settings to deal with precipitation: 30 mm/h, 50 mm/h, and 80 mm/h. Fog data of that also contain three settings to handle the degree of fog: 15 m, 30 m, and 80 m visible distance. These weather data include day and night data, along with two vehicle velocities: 5 km/h and 40 km/h. There are various obstacles in the environments where the data were collected, including real vehicles, real pedestrians, and white lines. Table 1 summarizes the configuration of the JARI data set. The JARI data set was utilized to prove that the proposed method can perform stable translation to adverse weather conditions.
We also utilized the Spain data set to prove that the proposed method can handle various real-world environments. Since the Spain data were collected on real roads in Spain, it cannot control degrees of various weathers. Additionally, due to the climatic characteristics of the country, the Spain data set does not include any fog data. Figure 4 shows examples of the JARI and Spain data sets. The JARI LiDAR data were collected by an Ouster-64 (i.e., Ouster LiDAR with 64 vertical channels), while those in Spain were obtained by an Ouster-128 (i.e., Ouster LiDAR with 128 vertical channels). Thus, the data configurations are slightly different. However, we designed the architectures separately, depending on each LiDAR data set.

4.2. Experiment 1: The Most Optimal Input Format and Architecture

For image-to-image translation, inputs consisting of RGB values are common. However, LiDAR data do not contain RGB values and there is no common format for LiDAR training. Thus, we conducted an experiment to determine the most optimal input format and architecture.
LiDAR data are different from image data. While image data contain limited values (e.g., RGB values), LiDAR data consist of unlimited values (e.g., distance, intensity, and reflectivity). In the proposed method, we utilize distance and intensity in LiDAR data, as these values are the most affected by weather and are important for the LiDAR representation. As mentioned above, we utilized the 2D PGM. The input array size was equal to (64,1024,2) for the JARI data set and (128,1024,2) for the Spain data set. The first channel denotes the number of vertical channels, and the second channel indicates the number of horizontal channels. The last channel indicates the number of values, which means distance and intensity values in this study. The distributions of distance and intensity values were 0–250 m and 0–3700, respectively, in the JARI data set.
Table 2 and Table 3 show the results depending on several distances and intensity pairs, with Sunny(dry)–Day and Fog(15 m)–Day in the JARI data set. The result was obtained by summing the errors between the ground truth and the generated result. The equation is given as follows:
X d i s t a n c e = 1 a l l D G D T ,
X i n t e n s i t y = 1 a l l I G I T ,
where X d i s t a n c e and X i n t e n s i t y are the sum of distance errors and intensity errors, respectively; D G and I G are distance and intensity values of the ground truth, respectively; and we define D T and I T as the distance and intensity values in the generated result, respectively.
As shown in Table 2 and Table 3, the intensity/15 result outperformed the others. There are two possible reasons for such a result. First, the distance–intensity/15 pair has a distribution highly similar to that of RGB values. As the architecture of the proposed method is based on the CycleGAN [34], which was originally designed for image translation, the results with that pair were superior. Second, when the intensity value is high, the intensity error results (with square root and log) are also likely to increase. Even though the square root and log values are small, the values restored to original intensity could be high, especially when considering high-intensity values.
In the original CycleGAN [34] for image translation, the ResNet block [49] is utilized to extract features in the down-sampling of generator and discriminator. There are nine ResNet blocks in the generator and four ResNet blocks in the discriminator for image translation. As this configuration was designed for image translation, we experimented and compared the results when changing the number of ResNet blocks in each generator and discriminator, as shown in Table 4. We utilized the Sunny(dry)–Day and Fog(80 m)–Day data in the JARI data set. The result was obtained by summing the errors between the ground truth and the generated result, the same as in Table 2 and Table 3. Table 4 indicates that the most optimal architecture for LiDAR translation in the JARI data set was found to be four ResNet blocks for the generator and one ResNet block for the discriminator.

4.3. Experiment 2: Advantages of Skip-Connections

In this experiment, we evaluated the effectiveness of skip-connections [50] for LiDAR translation. Skip-connections are connections in deep neural networks that feed the output of a particular layer to later layers in the network. Skip-connections are normally utilized in image training to keep features of lower layers and output higher quality. We expect that these connections allow the proposed method to generate a higher resolution output.
We compared the results obtained with and without skip-connections, as shown in Table 5. We utilized the Sunny(dry)–Day and Fog(80 m)–Day data in the JARI data set. The results were obtained by summing the errors between the ground truth and the generated result. The results in Figure 5 are the reconstructed 3D points, as seen above. The green points indicate the reconstructed 3D points. Table 5 and Figure 5 show that the results with skip-connections can generate LiDAR rings clearly, while, without skip-connections, it is hard to represent LiDAR rings (see the red rectangles in Figure 5 for comparison). A LiDAR ring is an arrangement of LiDAR points, which form a ring. Through this experiment, the proposed method with skip-connections outperformed that without skip-connections.

4.4. Experiment 3: LiDAR Translation to Handle Adverse Weathers

In the third experiment, we tested the ability of the proposed method to translate between sunny and adverse weather conditions utilizing the JARI and Spain data sets. The experiment with the JARI data set was carried out for parametrical evaluation, depending on the various degrees of adverse weather conditions. Figure 6 and Figure 7 show example results of translation between sunny and adverse weather conditions using the JARI data set. These results are reconstructed 3D points, as seen above. The green points are the reconstructed 3D points. The images on the left show the real data for the source domains, while those on the right are the generated data for the target domains. Table 6 provides the distance and intensity error results for translation. The results were obtained by summing the errors between the ground truth and the generated result, similarly to the other experiments.
The experiment with the Spain data set was carried out to obtain evidence that the proposed method applies to real-world scenarios. Figure 8 shows an example of results using the Spain data set. These results are also reconstructed 3D points, as seen above. As in Figure 6 and Figure 7, the reconstructed 3D points are represented by green points. The images on the left are the real data for the source domains, while those on the right are the generated data for the target domains. The result for the synthetic rain data was obtained by translation from the real day data, as denoted by the red arrow. In addition, the synthetic day data were generated by translation from the real rain data, denoted by the blue arrow. As the Spain data set does not include fog data, due to weather conditions of the country, we only considered translation between sunny and rainy conditions.
As shown in Figure 6, Figure 7 and Figure 8, the proposed method can translate successfully between sunny and adverse weather conditions. By comparing the real data of day and night for all weathers (e.g., the real data pairs of Sunny(dry)/Day–Sunny(dry)/Night and Fog(15 m)/Day–Fog(15 m)/Night) in Figure 6 and Figure 7, we can confirm that the LiDAR data between day and night are nearly the same. Although the timestamps of the real day and night data are slightly different, the arrangement and intensity values of the LiDAR points are almost identical. Through this confirmation, we can also prove that LiDAR is almost not influenced by darkness, according to the basic characteristic of LiDAR. The proposed method can easily translate from the day(dry) to day(wet), as there is no large difference between day(dry) and day(wet). However, the proposed method generated some wrong points when translating between sunny and rainy. This phenomenon also occurred when translating between sunny and foggy. This is because the proposed method may determine high errors in distance and intensity when the difference between the source and target domain is extremely large, such as when translating between rainy and foggy. More wrong points are generated farther away from the position of the LiDAR as the distance and intensity errors are high at those positions, particularly in the translation between rainy and foggy. This is the key limitation of the proposed method.
The proposed method still reflects the changes in the phenomena according to various weather conditions. In particular, the proposed method can successfully express the LiDAR rings, which take the form of a donut shape, in the translation between sunny and foggy conditions. These rings occur when a vehicle passes through fog, as shown in Figure 7. In addition, when it rains, distant points are erased, due to the visible distance of LiDAR. The proposed method also represents this phenomenon clearly when translating from sunny to rainy. Finally, we also showed that the proposed method can deal with real-world environments through the example results provided in Figure 8. As there was no large difference between the day and rain data, the translation results showed no noticeable difference. We further confirmed that the proposed method is applicable to real-world scenarios.

4.5. Discussion

In this section, we provide some qualitative assessments of the proposed method. In the first experiment, we compared the translation results of the proposed method, depending on various LiDAR formats and architectures. Through several trials, we confirmed that the proposed method performed the best when the distance and intensity values were similar to RGB values during training. This is also suitable for training, under the condition that the number of ResNet blocks is four for the generator and one for the discriminator. In experiment 2, we showed the effectiveness of skip connections in the proposed model. We showed that the proposed model can prevent the loss of local information in the encoder–decoder model generator. The results show that the proposed method is effective for preserving small regions, such as LiDAR rings. In the third experiment, we showed that the proposed method can perform a successful translation between sunny and adverse weather conditions. In particular, the proposed method successfully represented the LiDAR ring phenomenon, where the LiDAR points present a donut shape, which occurs while passing through the fog, in the translation to foggy conditions. The proposed method can also handle the phenomenon in which distant points are erased under rainy conditions. On the other hand, one limitation of this method is that the distance and intensity errors are relatively high when the gap between source and target domains is extremely big, such as when translating between rainy and foggy. Thus, we will consider this problem in future work. Since rain data in the JARI dataset are slightly different with those in real roads, we will also collect a large amount of rain data variously in real roads and modify the proposed method to optimize.

5. Conclusions

We proposed a method to perform LiDAR-to-LiDAR translation to handle adverse weather conditions. Traditional methods for LiDAR translation focus only on translation from simulated to real data. However, the proposed method can deal with translation between sunny and adverse weather data (i.e., rain and fog). Furthermore, the proposed method is very effective, in terms of data augmentation. Some traditional methods for LiDAR translation which can handle various weather conditions have been introduced recently; however, these methods cannot generate natural LiDAR data, as they are based on mathematical formulas. In contrast, the proposed method generates LiDAR data which can represent changes in weather conditions, by training various weather characteristics from the JARI and Spain data sets.
To train LiDAR data with considering various weather characteristics, we formulated the most optimal LiDAR input format and network architecture through diverse experiments. Since traditional translation networks are intended for image translation, formulating the LiDAR data format is a necessary step. Through various comparison experiments by changing LiDAR formats which are composed of distance and intensity values, we figured out that the most optimal format is when it is similar to the distribution of RGB values, while the RGB values in a typical image are limited, distance and intensity values of LiDAR are infinite. To deal with this issue, we designed the architecture with activation functions, which can handle unlimited value. Furthermore, the scale of the proposed LiDAR data format is totally different to that of images. To handle this problem, we revised a network based on the existing CycleGAN structure by modulating ResNet blocks which are utilized for feature extraction.
In our experiments, we used two data sets: JARI and Spain. We observed that the proposed model could generate natural LiDAR data for both environments. We expect that the proposed model can be adopted in driving simulators, for the examination of various autonomous driving algorithms before real-world testing. Thanks to the proposed method, we expect that researchers in the field of autonomous driving can save both time and costs.
In future work, we will improve the proposed method to handle more adverse conditions, such as heavy rain, heavy snow, and heavy fog. As we only tested the proposed method using the real roads in the Spain data set, which includes small weather variations, we, therefore, plan to collect various data from real roads and modify the proposed method to optimize it to real road conditions.

Author Contributions

J.L. contributed to the guidance and revision of the paper. J.L. devised the experimental strategy, conducted the experiments, wrote the manuscript, developed the algorithm, and contributed to the revisions. K.N. and S.K. undertook the leadership of this work. J.L., D.S. and T.N. contributed to the validation and data acquisition. D.S. and T.N. assisted with the experimentation and data preparation. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data are available upon request due to restrictions. The data presented in this study are available from the corresponding authors upon request. The data are not publicly available due to the project’s contract.

Conflicts of Interest

The authors declare no conflict of interest. Jinho Lee, Toshiaki Nishimori, Kenta Nakao, and Shunsuke Kamijo are the inventors of patent 2022-113347

References

  1. Chen, X.; Ma, H.; Wan, J.; Li, B.; Xia, T. Multi-view 3d object detection network for autonomous driving. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1907–1915. [Google Scholar]
  2. Chen, X.; Kundu, K.; Zhang, Z.; Ma, H.; Fidler, S.; Urtasun, R. Monocular 3d object detection for autonomous driving. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2147–2156. [Google Scholar]
  3. Feng, D.; Haase-Schütz, C.; Rosenbaum, L.; Hertlein, H.; Glaeser, C.; Timm, F.; Wiesbeck, W.; Dietmayer, K. Deep multi-modal object detection and semantic segmentation for autonomous driving: Datasets, methods, and challenges. IEEE Trans. Intell. Transp. Syst. 2020, 22, 1341–1360. [Google Scholar] [CrossRef] [Green Version]
  4. Zhang, Z.; Fidler, S.; Urtasun, R. Instance-level segmentation for autonomous driving with deep densely connected mrfs. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 669–677. [Google Scholar]
  5. De Brabandere, B.; Neven, D.; Van Gool, L. Semantic instance segmentation for autonomous driving. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2017; pp. 7–9. [Google Scholar]
  6. Fu, H.; Gong, M.; Wang, C.; Batmanghelich, K.; Tao, D. Deep ordinal regression network for monocular depth estimation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 2002–2011. [Google Scholar]
  7. Godard, C.; Mac Aodha, O.; Firman, M.; Brostow, G.J. Digging into self-supervised monocular depth estimation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 3828–3838. [Google Scholar]
  8. Wofk, D.; Ma, F.; Yang, T.J.; Karaman, S.; Sze, V. Fastdepth: Fast monocular depth estimation on embedded systems. In Proceedings of the 2019 International Conference on Robotics and Automation (ICRA), Montreal, QC, Canada, 20–24 May 2019; pp. 6101–6108. [Google Scholar]
  9. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. In Proceedings of the Conference on Neural Information Processing Systems, Montreal, QC, Canada, 8–13 December 2014; Volume 27. [Google Scholar]
  10. El Sallab, A.; Sobh, I.; Zahran, M.; Essam, N. LiDAR sensor modeling and data augmentation with GANs for autonomous driving. arXiv 2019, arXiv:1905.07290. [Google Scholar]
  11. Dosovitskiy, A.; Ros, G.; Codevilla, F.; Lopez, A.; Koltun, V. CARLA: An open urban driving simulator. In Proceedings of the Conference on Robot Learning, Mountain View, CA, USA, 13–15 November 2017; Volume 78, pp. 1–16. [Google Scholar]
  12. Arioui, H.; Hima, S.; Nehaoua, L.; Bertin, R.J.; Espié, S. From design to experiments of a 2-DOF vehicle driving simulator. IEEE Trans. Veh. Technol. 2010, 60, 357–368. [Google Scholar] [CrossRef] [Green Version]
  13. Lee, W.-S.; Kim, J.-H.; Cho, J.-H. A driving simulator as a virtual reality tool. In Proceedings of the 1998 IEEE International Conference on Robotics and Automation (Cat. No. 98CH36146), Leuven, Belgium, 20 May 1998; pp. 71–76. [Google Scholar]
  14. Kim, I.I.; McArthur, B.; Korevaar, E.J. Comparison of laser beam propagation at 785 nm and 1550 nm in fog and haze for optical wireless communications. In Proceedings of the Optical Wireless Communications III. International Society for Optics and Photonics, Boston, MA, USA, 5–8 November 2001; pp. 26–37. [Google Scholar]
  15. Rasshofer, R.H.; Spies, M.; Spies, H. Influences of weather phenomena on automotive laser radar systems. Adv. Radio Sci. 2011, 9, 49–60. [Google Scholar] [CrossRef] [Green Version]
  16. Wojtanowski, J.; Zygmunt, M.; Kaszczuk, M.; Mierczyk, Z.; Muzal, M. Comparison of 905 nm and 1550 nm semiconductor laser rangefinders’ performance deterioration due to adverse environmental conditions. Opto-Electron. Rev. 2014, 22, 183–190. [Google Scholar] [CrossRef]
  17. Filgueira, A.; González-Jorge, H.; Lagüela, S.; Díaz-Vilariño, L.; Arias, P. Quantifying the influence of rain in LiDAR performance. Measurement 2017, 95, 143–148. [Google Scholar] [CrossRef]
  18. Kutila, M.; Pyykönen, P.; Holzhüter, H.; Colomb, M.; Duthon, P. Automotive LiDAR performance verification in fog and rain. In Proceedings of the 2018 21st International Conference on Intelligent Transportation Systems (ITSC), Maui, HI, USA, 4–7 November 2018; pp. 1695–1701. [Google Scholar]
  19. Jokela, M.; Kutila, M.; Pyykönen, P. Testing and validation of automotive point-cloud sensors in adverse weather conditions. Appl. Sci. 2019, 9, 2341. [Google Scholar] [CrossRef] [Green Version]
  20. Heinzler, R.; Schindler, P.; Seekircher, J.; Ritter, W.; Stork, W. Weather influence and classification with automotive lidar sensors. In Proceedings of the 2019 IEEE intelligent vehicles symposium (IV), Paris, France, 9–12 June 2019; pp. 1527–1534. [Google Scholar]
  21. Wallace, A.M.; Halimi, A.; Buller, G.S. Full waveform LiDAR for adverse weather conditions. IEEE Trans. Veh. Technol. 2020, 69, 7064–7077. [Google Scholar] [CrossRef]
  22. Li, Y.; Duthon, P.; Colomb, M.; Ibanez-Guzman, J. What happens for a ToF LiDAR in fog? IEEE Trans. Intell. Transp. Syst. 2020, 22, 6670–6681. [Google Scholar] [CrossRef]
  23. Kutila, M.; Pyykönen, P.; Ritter, W.; Sawade, O.; Schäufele, B. Automotive lidar sensor development scenarios for harsh weather conditions. In Proceedings of the IEEE International Conference on Intelligent Transportation Systems (ITSC), Rio de Janeiro, Brazil, 1–4 November 2016. [Google Scholar]
  24. Sakaridis, C.; Dai, D.; Van Gool, L. Semantic foggy scene understanding with synthetic data. Int. J. Comput. Vis. 2018, 126, 973–992. [Google Scholar] [CrossRef] [Green Version]
  25. Cordts, M.; Omran, M.; Ramos, S.; Rehfeld, T.; Enzweiler, M.; Benenson, R.; Franke, U.; Roth, S.; Schiele, B. The cityscapes dataset for semantic urban scene understanding. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 3213–3223. [Google Scholar]
  26. Hahner, M.; Dai, D.; Sakaridis, C.; Zaech, J.N.; Van Gool, L. Semantic understanding of foggy scenes with purely synthetic data. In Proceedings of the 2019 IEEE Intelligent Transportation Systems Conference (ITSC), Auckland, New Zealand, 27–30 October 2019; pp. 3675–3681. [Google Scholar]
  27. Wrenninge, M.; Unger, J. Synscapes: A photorealistic synthetic dataset for street scene parsing. arXiv 2018, arXiv:1810.08705. [Google Scholar]
  28. Sakaridis, C.; Dai, D.; Van Gool, L. ACDC: The adverse conditions dataset with correspondences for semantic driving scene understanding. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 11–17 October 2021; pp. 10765–10775. [Google Scholar]
  29. Bijelic, M.; Gruber, T.; Mannan, F.; Kraus, F.; Ritter, W.; Dietmayer, K.; Heide, F. Seeing through fog without seeing fog: Deep multimodal sensor fusion in unseen adverse weather. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 951–963. [Google Scholar]
  30. Hahner, M.; Sakaridis, C.; Dai, D.; Van Gool, L. Fog Simulation on Real LiDAR Point Clouds for 3D Object Detection in Adverse Weather. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 11–17 October 2021; pp. 15283–15292. [Google Scholar]
  31. Yang, T.; Li, Y.; Ruichek, Y.; Yan, Z. Performance Modeling a Near-Infrared ToF LiDAR Under Fog: A Data-Driven Approach. IEEE Trans. Intell. Transp. Syst. 2021, 1–10. [Google Scholar] [CrossRef]
  32. Vargas Rivero, J.R.; Gerbich, T.; Buschardt, B.; Chen, J. Data Augmentation of Automotive LIDAR Point Clouds under Adverse Weather Situations. Sensors 2021, 21, 4503. [Google Scholar] [CrossRef] [PubMed]
  33. Milz, S.; Rudiger, T.; Suss, S. Aerial ganeration: Towards realistic data augmentation using conditional gans. In Proceedings of the European Conference on Computer Vision (ECCV) Workshops, Munich, Germany, 8–14 September 2018. [Google Scholar]
  34. Zhu, J.-Y.; Park, T.; Isola, P.; Efros, A.A. Unpaired image-to-image translation using cycle-consistent adversarial networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2223–2232. [Google Scholar]
  35. Geiger, A.; Lenz, P.; Stiller, C.; Urtasun, R. Vision meets robotics: The kitti dataset. Int. J. Robot. Res. 2013, 32, 1231–1237. [Google Scholar] [CrossRef] [Green Version]
  36. Wang, L.; Goldluecke, B.; Anklam, C. L2R GAN: LiDAR-to-radar translation. In Proceedings of the Asian Conference on Computer Vision, Kyoto, Japan, 30 November–4 December 2020. [Google Scholar]
  37. Li, R.; Li, X.; Fu, C.W.; Cohen-Or, D.; Heng, P.A. Pu-gan: A point cloud upsampling adversarial network. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27–28 October 2019; pp. 7203–7212. [Google Scholar]
  38. Arjovsky, M.; Chintala, S.; Bottou, L. Wasserstein generative adversarial networks. In Proceedings of the 34th International Conference on Machine Learning, Sydney, Australia, 6–11 August 2017; pp. 214–223. [Google Scholar]
  39. Mao, X.; Li, Q.; Xie, H.; Lau, R.Y.; Wang, Z.; Paul Smolley, S. Least squares generative adversarial networks. In Proceedings of the IEEE International Conference on computer Vision, Venice, Italy, 22–29 October 2017; pp. 2794–2802. [Google Scholar]
  40. Chen, X.; Duan, Y.; Houthooft, R.; Schulman, J.; Sutskever, I.; Abbeel, P. Infogan: Interpretable representation learning by information maximizing generative adversarial nets. In Proceedings of the 30th International Conference on Neural Information Processing Systems, Barcelona, Spain, 5–10 December 2016; pp. 2180–2188. [Google Scholar]
  41. Mirza, M.; Osindero, S. Conditional generative adversarial nets. arXiv 2014, arXiv:1411.1784. [Google Scholar]
  42. Radford, A.; Metz, L.; Chintala, S. Unsupervised representation learning with deep convolutional generative adversarial networks. arXiv 2015, arXiv:1511.06434. [Google Scholar]
  43. Zhang, H.; Goodfellow, I.; Metaxas, D.; Odena, A. Self-attention generative adversarial networks. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 10–15 June 2019; pp. 7354–7363. [Google Scholar]
  44. Isola, P.; Zhu, J.Y.; Zhou, T.; Efros, A.A. Image-to-image translation with conditional adversarial networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1125–1134. [Google Scholar]
  45. Kim, T.; Cha, M.; Kim, H.; Lee, J.K.; Kim, J. Learning to discover cross-domain relations with generative adversarial networks. In Proceedings of the 34th International Conference on Machine Learning, Sydney, Australia, 6–11 August 2017; pp. 1857–1865. [Google Scholar]
  46. Yi, Z.; Zhang, H.; Tan, P.; Gong, M. Dualgan: Unsupervised dual learning for image-to-image translation. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2849–2857. [Google Scholar]
  47. Liu, M.Y.; Breuel, T.; Kautz, J. Unsupervised image-to-image translation networks. In Proceedings of the 30th International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Volume 30. [Google Scholar]
  48. Liu, M.Y.; Tuzel, O. Coupled generative adversarial networks. In Proceedings of the 30th International Conference on Neural Information Processing Systems, Barcelona, Spain, 5–10 December 2016; Volume 29. [Google Scholar]
  49. He, K.; Zhang, X.; Ren, S.; Sun, J. Identity mappings in deep residual networks. In European Conference on Computer Vision; Springer: Cham, Switzerland, 2016; pp. 630–645. [Google Scholar]
  50. Tong, T.; Li, G.; Liu, X.; Gao, Q. Image super-resolution using dense skip connections. In Proceedings of the IEEE International Conference on computer Vision, Venice, Italy, 22–29 October 2017; pp. 4799–4807. [Google Scholar]
Figure 1. Purpose of the proposed method. As shown in upper figures, GAN-based networks are generally utilized for image translation. Sunny2Rainy indicates the translation from sunny data to rainy data and Sunny2Foggy is that from sunny data to foggy data. We present the first LiDAR-to-LiDAR translation which deals with various adverse weather conditions based on a GAN method.
Figure 1. Purpose of the proposed method. As shown in upper figures, GAN-based networks are generally utilized for image translation. Sunny2Rainy indicates the translation from sunny data to rainy data and Sunny2Foggy is that from sunny data to foggy data. We present the first LiDAR-to-LiDAR translation which deals with various adverse weather conditions based on a GAN method.
Sensors 22 05287 g001
Figure 2. Overview of the proposed method. The blue boxes represent sunny data, and the green boxes represent adverse weather data. G A B is the generator from sunny to adverse weather, and G B A is the generator from adverse weather to sunny. D A B and D B A are the discriminators for each domain. The red arrows below G A B and G B A represent skip connections. The proposed method adopts ResNet-based generators and includes skip connections to improve the performance.
Figure 2. Overview of the proposed method. The blue boxes represent sunny data, and the green boxes represent adverse weather data. G A B is the generator from sunny to adverse weather, and G B A is the generator from adverse weather to sunny. D A B and D B A are the discriminators for each domain. The red arrows below G A B and G B A represent skip connections. The proposed method adopts ResNet-based generators and includes skip connections to improve the performance.
Sensors 22 05287 g002
Figure 3. Examples of the JARI driving laboratory and 2D PGM. The left figure shows an example image of the JARI environment where the data are collected. The right ones show the Polar-Grid map (i.e., 2D PGM), which represents the LiDAR 3D point cloud as a 2D grid. The top figure of right ones is an example of 2D PGM for distance and the figure below those is that of 2D PGM for intensity in the JARI environment. Black is higher value and white is lower value in the 2D PGM figures.
Figure 3. Examples of the JARI driving laboratory and 2D PGM. The left figure shows an example image of the JARI environment where the data are collected. The right ones show the Polar-Grid map (i.e., 2D PGM), which represents the LiDAR 3D point cloud as a 2D grid. The top figure of right ones is an example of 2D PGM for distance and the figure below those is that of 2D PGM for intensity in the JARI environment. Black is higher value and white is lower value in the 2D PGM figures.
Sensors 22 05287 g003
Figure 4. Examples of the JARI and Spain data sets. The JARI data were collected using an Ouster-64, while the Spain data were obtained by an Ouster-128. Thus, the data sets are slightly different. However, we designed the respective architectures separately. Furthermore, fog data do not exist in the Spain data set, due to the climatic characteristics of Spain.
Figure 4. Examples of the JARI and Spain data sets. The JARI data were collected using an Ouster-64, while the Spain data were obtained by an Ouster-128. Thus, the data sets are slightly different. However, we designed the respective architectures separately. Furthermore, fog data do not exist in the Spain data set, due to the climatic characteristics of Spain.
Sensors 22 05287 g004
Figure 5. Examples of results without skip-connections (a) and with skip-connections (b). The results are the reconstructed 3D points seen from above. The green points indicate reconstructed 3D points. The red points indicate the comparatively high value of intensity. While the result without skip-connections did not generate LiDAR rings, that with skip-connections generated LiDAR rings clearly (see the red rectangles for comparison). The red rectangle on the right indicates the LiDAR rings, which are arrangements of LiDAR points that form rings.
Figure 5. Examples of results without skip-connections (a) and with skip-connections (b). The results are the reconstructed 3D points seen from above. The green points indicate reconstructed 3D points. The red points indicate the comparatively high value of intensity. While the result without skip-connections did not generate LiDAR rings, that with skip-connections generated LiDAR rings clearly (see the red rectangles for comparison). The red rectangle on the right indicates the LiDAR rings, which are arrangements of LiDAR points that form rings.
Sensors 22 05287 g005
Figure 6. Examples of translation between sunny and adverse weather conditions with the JARI data set. The results are reconstructed 3D points seen from above. The green points indicate reconstructed 3D points. The red points indicate the comparatively high value of intensity. We show the translation results between sunny(wet) and rainy conditions. The images on the left are the real data for the source domains, while those on the right are the generated data for the target domains. Even though there was some noise in the generated data, the proposed method can reflect the changes according to various environmental conditions.
Figure 6. Examples of translation between sunny and adverse weather conditions with the JARI data set. The results are reconstructed 3D points seen from above. The green points indicate reconstructed 3D points. The red points indicate the comparatively high value of intensity. We show the translation results between sunny(wet) and rainy conditions. The images on the left are the real data for the source domains, while those on the right are the generated data for the target domains. Even though there was some noise in the generated data, the proposed method can reflect the changes according to various environmental conditions.
Sensors 22 05287 g006
Figure 7. Examples of translation between sunny and adverse weather with the JARI data set. The results are reconstructed 3D points seen from above. The green points indicate reconstructed 3D points. The red points indicate the comparatively high value of intensity. We show the translation results between rainy and foggy conditions. The images on the left are the real data for the source domains, while those on the right are the generated data for the target domains. Even though there was some noise in the generated data, the proposed method can reflect the changes according to various environmental conditions.
Figure 7. Examples of translation between sunny and adverse weather with the JARI data set. The results are reconstructed 3D points seen from above. The green points indicate reconstructed 3D points. The red points indicate the comparatively high value of intensity. We show the translation results between rainy and foggy conditions. The images on the left are the real data for the source domains, while those on the right are the generated data for the target domains. Even though there was some noise in the generated data, the proposed method can reflect the changes according to various environmental conditions.
Sensors 22 05287 g007
Figure 8. Example of translation between sunny and rainy conditions with the Spain data set. The results are reconstructed 3D points seen from above. The green points indicate reconstructed 3D points. In this figure, we show the translation from sunny to rainy condition. The images on the left are the real data for the source domains, while those on the right are the generated data for the target domains. The synthetic rain data were obtained by translation from the real day data, denoted by the red arrow. In addition, the synthetic day data were generated by translation from real rain data, as shown by the blue arrow. As there is no large difference between the day and rain data, the translation results showed no noticeable difference. We thus confirmed that the proposed method is applicable to real-world scenarios.
Figure 8. Example of translation between sunny and rainy conditions with the Spain data set. The results are reconstructed 3D points seen from above. The green points indicate reconstructed 3D points. In this figure, we show the translation from sunny to rainy condition. The images on the left are the real data for the source domains, while those on the right are the generated data for the target domains. The synthetic rain data were obtained by translation from the real day data, denoted by the red arrow. In addition, the synthetic day data were generated by translation from real rain data, as shown by the blue arrow. As there is no large difference between the day and rain data, the translation results showed no noticeable difference. We thus confirmed that the proposed method is applicable to real-world scenarios.
Sensors 22 05287 g008
Table 1. JARI data set configuration. “O” indicates that data exist. Rain data are composed of the three settings to deal with precipitation (30 mm/h, 50 mm/h, and 80 mm/h) and fog data also contain three settings to handle the degree of fog (15 m, 30 m, and 80 m visible distance).
Table 1. JARI data set configuration. “O” indicates that data exist. Rain data are composed of the three settings to deal with precipitation (30 mm/h, 50 mm/h, and 80 mm/h) and fog data also contain three settings to handle the degree of fog (15 m, 30 m, and 80 m visible distance).
ConditionsDay/Night5 km/h40 km/h
Sunny
(dry)
DayOO
NightOO
Sunny
(wet)
DayOO
NightOO
Rain
(30 mm/h)
DayOO
NightOO
Rain
(50 mm/h)
DayOO
NightOO
Rain
(80 mm/h)
DayOO
NightOO
Fog
(15 m)
DayOO
NightOO
Fog
(30 m)
DayOO
NightOO
Fog
(80 m)
DayOO
NightOO
Table 2. Distance error results depending on LiDAR data format. The numbers in the table indicate distance errors (in m). The bold numbers show that the most optimal format is Intensity/15.
Table 2. Distance error results depending on LiDAR data format. The numbers in the table indicate distance errors (in m). The bold numbers show that the most optimal format is Intensity/15.
From Source
to Target
OriginalIntensity/15Intensity
Square-Root
log 2 (Intensity)
from Sunny(dry)–Day to Fog(15 m)–Day3.297721.893002.003922.29613
from Fog(15 m)–Day to Sunny(dry)–Day11.292315.318866.692137.00921
Table 3. Intensity error results depending on LiDAR data format. The numbers in the table indicate intensity errors. The bold numbers prove that the most optimal format is Intensity/15.
Table 3. Intensity error results depending on LiDAR data format. The numbers in the table indicate intensity errors. The bold numbers prove that the most optimal format is Intensity/15.
From Source
to Target
OriginalIntensity/15Intensity
Square-Root
log 2 (Intensity)
from Sunny(dry)–Day to Fog(15 m)–Day41.2219818.1787422.5592421.61834
from Fog(15 m)–Day to Sunny(dry)–Day360.95213178.30045189.13991190.81837
Table 4. Error results depending on the number of ResNet blocks in the generator and discriminator. G_Num and D_Num indicate the number of ResNet blocks in the generator and discriminator, respectively. The numbers in the table indicate the distance and intensity errors. The bold numbers show that the most architecture is four ResNet blocks for the generator and one ResNet block for the discriminator.
Table 4. Error results depending on the number of ResNet blocks in the generator and discriminator. G_Num and D_Num indicate the number of ResNet blocks in the generator and discriminator, respectively. The numbers in the table indicate the distance and intensity errors. The bold numbers show that the most architecture is four ResNet blocks for the generator and one ResNet block for the discriminator.
G_Num
&
D_Num
-Distance Error [m]-
from
(Fog(80 m)–Day
to
Sunny(dry)–Day)
-Distance Error [m]-
from
(Sunny(dry)–Day
to
Fog(80 m)–Day)
-Intensity Error-
from
(Fog(80 m)–Day
to
Sunny(dry)–Day)
-Intensity Error-
from
(Sunny(dry)–Day
to
Fog(80 m)–Day)
1_14.671282.62430141.6624597.37588
2_13.992602.50344133.3904590.80555
3_14.085082.45496126.6100587.46522
4_13.715192.40176118.9772887.32463
4_27.300832.59537138.4292488.39707
5_14.356192.41132139.3271988.23212
6_14.608282.43462165.0402789.37218
Table 5. Error results with and without skip-connections. The numbers in the table indicate the distance and intensity errors. The bold numbers prove that the proposed method with skip-connections outperforms that without.
Table 5. Error results with and without skip-connections. The numbers in the table indicate the distance and intensity errors. The bold numbers prove that the proposed method with skip-connections outperforms that without.
-Distance Error [m]-
from
(Fog(80 m)–Day
to
Sunny(dry)–Day)
-Distance Error [m]-
from
(Sunny(dry)–Day
to
Fog(80 m)–Day)
-Intensity Error-
from
(Fog(80 m)–Day
to
Sunny(dry)–Day)
-Intensity Error-
from
(Sunny(dry)–Day
to
Fog(80 m)–Day)
w/o skip
connections
3.715192.40176118.9772887.32463
w skip
connections
3.629872.40018109.6212087.00812
Table 6. Translation results under various conditions with the JARI data set.
Table 6. Translation results under various conditions with the JARI data set.
From Source to TargetDistance
Error [m]
Intensity
Error
from Sunny(dry)–Day to Sunny(dry)–Night1.4949785.98490
from Sunny(dry)–Night to Sunny(dry)–Day1.4365394.33114
from Sunny(dry)–Day to Sunny(wet)–Day1.7574185.57328
from Sunny(wet)–Day to Sunny(dry)–Day1.8152392.17662
from Sunny(dry)–Day to Sunny(wet)–Night1.7610087.66818
from Sunny(wet)–Night to Sunny(dry)–Day1.8454094.66367
from Sunny(dry)–Day to Rain(30 mm/h)–Day2.7851868.59065
from Rain(30 mm/h)–Day to Sunny(dry)–Day3.0875199.64109
from Sunny(dry)–Day to Rain(30 mm/h)–Night2.7459165.87693
from Rain(30 mm/h)–Night to Sunny(dry)–Day3.36074109.62120
from Sunny(dry)–Day to Rain(50 mm/h)–Day2.7700870.83799
from Rain(50 mm/h)–Day to Sunny(dry)–Day3.47687112.60500
from Sunny(dry)–Day to Rain(50 mm/h)–Night2.8960369.15411
from Rain(50 mm/h)–Night to Sunny(dry)–Day3.28403103.78098
from Sunny(dry)–Day to Rain(80 mm/h)–Day2.6901759.15411
from Rain(80 mm/h)–Day to Sunny(dry)–Day3.6037498.61141
from Sunny(dry)–Day to Rain(80 mm/h)–Night2.6095162.18444
from Rain(80 mm/h)–Night to Sunny(dry)–Day3.1533396.81059
from Sunny(dry)–Day to Fog(15 m)–Day1.7313417.84077
from Fog(15 m)–Day to Sunny(dry)–Day4.19671111.54227
from Sunny(dry)–Day to Fog(15 m)–Night1.6925816.66186
from Fog(15 m)–Night to Sunny(dry)–Day4.19542115.00208
from Sunny(dry)–Day to Fog(30 m)–Day1.8887156.95320
from Fog(30 m)–Day to Sunny(dry)–Day3.62490109.74681
from Sunny(dry)–Day to Fog(30 m)–Night1.9034155.06957
from Fog(30 m)–Night to Sunny(dry)–Day3.3919197.75404
from Sunny(dry)–Day to Fog(80 m)–Day2.2001867.00812
from Fog(80 m)–Day to Sunny(dry)–Day2.6298789.62120
from Sunny(dry)–Day to Fog(80 m)–Night2.3504463.26001
from Fog(80 m)–Night to Sunny(dry)–Day2.5770185.65783
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Lee, J.; Shiotsuka, D.; Nishimori, T.; Nakao, K.; Kamijo, S. GAN-Based LiDAR Translation between Sunny and Adverse Weather for Autonomous Driving and Driving Simulation. Sensors 2022, 22, 5287. https://doi.org/10.3390/s22145287

AMA Style

Lee J, Shiotsuka D, Nishimori T, Nakao K, Kamijo S. GAN-Based LiDAR Translation between Sunny and Adverse Weather for Autonomous Driving and Driving Simulation. Sensors. 2022; 22(14):5287. https://doi.org/10.3390/s22145287

Chicago/Turabian Style

Lee, Jinho, Daiki Shiotsuka, Toshiaki Nishimori, Kenta Nakao, and Shunsuke Kamijo. 2022. "GAN-Based LiDAR Translation between Sunny and Adverse Weather for Autonomous Driving and Driving Simulation" Sensors 22, no. 14: 5287. https://doi.org/10.3390/s22145287

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop