Next Article in Journal
Kinematic and Dynamic Structure of the 18 May 2020 Squall Line over South Korea
Previous Article in Journal
ITS Efficiency Analysis for Multi-Target Tracking in a Clutter Environment
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Wind Profile Reconstruction Based on Convolutional Neural Network for Incoherent Doppler Wind LiDAR

1
School of Earth and Space Sciences, University of Science and Technology of China, Hefei 230026, China
2
Hefei National Laboratory, Hefei 230088, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2024, 16(8), 1473; https://doi.org/10.3390/rs16081473
Submission received: 30 January 2024 / Revised: 13 April 2024 / Accepted: 17 April 2024 / Published: 22 April 2024

Abstract

:
The rapid development of artificial intelligence (AI) and deep learning has revolutionized the field of data analysis in recent years, including signal data acquired by remote sensors. Light Detection and Ranging (LiDAR) technology is widely used in atmospheric research for measuring various atmospheric parameters. Wind measurement using LiDAR data has traditionally relied on the spectral centroid (SC) algorithm. However, this approach has limitations in handling LiDAR data, particularly in low signal-to-noise ratio (SNR) regions. To overcome these limitations, this study leverages the capabilities of customized deep-learning techniques to achieve accurate wind profile reconstruction. The study uses datasets obtained from the European Centre for Medium Weather Forecasting (ECMWF) Reanalysis v5 (ERA5) and the mobile Incoherent Doppler LiDAR (ICDL) system constructed by the University of Science and Technology of China. We present a simulation-based approach for generating wind profiles from the statistical data and the associated theoretical calculations. Whereafter, our team constructed a convolutional neural network (CNN) model based on the U-Net architecture to replace the SC algorithm for LiDAR data post-processing. The CNN-generated results are evaluated and compared with the SC results and the ERA5 data. This study highlights the potential of deep learning-based techniques in atmospheric research and their ability to provide more accurate and reliable results.

Graphical Abstract

1. Introduction

Undoubtedly, artificial intelligence (AI) has the potential to change almost every aspect of our lives. Among all of the AI techniques, deep learning has gained the most significant attention due to its ability to construct intricate structures that mimic the human neuron system [1,2]. Unlike traditional algorithms, deep neural networks have the capability to learn, memorize, and generate new content. This has greatly enhanced the compatibility of data analysis in all aspects, especially in image recognition [3,4,5,6,7] and natural language processing [8,9,10,11]. Recently, deep learning has become an essential tool in numerous applications, scoping from big-data analytics in daily APPs to AI-generated content products like ChatGPT. In the field of remote sensing, in particular, it has also contributed a significant improvement from the CNN techniques, with promising potential for further innovations [12,13,14].
The acquisition of atmospheric measurement data plays a crucial role in understanding atmospheric weather patterns, climate changes, and other related phenomena. LiDAR technology has been widely deployed as an effective method for collecting various atmospheric parameters, including temperature, air densities, wind fields, humidity, and more. For wind measurement, two categories of LiDAR can be classified: coherent (or heterodyne) Doppler LiDAR [15,16] and incoherent (or direct detection) Doppler LiDAR (ICDL) [17,18,19]. Regardless of its category, its theoretical principle is to utilize the Doppler shift information from the backscattered signal to determine the wind speed. Conventionally, the spectral centroid (SC) algorithm [12,20,21] usually works as the data post-processing method for wind profile reconstruction.
The utilization of machine learning algorithms in the area of wind detection is not a novelty [12,13,14]. Some traditional algorithms such as polynomial fitting, linear regression, and spline transformation have already been implemented in analyzing and determining wind fields in several applications, such as background wind field deduction and wind profile trend fitting [22,23,24,25,26,27]. These algorithms have provided insights into wind behaviors, but their limitations in processing local delicate wind perturbations have led to expectations for new techniques. With the rise of the deep learning and CNN technique, based on its remarkable prediction and classification capability, advanced manipulations can be performed. Consequently, a more accurate wind profile reconstruction based on this new CNN approach can be utilized with the assistance of deep learning algorithms.
Previous studies have explored wind speed prediction and wind field reconstruction, yet the insight into atmospheric wind field trends remains uninvestigated [13,28]. These efforts have been limited to relatively low altitudes and bounded ranges in the hundreds or thousands of meters. Additionally, their convolutional neural network (CNN) structures were not designed specifically for this task, thus obscuring the full potential of the CNN algorithm. Here, we exploit the power of deep learning architecture to achieve precise wind profile reconstruction.
To enhance the reliability of our work, we collected data from both the mobile ICDL system shown in Figure 1 and the ERA5 dataset produced by ECMWF. Addressing the lack of a substantial quantity of dependable ground truth data for wind profiles between altitudes of 30 km and 60 km, we present a simulation-based approach for constructing wind profiles using theoretical calculations and mathematical statistics; consisting of components in wind perturbation, signal-to-noise (SNR) error, and background wind field trends. The data preprocessing involves integrating the simulated and LiDAR-derived data for format unification. To displace the spectral centroid algorithm, we construct a U-net-based [29] CNN model, due to its efficiency and detail-capturing capability, for more accurate post-processing of the LiDAR data. The network integrates both simulated wind profile lines and wind profile backscattered data into the training procedures. We subsequently evaluate the CNN-generated results and compare them with the ground truth labels from our simulations, as well as with the spectral centroid results. Moreover, we conduct an evaluation of real wind profile data obtained from our LiDAR with no ground truth and compare it to the spectral centroid results and the ERA5 dataset in the last step.

2. Methods

2.1. Data Acquisition

2.1.1. Measurement from ICDL

The raw LiDAR data, available at Science Data Bank [30], were obtained from the mobile Rayleigh Doppler LiDAR developed by the University of Science and Technology of China [31,32], during the night period at Kolar, Xinjiang (41.1°N, 87.1°E), China in 2019. The LiDAR system has been installed in the vehicles shown in Figure 1. Our ICDL employs the double-edge technique [18,28,33,34] as the primary frequency discriminator to determine the Doppler shift. The system description along with a detailed explanation of the theoretical principle can be found in our previous studies [18,32,35]. The telescope of the LiDAR is pointed at a zenith angle (ϕ) of 30 degrees with the correlated azimuth angles (ϴ) in the east–west direction.
The data collected were stored in the form of a text document, recording photon counts from each of the two photodetectors corresponding to the double-edge technique. The photon counts ratio R(∆vd) after the two edge filters is under the following relationship:
R v d = I 1 I 2 I 1 + I 2
The response curve for discriminating frequency shift from the photon counts ratio will be scanned and calculated on a daily basis at the beginning of the operation. The line-of-sight (LOS) wind velocity VLOS related to the frequency shift then can be calculated under the Doppler shift formula:
V L O S = λ L a s e r 2 v d
The horizontal wind velocity, characterized by an east–west orientation corresponding to the direction targeted by our LiDAR, adheres to the relationship delineated in Equation (3), assuming the vertical wind flow in the stratosphere remains below 2 m/s [18] and is considered negligible.
V E W = V L O S s i n ( ϕ ) c o s ( ϴ )
Across approximately 40,000 files, each text file is generated from the repetitional detection of 40 s, recording the backscattered signal from 4000 laser shots that have been transmitted to the targeted direction of the atmosphere. A single text file contains a total of 16,000 bins, where each bin logs the photon counts on a 50 ns interval, corresponding to a line-of-sight (LOS) distance of 7.5 m. To ensure sufficient photon counts for a feasible signal-to-noise ratio (SNR), a range gate of 30 bins and 60 bins’ combination is selected at an altitude threshold of around 40 km, where each resolution grid corresponds to 225 m and 450 m for the LOS distance, equivalent to approximately 195 m and 390 m in height, respectively. To compensate for the low SNR at high altitudes, each of our wind profiles represents an average horizontal wind speed of 30 min. An example wind data image of the real LiDAR signal after pre-processing is shown in Section 2.2.3, and the wind profile example is shown in Figure 2. More details of the signal pre-processing are described in Section 2.3.2.

2.1.2. ECMWF: ERA5 Dataset

The ERA5 dataset [36] is a comprehensive atmospheric study that employs observation data gathered from an extensive list of satellites, aircraft, and onsite stations [37]. Produced by the Copernicus Climate Change Service (C3S) at ECMWF, this global reanalysis model incorporates a variety of atmospheric indices, including pressure, temperature, density, vorticity levels, and wind field decompositions, in both zonal and meridional directions. The dataset offers a spatial resolution of 0.25° × 0.25° for longitude and latitude and covers altitudes ranging from 10 m to 80 km with an hourly time resolution, including historical data up to five days prior to the present time. Additionally, the ERA5 dataset offers forecasting capabilities, extending its calculations into future weeks. This highly accurate atmosphere re-analysis model contains wind profile data at precisely the same location and time as our LiDAR’s detection introduced in Section 2.1.1, with a neglectable direct linear distance error of approximately 14 km. Given the ERA5’s lower altitude resolution, we extract only its background wind trend to assist in simulating the background wind of our wind profile. Detailed wind extraction is specified in the next section.

2.2. Simulation of Wind Profile Reconstruction

The LiDAR-measured wind profile can be divided into two components: the true wind velocity and the uncertainty caused by errors. Traditionally, the LiDAR’s data are calculated based on the spectral centroid algorithm, which estimates the average radial wind speed of a given time and space interval. In our case, the objective is to find the averaged horizontal wind speed for the east–west direction of a 30 min interval, and the space interval depends on the range gate of 195 m and 390 m. More details of the space interval are associated with LiDAR design and introduced in Section 2.1.1. Wind profiles calculated by this method consist of a large portion of uncertainty, especially in regions with low SNR, where the resulting wind profile cannot reflect an accurate true wind field of the measured air flow.
In this paper, we introduce a newly designed CNN network to improve the accuracy of the LiDAR. When training a neural network model, the ground truth answer, known as the label of the input image, has to be fed to the network to give the machine an example of the ideal outcome of the given input data. However, using wind data based on the SC result as the label for the CNN model may not yield the desired outcome, as the label itself is not accurate. To overcome this problem, we decided to first simulate the “ground truth” of the wind profile that does not contain any errors, while the errors can be generated separately and combined into the ground truth to imitate the raw data collected from the LiDAR.
On account of various factors of atmospheric parameters, including air density, temperature, geostrophic force, atmospheric circulation, etc., the wind field can be further split into two components: wind perturbation from the local oscillation, and the background wind field. The wind perturbation and the background wind profile change in both time periods and altitude levels. To extract the background wind from the wind perturbation, several methods from former research [22,23,24,25,26,27] are all capable of this task. We used the Butterworth filter from Python’s built-in library to split the two for convenience: the background wind profile, exemplified by the orange curve shown in Figure 2, separated from the original wind profile, represented by the blue curve. After extracting the background wind, the wind profile is simply an oscillation consisting of the actual perturbation and the error. To emulate the true wind profile, the perturbation and the error need to be further discriminated.

2.2.1. Error Calculation

Previous works evaluating LiDAR errors [28,38,39] have identified several types of errors: systematic errors and random errors. The systematic errors can be divided into two categories: the sensitivity of the Fabry–Pérot Interferometer (FPI) and the signal-noise-ratio of the received photon counts. The sensitivity of the FPI is calculated based on Equation (4):
Θ ( v ) = d R ( v ) d v 1 R ( v )
where Θ denotes the sensitivity of the system, R(v) is the I1/I2 ratio of the photon counts detected by the two photodetectors, and the dR(v)/dv term represents the FPI resolution for the rate of change of the detected frequency from the Doppler effect. This parameter is measured and curve-fitted based on the spectral scan from our LiDAR system mentioned in Section 2.1.1.
The error associated with the SNR is dependent on the photon counts from the photodetector; three major sources contribute to this noise: photodetector’s shot noise, background noise from other light sources, and detection errors. The SNR is calculated based on Equation (5):
S N R 1 = S i g n a l 1 S i g n a l 1 + N o i s e 1 ,   a n d     S N R 2 = S i g n a l 2 S i g n a l 2 + N o i s e 2 S N R T o t a l = 1 1 S N R 1 2 + 1 S N R 2 2
where Signal represents the photon counts from the backscattered laser signal, Noise represents the photon counts when no laser source is activated, and SNRtotal is the total SNR of edge-channel 1 and 2.
A relation of the standard deviation of the error proportional to the inverse of the SNR (error ∝ 1/SNR) is considered [28], and the total error is written as
E r r o r v = 1 Θ v 1 S N R T o t a l
The relationship between the average standard deviation (SD) of the wind speed error and altitude, as calculated from the actual LiDAR signal, is illustrated in Figure 3b.

2.2.2. Wind Perturbation

The intricate interplay of various factors, including gravity waves, planetary waves, and jet streams, among others, contributes to the apparently random patterns of wind in the middle atmosphere [40,41,42]. Leveraging the unique capabilities of our LiDAR location, which is strategically positioned to observe mountain gravity waves, we have gained remarkable insights into the nature of wind perturbations.
The altitude-dependent variation in air density plays a pivotal role in shaping the amplitude of wind perturbations, as it induces the changes in the oscillatory behavior of upward-propagating waves [43,44]. As the air density decreases with increasing altitude, upward waves tend to oscillate with greater amplitudes. At each altitude level, wind perturbation variations adhere to a Gaussian distribution [26,45], represented by the standard deviation (SD), denoted as “σ sigma”, which exhibits a distinct trend with altitude. To quantify this phenomenon, we systematically analyze a vast dataset comprising 40,000 raw wind measurements collected over 3 months, from October to December 2019. Our analysis focuses primarily on zonal wind profiles across various altitudes.
As the upward-propagating waves reach a certain altitude, the wave may be resonating with and absorbed by the background wind field. Such incidence of gravity wave-breaking appears to dominantly take place around 35~50 km in altitude [23,46,47,48]. Figure 3a shows the plot of the SD of the amplitude of the wind perturbation against the altitude. This aligns with our perturbation theory, where the perturbation increases along with the altitude and eventually decreases around the 35~50 km mark. At the same time, this perturbation trend follows a similar trend from ERA5’s modeling, which will later be used to generate the ground-truth wind profiles as the labels for the CNN network.

2.2.3. Wind Profile Reconstruction

In this step, an exhaustive analysis of each component of the wind profile is conducted, and wind profiles are simulated based on the information obtained above. Initially, 10,000 samples of background wind trends are generated using the wind trends extracted from both the LiDAR system and ERA5 dataset by applying a low pass Butterworth filter to their original wind profiles. Subsequently, the wind perturbations calculated in the preceding section are incorporated into the background wind profile, resulting in the generation of 40,000 distinct combinations of the “ground truth” of wind profiles. These ground-truth wind profiles serve as the labels or references for the CNN model. Leveraging the ground truth data, the raw LiDAR wind profile is simulated by incorporating the errors identified in Section 2.2.1 into the ground truth samples. The magnitude of the assigned error is determined by the randomly generated exponential signal-to-noise ratio (SNR) profiles during the simulation process, which emulates the signal attenuation from the increasing distance and the visibility of weather conditions. This is crucial because the range of the LiDAR signals can vary significantly, with certain days exhibiting signal ranges of only up to 30 km, while on others, detectable signals may extend up to 50–60 km.
The reconstructed LiDAR signal of the wind profile through our simulation is shown in Figure 4 (top). To boost accuracy, we also generated two additional channels: the SC wind profile and the Spline Transformer wind profile of the LiDAR data. We propose that the SC can detect the local details of the minor wind transitions, while the Spline Transformer profile can assist in ascertaining the overall wind trend.

2.3. Machine Learning

2.3.1. CNN Structure

The present work describes a customized U-net architecture implemented for wind speed analysis. Several attempts with other architectures such as SqueezeNet [5], ResNet [3], and purely customized CNN structure have also been investigated, but with less promising result. Comparing these architectures’ performances is beyond this paper’s scope but may be explored in future work.
The U-net was originally designed for image segmentation and noise cancellation by dropping redundant spatial information during its middle bottleneck layers. The encoder of the U-net is composed of convolutional layers, followed by ReLU activation layers [49], and a maxpooling layer [50] that contracts the spatial dimensions of the image while boosting the feature/channel space. In the middle layers, the same principle is followed as in the encoder, where the feature space is further extended. In the decoder, convolutional and ReLU layer combinations are employed to reduce both the spatial and feature space, followed by two convolutional transpose layers to enlarge the spatial dimension to the size of the original image. The full structure of the U-net is shown in Figure 5.
The input of the network is a 3-channel image of 220 × 500 pixels, and the output is a single-channel image of the same size. During the spatial contraction, the vertical space is only contracted by a small portion, while the horizontal space is contracted to one-quarter of the original dimension. This design concept aims to let the network focus on analyzing the noised raw data obtained at each altitude level of the wind, assuming a strong randomicity of the wind profile, and considering that the vertical relation of the wind profile has less effect. Each horizontal line of the network’s output image shows the probability distribution of the wind speed at a particular altitude. For computation along each line, it forms a classification problem at each pixel, each representing a class of the 1 m/s range of wind speed. Thus, the whole architecture can be seen as a multi-classification process.
To train the network, a total of 40,000 simulated samples with ground truth labels and 10,000 true LiDAR data with spectral centroid calculated labels are used. The data are split into a 60:20:20 train, test, and validation set ratio, respectively. The ADAM optimizer [51] and mean square error criterion is adopted with a selection of 3 × 10−4 learning rate and 50 batch-size for 25 epochs, and the overfitting occurs at approximately epoch 23.

2.3.2. Pre-Processing

The raw data collected by the LiDAR instrument are stored in the text file format. Each file contains photon counts from two Fabry–Perot Interferometer (FPI) channels, resulting in a total of 16,000 bin measurements arranged along the text rows. To ensure adequate signal strength, the range of 1600 to 8200 bins was selected for analysis, providing a maximum detection range of 53 km in altitude. The LiDAR data were commonly acquired over a daily period spanning from 11:00 p.m. to 8:00 a.m. at the Xinjiang location, with a detection rate of 40 s between each file.
To enhance signal quality at the far end, 48 files, representing the averaged wind profile of 30 min, were combined for each wind profile. The resulting wind profile was represented as a 220 × 500-pixel image, where the vertical axis corresponds to range gates from 10 km up to 53 km, and the horizontal axis is defined by the wind speed domain of −250 m/s to +250 m/s. Each pixel in this configuration represents the wind speed at a specific range gate level. In the near field between 10 km and 15 km, where the signal intensity is relatively strong, each of the 48 files generated one pixel for a total of 48 pixels in each row. As the measured height increases, the quantity of files utilized to form one pixel adjusts in response to the descending SNR. Conversely, at the far field beyond 45 km, the signal from all 48 files was combined into a single pixel point to compensate for the low signal strength. An example of the pre-processed simulated LiDAR signal is shown in Figure 4 (top), and the real LiDAR signal is shown in Figure 6.

2.3.3. Post-Processing

Figure 7 (top) illustrates the output of the CNN in the form of a single-channel image that is equal in size to the input. The computation of the wind profile, based on this CNN output, can be conducted using two distinct methods. The first approach involves utilizing the “argmax” function, as this represents the standard method in classification problems. As previously mentioned, the output forms a classification problem, where each pixel on the output image corresponds to a specific wind speed value within the 1 m/s domain, ranging from −250 m/s to 250 m/s, resulting in a total of 500 pixel sections in each horizontal line of the image that represents a single altitude level. The argmax function identifies the position of the top class of the wind speed section indicating the highest probability category for each horizontal line. By connecting all wind speed section positions from each altitude level, the resulting wind profile is depicted in Figure 8a.
For the second method, we implement the spectral centroid algorithm on the output image of the CNN, and the result is shown in Figure 8b. The methodology proffers several merits: (1) It is advantageous compared to conventional object classification, where classes are not invariably linearly connected. In contrast, the classifications within our wind speed section present a linear relation within the wind speed range of −250 m/s to +250 m/s. Such arrangement enables the spectral centroid algorithm to calculate the mean wind speed class on a per-line basis. (2) The spectral centroid technique takes a greater multitude of classes into consideration, which is especially beneficial for the low signal-to-noise ratio (SNR) domain. In such regions, the most probable wind speed class may not provide sufficient accuracy for precise classification, emphasizing the value of our method’s extensive class evaluation.

3. Results

3.1. CNN Outputs

In this section, we present a comparison between the outputs of the U-Net model and the traditional raw spectral centroid algorithm for both simulated and real LiDAR signals. Figure 7(top) illustrates the output of the simulated LiDAR signal, and the output of the real LiDAR signal shows a comparable pattern. The output images clearly depict the wind profile in the high SNR and intensity region at lower altitudes, where the certainty of top classes is relatively high. Conversely, in regions with lower SNR, the probabilities of top classes are widely distributed along each altitude level. As expected, the CNN produces lower intensity and probability at higher altitudes, in accordance with our theoretical understanding. Overall, the results demonstrate the effectiveness of the proposed approach in wind speed analysis, showing promise for future applications.

3.2. Wind Profile Evaluation for Simulated Signal

By employing the two techniques presented in Section 2.3.3, we derived the wind profile depicted in Figure 8 and juxtaposed it against the ground truth. Additionally, we included the SC and Gaussian-smoothed SC (GSSC) [21] outcomes, which were obtained using the raw input image data, for comparative purposes. We utilized the R-squared (R2) coefficient of determination (Formula (7)) as our assessment metric.
R 2 = 1 ( V a c t u a l V p r e d i c t ) 2 ( V a c t u a l V a c t u a l _ m e a n ) 2
Our evaluation was conducted on the validation set to eliminate any bias arising from the training and testing sets. The results of the four methods are presented in Table 1, where Mean, Max., and Min. represent the average, the minimum, and the maximum performance of a single wind profile after comparing the method’s result with the ground truth label, correspondingly. Notably, the spectral centroid of the CNN (SCCNN) outperforms the argmax method (AMCNN) in nearly all scenarios. A comparison of the two results can be found in Figure 8a,b. Among the traditional algorithms, shown in Figure 8c,d, the Gaussian-smoothed SC method (GSSC) outperforms the raw SC method (SCR). In terms of the R2 performance, the SCCNN achieved a superior mean score of 0.607, which was 0.137 points higher than the GSSC method’s score of 0.470, while the other two methods underperformed at the mean of 0.0126 and −0.1680. Thus, we selected the SCCNN profile from the CNN output as our primary result and the GSSC profile from the raw data as the traditional processing method for all of the subsequent comparative analyses.
In Figure 9, we observe a high correlation of the regression analysis results for the SCCNN and GSSC, indicated by a majority of data points congregating near the ground truth line. However, a portion of these points diverges substantially from this line. These points with the high deviation are dominated primarily by the detection errors in the high-altitude domain due to the weak SNR signal, subsequently affected by the increasing fluctuation of the R ratio in Formula 1. Precise wind speed restoration of the high-uncertainty domain is notably challenging via conventional techniques, such as the GSSC. In contrast, the SCCNN is able to provide conditional predictions informed by its pre-training, addressing these high-uncertainty scenarios more adeptly. It can estimate each wind speed point through the adjacent points, the overall wind profile deflection, as well as the prevailing seasonal background wind trend, consequently leading to considerable error reduction.

3.2.1. Low SNR Scenario

During the validation process, we also assessed the performance of the cases under low signal-to-noise ratio (SNR) conditions. We simulated the LiDAR signal for low visibility days, such as those with cloudy, sandstorm, or strong moonlight backgrounds, by manually increasing the noise and decreasing the SNR at 30~40 km from 35~15 to 17~9. We randomly selected 100 low SNR wind profiles from the validation set and compared the results of all four methods. In the worst case, the SCR and AMCNN methods produced negative R2 values that were close to −1, as shown in Table 1. Meanwhile, the SCCNN achieved an R2 value of 0.3119, whereas the GSSC from the raw data obtained an R2 value of 0.0241 in their minimal performance. On average, GSSC and SCCNN achieved an overall performance of R2 values of 0.423 and 0.543, respectively.

3.2.2. High SNR (Near Field) and Low SNR (Far Field) Region Reanalysis

To perform a comprehensive analysis of the benefits of the CNN result, we segregated the low and high SNR regions (near field and far field) from each wind profile, allowing us to compare the two regions independently. The boundary line was set at the 37 km altitude level, that is, where the vertical resolution changes from 30 to 60 bins for a single range gate. We reanalyzed the data solely in the high SNR region of the same wind profiles from the validation set. Both GSSC and SCCNN methods performed admirably when the signal strength was relatively strong, achieving a remarkably high R2 score of 0.826 and 0.912. The regression analysis score testified to the CNN result’s reliability by accurately restoring the wind profile in high SNR scenarios.
Next, we examined the low SNR profile region. After surpassing the 37 km boundary line, the accuracy of the traditional method GSSC dropped significantly, to 0.279, whereas the SCCNN outputs still exhibited an acceptable R2 of 0.457. Therefore, we can conclude that the CNN outperforms the traditional SC method, notably in the low SNR region, where the CNN’s capability to restore the wind profile pattern is particularly evident. Instead of treating each LiDAR signal’s wind profile data individually, the CNN adopts the wind pattern at a specific level from the dataset at the given time interval. This confers a significant advantage to the CNN, enabling it to partially predict the wind speed even under low signal strength conditions, thereby achieving a substantial boost in wind profile reconstruction performance.

3.3. Wind Profile Evaluation for Real LiDAR Data

After a full assessment of the CNN outputs for simulated LiDAR signals, we evaluated its performance on real LiDAR data, as depicted in Figure 10 and Figure 11. A comparative analysis is performed between GSSC, SCCNN, and the ERA5 data at the same time period and location. The constructed real LiDAR image, shown in Figure 6, is identically manipulated with the simulated data by inputting it into our U-Net and applying the SC method on the output end. The LiDAR signal measurement was collected during the nocturnal hours of 31 October 2019, from Kolar, Xinjiang. The wind profiles for the duration were derived using both GSSC and SCCNN methodologies. The overview horizontal wind profile of the east–west orientation with a 30° zenith angle on a 6.5-h timeline is shown in Figure 10, providing a basis for comparative analysis.
The result of the GSSC, depicted in the top section of Figure 10 (top), reveals a conspicuous absence of the wind speed data in the higher-altitude regions, attributed to the application of an SNR threshold of 10, as determined by Equation (6). Contrarily, at lower altitudes, specifically below 40 km, the wind profiles generated by both GSSC and the SCCNN exhibit a significant overlap, with an R2 score of 0.9343. This substantial concordance not only underscores the precision of our SCCNN model but also affirms the efficacy of the traditional GSSC approach, especially in cases where LiDAR signals are abundant and the GSSC result is reliable.
Referring to the performance of the simulated signal in Section 3.2, we posit that the SCCNN outputs yield a suitable precision under low SNR circumstances, thus surpassing GSSC in the detection range. As depicted in Figure 11a, there is considerable overlap between the results obtained from the GSSC and SCCNN methodologies; however, the SCCNN exhibits lower oscillation amplitudes as the altitude rises. Upon juxtaposing these results with ERA5 data, it is evident that the ERA5 dataset aligns with the background wind trends of our wind profiles. This alignment is attributed to the lower resolution of the ERA5 dataset, especially at altitudes beyond 37 km where its vertical resolution extends to 1 km. Given that the ERA5 wind profiles lack details on local wind perturbations, we compared them against our results’ background wind, which was derived using the previously mentioned Butterworth filter. Figure 11b reveals a strong alignment between the background wind trends and the ERA5 wind profile, with the SCCNN method achieving an R2 score of 0.7653. This score is significantly higher, by 0.0738, compared to the GSSC method’s score of 0.6915. This outcome represents the most favorable performance in the comparative analyses of seven ERA5 wind profiles recorded from 9:00 PM to 3:00 AM, consistent with the timeline presented in Figure 10 and within the constraints of ERA5’s hourly temporal resolution. The SCCNN method’s average performance yielded a score of 0.3157, notably higher than the GSSC method’s score of 0.1693, thus establishing a performance advantage of 0.1464 for SCCNN. Further details on the R2 scores are provided in Table 2.
We conducted error estimations for both the SCCNN result and the traditional GSSC method from a randomly selected set of 100 wind profiles. In the case of the SCCNN, we estimated the error based on its performance during simulation, while the error of the GSSC was derived from real signals utilizing Formula 6 outlined in Section 2.2.1. Our evaluation reveals that the error bar reaches 10.4 m/s for SCCNN and 11.5 m/s for GSSC at the 30 km mark on average. Notably, the error increments start escalating sharply beyond 40 km, and the SCCNN changes at a slower rate than that of the GSSC. Specifically, at 40 km, the SD of the error for SCCNN is around 16.2 m/s, increasing to 27.5 m/s at 43 km. In contrast, the SD of GSSC’s error is estimated at about 24.3 m/s at 40 km, rising to 35.3 m/s at 43 km. This comparative analysis underscores a significant enhancement in accuracy attributable to our methodology.

4. Discussion

In this study, we investigated the performance of convolutional neural network (SCCNN, AMCNN) and traditional spectral centroid (SCR, GSSC) methods for wind profile signal processing of an ICDL. Our results show that the SCCNN method outperforms the GSSC method with a mean score of 0.6071 compared to 0.4702 in simulation; and a score of 0.3157 compared to 0.1693 in field measurement compared to ERA5 data. This is particularly evident at higher altitudes, where the LiDAR signal with the traditional algorithm is insufficient to provide reliable wind field information.
Moreover, the SCCNN method demonstrates superior overall performance in accomplishing this task. Despite these findings, it is important to acknowledge the limitations of our work. Firstly, the wind profile generated or detected in this study does not represent the true wind field flowing in the atmosphere. Instead, we adopted the average wind speed within each altitude level based on the range gate’s resolution, resulting in reduced accuracy in the spatial domain.
Additionally, during the wind profile reconstruction, we made several assumptions: (1) we assumed the random error to be 2 m/s [18] to include the vertical wind perturbation and other unconsidered aspects that may affect our measurement; (2) we assumed that the wind profile mainly consisted of perturbations from gravity waves, allowing us to separate and consider the perturbation and background individually during the wind profile reconstruction; and (3) we assumed that the amplitude of the wind perturbation at a single altitude level followed an idealized Gaussian distribution. These assumptions should be taken into account when interpreting the results of this study, and further improvements can be made to address these limitations.
Lastly, in Section 3.3, our LiDAR observations could potentially be verified through additional measurements from balloons or rockets. Currently, our analysis is confined to comparing the outcomes of SCCNN and GSSC against the ERA5 dataset. It is important to note that the ERA5 dataset may not accurately capture the original wind field and could include errors that surpass those in our observations. Moreover, the ERA5 dataset differs from our LiDAR in both temporal and spatial resolutions. Specifically, our LiDAR data are collected over 30 min, whereas ERA5 operates on a one-hour temporal resolution. The spatial resolution of the ERA5 dataset is 0.25° × 0.25° for both longitude and latitude, which does not align exactly with our observation site. Therefore, it should be recognized that the measurements from our LiDAR and the ERA5 dataset do not originate from an identical wind field. Despite these differences, such discrepancies are not deemed crucial for the purposes of our analysis.

5. Conclusions

In conclusion, this study highlights the transformative potential of deep learning in enhancing ICDL data processing research methodologies. We commenced with a rigorous analysis of LiDAR data alongside the ERA5 dataset, utilizing these as foundational elements for our wind profile simulations. These simulations were instrumental in training a customized U-Net architecture, specifically tailored for the task of wind profile reconstruction. The efficacy of our approach was validated by comparing the performance of our U-Net model against the traditional SC algorithm, with evaluations conducted on both the simulated dataset and the actual LiDAR signal data. Our findings indicate that the U-Net model outperforms the SC algorithm significantly, particularly in low signal-to-noise ratio environments.
In our forthcoming research, we anticipate proposing several improvements based on the work presented herein. Specifically, this study focused solely on analyzing the zonal wind, given that the zonal wind at Xinjiang exhibits a highly characteristic profile; the wind field in this region mainly produces robust eastward wind [18,32,52] at the observation time and height. Subsequent analysis of the meridional wind will enable us to combine the two wind profiles, thereby obtaining full wind speed and direction. In our verification process, incorporating measurements from additional instruments could enhance the comparative analysis, thereby increasing the reliability and validity of the LiDAR measurement. Moreover, while the U-net architecture presented in this paper is relatively simple, more extensive comparisons and fine-tuning of different convolutional neural network architectures could prove to be a fruitful avenue for further investigation. Another approach is to expand our findings by considering the effects in the time domain; with a fixed data acquisition rate of the LiDAR, the CNN can be adopted to generate contemporary wind profiles between two consecutive LiDAR detections, thus increasing the LiDAR’s detection rate in time resolution. Finally, we plan to integrate a field-programmable gate array (FPGA) into our LiDAR’s hardware to enhance the speed of both our CNN’s training and operation. We believe that integrating such advanced technologies will revolutionize ICDL data processing research, setting a new benchmark for future studies.

Author Contributions

Conceptualization, J.L., C.C. and Y.H.; methodology, J.L. and C.C.; software, J.L.; validation, C.C.; formal analysis, J.L.; investigation, J.L., H.L., S.Z. and J.Y.; resources, C.C., T.C., X.X. and D.S.; data curation, J.L. and C.C.; writing—original draft preparation, J.L.; writing—review and editing, J.L., C.C., Y.H. and D.S.; visualization, J.L.; supervision, D.S.; project administration, D.S.; funding acquisition, D.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (Grant Nos. 42374182 and 42104146) and the Innovation Program for Quantum Science and Technology (Grant No. 2021ZD0300302).

Data Availability Statement

The USTC lidar wind profiler data presented in this paper are publicly available via “Chen, C., & Zhang, N. (2022). Wind observations of USTC Rayleigh Doppler Lidar in 2019 (Xinjiang) [Dataset]. Science Data Bank. https://doi.org/10.12176/01.99.02468” [30]. The ECMWF/ERA5 dataset can be downloaded from https://cds.climate.copernicus.eu/.

Acknowledgments

The authors would like to express their deepest gratitude to Dongsong Sun, Yuli Han and Xianghui Xue, whose invaluable guidance, expertise, and mentorship contributed immeasurably to the successful completion of this work.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. 2012, 60, 1097–1105. [Google Scholar] [CrossRef]
  2. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  3. He, K.Z.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  4. Hu, J.; Shen, L.; Albanie, S.; Sun, G.; Wu, E. Squeeze-and-Excitation Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  5. Iandola, F.N.; Han, S.; Moskewicz, M.W.; Ashraf, K.; Dally, W.J.; Keutzer, K. SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and <0.5 MB model size. arXiv 2016, arXiv:1602.07360. [Google Scholar]
  6. Howard, A.; Pang, R.; Adam, H.; Le, Q.; Sandler, M.; Chen, B.; Wang, W.; Chen, L.C.; Tan, M.; Chu, G.; et al. Searching for MobileNetV3. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 1314–1324. [Google Scholar]
  7. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  8. Devlin, J.; Chang, M.-W.; Lee, K.; Toutanova, K. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. arXiv 2018, arXiv:1810.04805. [Google Scholar]
  9. Radford, A.; Wu, J.; Child, R.; Luan, D.; Amodei, D.; Sutskever, I. Language models are unsupervised multitask learners. OpenAI Blog 2019, 1, 8. [Google Scholar]
  10. Brown, T.B.; Mann, B.; Ryder, N.; Subbiah, M.; Kaplan, J.; Dhariwal, P.; Neelakantan, A.; Shyam, P.; Sastry, G.; Askell, A.; et al. Language Models are Few-Shot Learners. Adv. Neural Inf. Process. Syst. 2020, 33, 1877–1901. [Google Scholar]
  11. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  12. Kliebisch, O.; Uittenbosch, H.; Thurn, J.; Mahnke, P. Coherent Doppler wind lidar with real-time wind processing and low signal-to-noise ratio reconstruction based on a convolutional neural network. Opt. Express 2022, 30, 5540–5552. [Google Scholar] [CrossRef] [PubMed]
  13. Mohandes, M.A.; Rehman, S. Wind Speed Extrapolation Using Machine Learning Methods and LiDAR Measurements. IEEE Access 2018, 6, 77634–77642. [Google Scholar] [CrossRef]
  14. Song, Y.; Su, Z.; Chen, C.; Sun, D.; Chen, T.; Xue, X. Denoising coherent Doppler lidar data based on a U-Net convolutional neural network. Appl. Opt. 2024, 63, 275–282. [Google Scholar] [CrossRef]
  15. Huffaker, R.M.; Hardesty, R.M. Remote sensing of atmospheric wind velocities using solid-state and CO2 coherent laser systems. Proc. IEEE 1996, 84, 181–204. [Google Scholar] [CrossRef]
  16. Huffaker, R.M.; Jelalian, A.V.; Thomson, J.A.L. Laser-Doppler system for detection of aircraft trailing vortices. Proc. IEEE 1970, 58, 322–326. [Google Scholar] [CrossRef]
  17. Baumgarten, K.; Gerding, M.; Lübken, F.-J. Seasonal variation of gravity wave parameters using different filter methods with daylight lidar measurements at midlatitudes. J. Geophys. Res. Atmos. 2017, 122, 2683–2695. [Google Scholar] [CrossRef]
  18. Dou, X.; Han, Y.; Sun, D.; Xia, H.; Shu, Z.; Zhao, R.; Shangguan, M.; Guo, J. Mobile Rayleigh Doppler lidar for wind and temperature measurements in the stratosphere and lower mesosphere. Opt. Express 2014, 22, A1203–A1221. [Google Scholar] [CrossRef] [PubMed]
  19. Chanin, M.L.; Garnier, A.; Hauchecorne, A.; Porteneuve, J. A Doppler lidar for measuring winds in the middle atmosphere. Geophys. Res. Lett. 1989, 16, 1273–1276. [Google Scholar] [CrossRef]
  20. Besson, A.; Canat, G.; Lombard, L.; Valla, M.; Durécu, A.; Besson, C. Long-range wind monitoring in real time with optimized coherent lidar. Opt. Eng. 2017, 56, 031217. [Google Scholar] [CrossRef]
  21. Baumgarten, G.; Fiedler, J.; Hildebrand, J.; Lübken, F.J. Inertia gravity wave in the stratosphere and mesosphere observed by Doppler wind and temperature lidar. Geophys. Res. Lett. 2015, 42, 10929–10936. [Google Scholar] [CrossRef]
  22. Ehard, B.; Kaifler, B.; Kaifler, N.; Rapp, M. Evaluation of methods for gravity wave extraction from middle-atmospheric lidar temperature measurements. Atmos. Meas. Tech. 2015, 8, 4645–4655. [Google Scholar] [CrossRef]
  23. Yamashita, C.; Chu, X.; Liu, H.L.; Espy, P.J.; Nott, G.J.; Huang, W. Stratospheric gravity wave characteristics and seasonal variations observed by lidar at the South Pole and Rothera, Antarctica. J. Geophys. Res. Atmos. 2009, 114. [Google Scholar] [CrossRef]
  24. Duck, T.J.; Whiteway, J.A.; Carswell, A.I. The gravity wave–Arctic stratospheric vortex interaction. J. Atmos. Sci. 2001, 58, 3581–3596. [Google Scholar] [CrossRef]
  25. Chane-Ming, F.; Molinaro, F.; Leveau, J.; Keckhut, P.; Hauchecorne, A. Analysis of gravity waves in the tropical middle atmosphere over La Reunion Island (21 S, 55 E) with lidar using wavelet techniques. Ann. Geophys. 2000, 18, 485–498. [Google Scholar] [CrossRef]
  26. Mzé, N.; Hauchecorne, A.; Keckhut, P.; Thétis, M. Vertical distribution of gravity wave potential energy from long-term Rayleigh lidar data at a northern middle-latitude site. J. Geophys. Res. Atmos. 2014, 119, 12069–12083. [Google Scholar] [CrossRef]
  27. Alexander, M.; Geller, M.; McLandress, C.; Polavarapu, S.; Preusse, P.; Sassi, F.; Sato, K.; Eckermann, S.; Ern, M.; Hertzog, A. Recent developments in gravity-wave effects in climate models and the global distribution of gravity-wave momentum flux from observations and models. Q. J. R. Meteorol. Soc. 2010, 136, 1103–1124. [Google Scholar] [CrossRef]
  28. Flesia, C.; Korb, C.L. Theory of the double-edge molecular technique for Doppler lidar wind measurement. Appl. Opt. 1999, 38, 432–440. [Google Scholar] [CrossRef] [PubMed]
  29. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015, Munich, Germany, 5–9 October 2015. [Google Scholar] [CrossRef]
  30. Chen, C.; Zhang, N. Wind Observations of USTC Rayleigh Doppler Lidar in 2019 (Xinjiang); Science Data Bank: Beijing, China, 2022. [Google Scholar] [CrossRef]
  31. Zhang, N.; Sun, D.; Han, Y.; Chen, C.; Wang, Y.; Zheng, J.; Liu, H.; Han, F.; Zhou, A.; Tang, L. Zero Doppler correction for Fabry–Pérot interferometer-based direct-detection Doppler wind LIDAR. Opt. Eng. 2019, 58, 054101. [Google Scholar] [CrossRef]
  32. Chen, C.; Xue, X.; Sun, D.; Zhao, R.; Han, Y.; Chen, T.; Liu, H.; Zhao, Y. Comparison of Lower Stratosphere Wind Observations From the USTC’s Rayleigh Doppler Lidar and the ESA’s Satellite Mission Aeolus. Earth Space Sci. 2022, 9, e2021EA002176. [Google Scholar] [CrossRef]
  33. Chanin, M.L.; Hauchecorne, A. Lidar observation of gravity and tidal waves in the stratosphere and mesosphere. J. Geophys. Res. Ocean. 1981, 86, 9715–9721. [Google Scholar] [CrossRef]
  34. Garnier, A.; Chanin, M.L. Description of a Doppler rayleigh LIDAR for measuring winds in the middle atmosphere. Appl. Phys. B 1992, 55, 35–40. [Google Scholar] [CrossRef]
  35. Zhao, R.; Dou, X.; Sun, D.; Xue, X.; Zheng, J.; Han, Y.; Chen, T.; Wang, G.; Zhou, Y. Gravity waves observation of wind field in stratosphere based on a Rayleigh Doppler lidar. Opt. Express 2016, 24, A581–A591. [Google Scholar] [CrossRef] [PubMed]
  36. Hersbach, H.; Bell, B.; Berrisford, P.; Hirahara, S.; Horányi, A.; Muñoz-Sabater, J.; Nicolas, J.; Peubey, C.; Radu, R.; Schepers, D.; et al. The ERA5 global reanalysis. Q. J. R. Meteorol. Soc. 2022, 146, 1999–2049. [Google Scholar] [CrossRef]
  37. ERA5: Data Documentation-Copernicus Knowledge Base-ECMWF Confluence Wiki. Available online: https://confluence.ecmwf.int/display/CKB/ERA5%3A+data+documentation#ERA5:datadocumentation-Observations (accessed on 9 March 2023).
  38. Sun, D.; Zhong, Z.; Zhou, J.; Hu, H.; Kobayashi, T. Accuracy Analysis of the Fabry–Perot Etalon Based Doppler Wind Lidar. Opt. Rev. 2005, 12, 409–414. [Google Scholar] [CrossRef]
  39. Korb, C.L.; Gentry, B.M.; Weng, C.Y. Edge technique: Theory and application to the lidar measurement of atmospheric wind. Appl. Opt. 1992, 31, 4202–4213. [Google Scholar] [CrossRef]
  40. Baumgarten, K.; Gerding, M.; Baumgarten, G.; Lübken, F.J. Temporal variability of tidal and gravity waves during a record long 10-day continuous lidar sounding. Atmos. Chem. Phys. 2018, 18, 371–384. [Google Scholar] [CrossRef]
  41. Senf, F.; Achatz, U. On the impact of middle-atmosphere thermal tides on the propagation and dissipation of gravity waves. J. Geophys. Res. Atmos. 2011, 116. [Google Scholar] [CrossRef]
  42. Eckermann, S.D.; Marks, C.J. An idealized ray model of gravity wave-tidal interactions. J. Geophys. Res. Atmos. 1996, 101, 21195–21212. [Google Scholar] [CrossRef]
  43. Alexander, M.; Holton, J. On the spectrum of vertically propagating gravity waves generated by a transient heat source. Atmos. Chem. Phys. 2004, 4, 923–932. [Google Scholar] [CrossRef]
  44. Holton, J.R. The Influence of Gravity Wave Breaking on the General Circulation of the Middle Atmosphere. J. Atmos. Sci. 1983, 40, 2497–2507. [Google Scholar] [CrossRef]
  45. Kreyszig, E.; Stroud, K.; Stephenson, G. Advanced engineering mathematics. Integration 2008, 9, 1045. [Google Scholar]
  46. Vadas, S.L.; Fritts, D.C.; Alexander, M.J. Mechanism for the Generation of Secondary Waves in Wave Breaking Regions. J. Atmos. Sci. 2003, 60, 194–214. [Google Scholar] [CrossRef]
  47. Andreassen, Ø.; Wasberg, C.E.; Fritts, D.C.; Isler, J.R. Gravity wave breaking in two and three dimensions: 1. Model description and comparison of two-dimensional evolutions. J. Geophys. Res. Atmos. 1994, 99, 8095–8108. [Google Scholar] [CrossRef]
  48. Fritts, D.C.; Isler, J.R.; Andreassen, Ø. Gravity wave breaking in two and three dimensions: 2. Three-dimensional evolution and instability structure. J. Geophys. Res. Atmos. 1994, 99, 8109–8123. [Google Scholar] [CrossRef]
  49. Agarap, A.F. Deep Learning using Rectified Linear Units (ReLU). arXiv 2019, arXiv:1803.08375. [Google Scholar]
  50. Nagi, J.; Ducatelle, F.; Di Caro, G.A.; Ciresan, D.; Meier, U.; Giusti, A.; Nagi, F.; Schmidhuber, J.; Gambardella, L.M. Max-pooling convolutional neural networks for vision-based hand gesture recognition. In Proceedings of the 2011 IEEE International Conference on Signal and Image Processing Applications, ICSIPA 2011, Kuala Lumpur, Malaysia, 16–18 November 2011. [Google Scholar]
  51. Kingma Diederik, B.J. Adam: A Method for Stochastic Optimization. In Proceedings of the International Conference on Learning Representations, Banff, AB, Canada, 14–16 April 2014. [Google Scholar]
  52. Zhao, R.; Dou, X.; Xue, X.; Sun, D.; Han, Y.; Chen, C.; Zheng, J.; Li, Z.; Zhou, A.; Han, Y.; et al. Stratosphere and lower mesosphere wind observation and gravity wave activities of the wind field in China using a mobile Rayleigh Doppler lidar. J. Geophys. Res. Space Phys. 2017, 122, 8847–8857. [Google Scholar] [CrossRef]
Figure 1. Mobile ICDL system developed by USTC.
Figure 1. Mobile ICDL system developed by USTC.
Remotesensing 16 01473 g001
Figure 2. LiDAR measured wind profile vs. background wind profile (east–west orientation) of 20 October 2019, at 10 p.m.
Figure 2. LiDAR measured wind profile vs. background wind profile (east–west orientation) of 20 October 2019, at 10 p.m.
Remotesensing 16 01473 g002
Figure 3. (a) Altitude vs. amplitude of the wind perturbation; (b) altitude vs. wind speed error of the LiDAR detection in logarithmic scale.
Figure 3. (a) Altitude vs. amplitude of the wind perturbation; (b) altitude vs. wind speed error of the LiDAR detection in logarithmic scale.
Remotesensing 16 01473 g003
Figure 4. Three-channeled input of the CNN (simulated LiDAR signal). Top: Raw data scatter plot (spectral profile) of the simulated LiDAR signal. Middle: SC wind profile calculated based on the simulated LiDAR signal. Bottom: Spline Transformer wind profile calculated based on the simulated LiDAR signal.
Figure 4. Three-channeled input of the CNN (simulated LiDAR signal). Top: Raw data scatter plot (spectral profile) of the simulated LiDAR signal. Middle: SC wind profile calculated based on the simulated LiDAR signal. Bottom: Spline Transformer wind profile calculated based on the simulated LiDAR signal.
Remotesensing 16 01473 g004
Figure 5. Proposed Customized U-net Structure.
Figure 5. Proposed Customized U-net Structure.
Remotesensing 16 01473 g005
Figure 6. LiDAR data spectral profile of the real LiDAR signal as the input of the CNN.
Figure 6. LiDAR data spectral profile of the real LiDAR signal as the input of the CNN.
Remotesensing 16 01473 g006
Figure 7. Top: Output of the CNN. Bottom: The ground truth label.
Figure 7. Top: Output of the CNN. Bottom: The ground truth label.
Remotesensing 16 01473 g007
Figure 8. Wind profile plots of average performance of the validation set: (a) Argmax plot of the CNN against ground truth; (b) SC plot of the CNN against ground truth; (c) SC plot of raw data against ground truth; (d) Gaussian-smoothed SC plot of raw data against ground truth.
Figure 8. Wind profile plots of average performance of the validation set: (a) Argmax plot of the CNN against ground truth; (b) SC plot of the CNN against ground truth; (c) SC plot of raw data against ground truth; (d) Gaussian-smoothed SC plot of raw data against ground truth.
Remotesensing 16 01473 g008aRemotesensing 16 01473 g008b
Figure 9. Regression analysis of average performance: (a) Regression analysis of GSSC against ground truth; (b) regression analysis of SCCNN against ground truth.
Figure 9. Regression analysis of average performance: (a) Regression analysis of GSSC against ground truth; (b) regression analysis of SCCNN against ground truth.
Remotesensing 16 01473 g009
Figure 10. East–west wind profile measurement of the GSSC vs. the SCCNN on 31 October 2019, at Kolar, Xinjiang (positive wind speed corresponds to east wind). The horizontal axis shows the timeline in HH:MM format, and the color indicates the wind speed and direction. The white blank region indicates the LiDAR signal is below sufficient SNR threshold for SC calculation. Top: Wind profile measurement of the GSSC. Bottom: Wind profile measurement of the SCCNN.
Figure 10. East–west wind profile measurement of the GSSC vs. the SCCNN on 31 October 2019, at Kolar, Xinjiang (positive wind speed corresponds to east wind). The horizontal axis shows the timeline in HH:MM format, and the color indicates the wind speed and direction. The white blank region indicates the LiDAR signal is below sufficient SNR threshold for SC calculation. Top: Wind profile measurement of the GSSC. Bottom: Wind profile measurement of the SCCNN.
Remotesensing 16 01473 g010
Figure 11. Wind profile plots of east–west orientation at 2:00 a.m. on 1 November 2019, at Kolar, Xinjiang: (a) wind profile plots of SCCNN vs. GSSC vs. ERA5; (b) background wind profile plots of SCCNN vs. GSSC vs. ERA5.
Figure 11. Wind profile plots of east–west orientation at 2:00 a.m. on 1 November 2019, at Kolar, Xinjiang: (a) wind profile plots of SCCNN vs. GSSC vs. ERA5; (b) background wind profile plots of SCCNN vs. GSSC vs. ERA5.
Remotesensing 16 01473 g011
Table 1. Score List of Coefficient of Determination for all Four Methods.
Table 1. Score List of Coefficient of Determination for all Four Methods.
MethodMeanMax.Min.
AMCNN0.01260.5564−0.5051
SCCNN0.61410.82320.3119
SCR−0.16800.5530−0.9087
GSSC0.47020.77790.0241
Table 2. Score List of Coefficient of Determination for ERA5 Data’s Comparative Analysis.
Table 2. Score List of Coefficient of Determination for ERA5 Data’s Comparative Analysis.
TimeSCCNNGSSC
31 October, 9:00 p.m.0.45360.3068
10:00 p.m.0.46690.3676
11:00 p.m.0.33120.2336
1 November, 00:00 a.m.0.09850.1552
1:00 a.m.−0.0429−0.5710
2:00 a.m.0.76530.6915
3:00 a.m.0.13710.0010
Mean0.31570.1693
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, J.; Chen, C.; Han, Y.; Chen, T.; Xue, X.; Liu, H.; Zhang, S.; Yang, J.; Sun, D. Wind Profile Reconstruction Based on Convolutional Neural Network for Incoherent Doppler Wind LiDAR. Remote Sens. 2024, 16, 1473. https://doi.org/10.3390/rs16081473

AMA Style

Li J, Chen C, Han Y, Chen T, Xue X, Liu H, Zhang S, Yang J, Sun D. Wind Profile Reconstruction Based on Convolutional Neural Network for Incoherent Doppler Wind LiDAR. Remote Sensing. 2024; 16(8):1473. https://doi.org/10.3390/rs16081473

Chicago/Turabian Style

Li, Jiawei, Chong Chen, Yuli Han, Tingdi Chen, Xianghui Xue, Hengjia Liu, Shuhua Zhang, Jing Yang, and Dongsong Sun. 2024. "Wind Profile Reconstruction Based on Convolutional Neural Network for Incoherent Doppler Wind LiDAR" Remote Sensing 16, no. 8: 1473. https://doi.org/10.3390/rs16081473

APA Style

Li, J., Chen, C., Han, Y., Chen, T., Xue, X., Liu, H., Zhang, S., Yang, J., & Sun, D. (2024). Wind Profile Reconstruction Based on Convolutional Neural Network for Incoherent Doppler Wind LiDAR. Remote Sensing, 16(8), 1473. https://doi.org/10.3390/rs16081473

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop