Next Article in Journal
Application of Biochar Functionalized with Layered Double Hydroxides: Improved Plant Growth Performance after Use as Phosphate Adsorbent
Next Article in Special Issue
Analysis and Identification of Nonlinear Acoustic Damping in Miniature Loudspeakers
Previous Article in Journal
Lp-Norm Inversion of Gravity Data Using Adaptive Differential Evolution
Previous Article in Special Issue
Application of Machine Learning to Include Honking Effect in Vehicular Traffic Noise Prediction
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Machine Learning-Based Tools for Wind Turbine Acoustic Monitoring

by
Giuseppe Ciaburro
1,
Gino Iannace
1,
Virginia Puyana-Romero
2 and
Amelia Trematerra
1,*
1
Department of Architecture and Industrial Design, Università degli Studi della Campania Luigi Vanvitelli, Borgo San Lorenzo, 81031 Aversa, Italy
2
Department of Sound and Acoustic Engineering, Universidad de Las Américas, Quito EC170125, Ecuador
*
Author to whom correspondence should be addressed.
Appl. Sci. 2021, 11(14), 6488; https://doi.org/10.3390/app11146488
Submission received: 16 June 2021 / Revised: 8 July 2021 / Accepted: 9 July 2021 / Published: 14 July 2021
(This article belongs to the Special Issue Modelling, Simulation and Data Analysis in Acoustical Problems Ⅱ)

Abstract

:
The identification and separation of sound sources has always been a difficult problem for acoustic technicians to tackle. This is due to the considerable complexity of a sound that is made up of many contributions at different frequencies. Each sound has a specific frequency spectrum, but when many sounds overlap it becomes difficult to discriminate between the different contributions. In this case, it can be extremely useful to have a tool that is capable of identifying the operating conditions of an acoustic source. In this study, measurements were made of the noise emitted by a wind turbine in the vicinity of a sensitive receptor. To identify the operating conditions of the wind turbine, average spectral levels in one-third octave bands were used. A model based on a support vector machine (SVM) was developed for the detection of the operating conditions of the wind turbine; then a model based on an artificial neural network was used to compare the performance of both models. The high precision returned by the simulation models supports the adoption of these tools as a support for the acoustic characterization of noise in environments close to wind turbines.

1. Introduction

In recent years, the global energy supply has been characterized by the development of various technologies defined as renewable, which make it possible to obtain energy in a clean and inexhaustible way by exploiting the wind, solar, hydroelectric, and geothermal resources present on Earth. The term “renewable energies” refers to energies that can be regenerated in a short time compared to human history [1]. The sources of these forms of energy are called renewable energy resources. For millennia, wind has represented a form of local energy that can be easily transformed, with examples such as mills and sailing boats illustrating that humans had already recognized its great significance in 2000 BC [2,3].
The machines used to extract energy from wind are wind turbines, ranging from their most archaic forms in ancient mills to the most modern and immense machines with which offshore parks are built, which retain the fundamental characteristic of extracting kinetic energy through blades that rotate around a vertical or horizontal axis [4,5,6]. Turbines can be classified according to two main aspects: aerodynamic operation and construction design. Aerodynamic operation defines the method by which the blades convert airflow into energy. According to this aspect, lift and drag turbines can be distinguished from one another [7,8]. The most common form of turbines are those that use lift; in which, compared to drag turbines, the wind flows on both sides of the blade (which has two different profiles) thus creating a zone of downforce on the upper surface compared to that on the lower surface. The wing profile of the blade leads to different fluid flow velocities on the upper and lower surfaces, which produces the pressure difference [9]. This pressure difference determines a force called aerodynamic lift, similar to that generated on the wings of aircraft. As for turbines based on dragging, they take advantage of the simple aerodynamic pressure that is exerted by the wind impacting the surface of the blades. Finally, classification according to the construction design is based on dividing the turbines into two large groups that are easily distinguishable: turbines with a vertical rotation axis and turbines with a horizontal rotation axis [10].
In the past, the main objective of the use of a wind turbine was to obtain maximum performance, whereas currently, there is a tendency to care about other aspects such as the generated noise, and to include it among the turbine’s specifications [11]. The noise emitted by wind turbines represents one of the most important problems affecting the quality of life of the residents in the vicinity of wind farms, limiting the large-scale diffusion of wind energy. The sound emission produced can be divided into two categories: noise produced by mechanical parts and aerodynamic noise [12]. The greatest contribution to the noise emitted by a wind turbine is attributable to the friction of the air with the blades and the support tower, whereas the modern machinery placed in the nacelle is relatively silent. Aerodynamic noise is caused by the movement of the turbine blades in the air. This motion leads to complex sound generation phenomena, and often, the emission associated with these sources is broadband. Three types of aerodynamic noise can be defined, according to the different mechanisms that cause them. Low-frequency noise is generated when a rotating blade encounters localized flow deficiencies due to the tower, when there is a change in wind speed or the wind spreads between the other blades [13]. Another noise is caused by the turbulent flow, which depends on the increase in atmospheric turbulence that appears as localized forces or pressure variations on the blade. Finally, there is a noise generated by the airfoil, which consists of the sound generated by the flow of air along the surface of the profile. This type of sound is typically broadband, but tonal components can occur due to the smoothing of the edges, or from the action of the airflow on cracks and holes [14].
In evaluating the noise emitted in a residential area, it is essential to identify all the sound sources that contribute to the annoyance caused to residents. Often this procedure is made difficult by the simultaneous contribution of different sources of a different nature which overlap. The acoustic characterization of these sources makes it possible, in some cases, to isolate these contributions, which can thus be removed or added. In this way, it is possible to evaluate the contribution of every single source to the perceived disturbance. When the frequency contributions of different sources overlap, this isolation is no longer possible, at least with the use of classic techniques. In these cases, it becomes essential in the noise disturbance assessment procedure to identify the presence of a specific source within the ambient noise [15].
In recent times, the use of machine learning-based algorithms for the identification of patterns in complex environments has become widespread [16,17,18,19]. At first, these technologies were widely used in the field of computer vision [20,21], but soon proved to be useful for identifying sources in complex sound environments [22,23]. The idea behind the concept of machine learning is to provide computers with the ability to learn and replicate certain operations without having been explicitly programmed. These operations are generally of a predictive or decision-making type and are based on the data available. The machine, therefore, will be able to learn to perform certain tasks by improving, through experience, its skills, responses, and functions. At the basis of machine learning, there are a series of different algorithms that, starting from primitive notions, will process a result. A distinction between the various algorithms is given by the type of learning performed: depending on the training mode, supervised and unsupervised learning can be distinguished. In the first case, the model is trained by providing labeled data as an input, so in addition to the value of the chosen feature, a label with the class to which the data belongs will also be available [24]. On the contrary, to algorithms that exploit unsupervised learning, datasets are given without labels, and it is therefore the program that must find the classes in which to divide the data through features [25]. Then, depending on the different types of task they perform, the algorithms take various names. Classifiers are algorithms that, having input data and the classes to which they belong, can develop a way to understand which class a datum belongs to if they have never seen it before. Another type of algorithm is clustering algorithms, which can divide data taken from inputs without knowing the classes to which they belong. There are also regression algorithms that try to predict the value of data. Using other inputs of the same phenomena, they try to understand their trends by predicting the value of an unknown datum via interpolation [26].
In this study, the measurements of the noise emitted by a wind turbine carried out inside a house were used to develop a classifier capable of discriminating between the operating conditions of the turbine. To do this, a model based on artificial neural networks has been built which uses the sound pressure levels at different frequencies as inputs and returns the operating conditions of the turbine. The article is developed as follows: Section 2 describes in detail the methodologies used in the study: first, the instruments used to carry out the acoustic measurements and the noise detection techniques are described, along with the algorithms used to elaborate the numerical simulation model based on the support vector machine (SVM). In Section 3 the results obtained from the measurements of the noise emitted by the wind turbines and subsequently the results obtained through the numerical simulation are reported and discussed in detail. Finally, Section 4 summarizes the results obtained from this study and discusses possible uses of the technology developed in real cases.

2. Materials and Methods

The wind farm under study is located in a hilly area of southern Italy with an agricultural vocation, where the cultivation of durum wheat, soft wheat, tobacco, sunflower, vegetables, and other minor crops has developed. The presence of tree crops is concentrated in some niches of the territory, and on small or very reduced surfaces (Figure 1).
In this area, there is a wind farm with several wind turbines. The turbines included in this study have a horizontal axis structure with a nominal power of 2 MW [27]. The towers are of a monopolar tubular type, with a hub height of 80 m, equipped with three blades with a diameter of 95 m, made of glass fibers. The maximum power is obtained in a speed range from 10.5 m/s to 25 m/s. Beyond this speed, the blades stop for safety reasons.
Acoustic measurements were carried out to evaluate the noise emissions by the wind turbines under study. The measurements were carried out inside a private house near the wind turbines, with windows open in order to detect the conditions of maximum disturbance. The instrument used for the noise detection was positioned about 1.2 m from the window and placed on a tripod at a height of 1.60 m from the ground. The room where the measurements were made was 3 m high, 4 m long, and 5 m wide, and was normally furnished [28,29,30].
The following instrumentation was used: a LXT1 Larson Davis integrating sound level meter model of “Class 1”, and a Larson Davis CAL 200 Calibrator. The instrumentation used complied with the requirements of the IEC61400-11 standard for wind turbines [31]; the sound level meter was configured for the acquisition of the equivalent level of linear sound pressure, weighted “A”, of the statistical levels with a fast time constant and of the frequency spectra of 1/3 of an octave.
The acoustic measurements were performed in different measurement sessions spread over several days. For every measurement session, measurements were made in different operating conditions of the turbines to assess the contribution made to the sound field by the operation of each wind turbine. The following operating conditions were identified and classified:
  • 0 = All turbines off;
  • 1 = All turbines on;
  • 2 = One turbine off;
  • 3 = Two turbines off.
The measured sound levels were compared in order to identify the operating conditions of the wind farm. To do this, the time history and the third-octave frequency spectra of the average and minimum values of the acoustic measurements were analyzed.
The results of the sound level measurements were used as inputs to train a classifier based on the support vector machine algorithm. The classifier will allow us to recognize the operating conditions of the wind turbine, thus providing the technician with the possibility of identifying scenarios in which the sources are turned on/off.
Classification consists of identifying to which category a newly observed datum belongs, based on a set of data known as training data. An algorithm that implements a classification method based on the observations found during training is called a classifier. This term is also used in mathematics to indicate a function, implemented by a classification algorithm, which maps the input data into a category [32]. In the case of support vector machines, classification is considered an instance of supervised learning. Classification is therefore an example of the more general problem of pattern recognition, which consists in assigning some kind of output value to a given input value. To obtain a clearer idea of what the purpose of the classification is, let us consider the case of binary classification and represent the data on a plane. We have N data, also called training points, where each input datum has a D attribute, so it is a vector of dimension D, and belongs to one of the two classes.
The training data depicted in Figure 2 consist of several instances belonging to two different classes. These instances are separable by straight lines, so they are linearly separable. Conversely, when instances of a data collection are not separable by straight lines, they are separable in a non-linear way. The data represented in the plan can appear to us, at this point, as being distributed in two different ways. We may be faced with linearly separable data or non-linearly separable data. In the first case it will be possible to draw in the graph a line (if D = 2) or a hyperplane (if D > 2) that exactly divides the two classes (Figure 2). In the first reported case, we can intuitively draw a straight line dividing the two classes, indicated with the colors blue and red. In the second case it is impossible to draw a line that divides the two classes in half because some points would be found on the wrong side of the separator line [33].
An SVM-based classification technique allows the classification of both linear and non-linear data collections. An SVM represents all instances of the training data collection on a plane consisting of a number of axes (dimensions) equal to the number of attributes that make up the instances. For example, if an instance consists of three attributes, then the training data will be represented on a three-dimensional plane. The three main characteristics of an SVM classifier are: (1) lines or hyperplanes, depending on whether the classifier has a two-dimensional or n-dimensional graph; (2) margins; and (3) support vectors. A line or a hyperplane constitutes a boundary that allows one to classify the instances belonging to different classes by dividing them between the classes. A margin is a distance between the two closest instances of different classes.
Instead, support vectors represent the most difficult instances to classify for an SVM, since they are found within the boundaries of a hyperplane. An SVM allows both the classification of data collections that are linearly separable between the various classes and those of which the instances are separable in a non-linear way. The instances of a data collection are linearly separable if, after having represented them on a plane, they are separable by one or more straight lines.
An SVM is applied differently depending on the type of instances of a collection of data to be classified, whether they are linearly separable or not. In the presence of linearly separable instances, it is necessary to find—among all the straight lines or among all the hyperplanes that separate them between the different classes—those that maximize the margin value. In fact, the line or hyperplane with the maximum margin value is selected since it allows us to minimize the classification error.
Regarding those collections of data that can be separated in a non-linear manner, the classification procedure is more complex, as it is necessary to operate in two separate phases and extend the behavior of the previous approach. In the first phase, the instances of the data collection are mapped to a larger dimensional space in order to make them separable in a linear way. Finally, in the second phase it is possible to use the previous approach to search for a line or a hyperplane that maximizes the size of the margin, since now the instances are linearly separable.
To use the classification through hyperplanes for data that would require non-linear functions to be separated, it is necessary to resort to the technique of feature spaces. This method, which is the basis of SVM theory, consists in mapping the initial data into a space of a higher dimension. Therefore, assuming m > n, for the map we use the function expressed in Equation (1):
f :   R n     R m                    
In Equation (1):
  • x and y are input vectors
  • f is a transformation function
The technique of image spaces is particularly interesting for algorithms that use training data only through scalar products. In this case, in the space Rm we do not have to find f (x) and f (y) explicitly but it is enough to calculate their scalar product f (x) · f (y). The function f is used to map the input from the n-dimensional space to the m-dimensional space, at a higher level than the n level. To make this last calculation simple—as it becomes very complicated in large spaces—a function called a kernel is used, which directly returns the scalar product of the images (Equation (2)).
K   ( x , y ) = f ( x )   · f ( y )            
In Equation (2), the · denotes the dot product operation that takes the dot product of the transformed input vectors. The goal of the kernel function is to use the data as an input and transform it into the required form if it is not possible to determine a linearly separable hyperplane, as is the case in most cases [34].
The most used kernels are:
  • Linear:
K ( x ,   y ) =   x   ·   y
This is the simplest type of kernel.
  • Polynomial:
K ( x ,   y ) = ( x   ·   y ) d  
This kernel contains a parameter: a degree of freedom d. A d value of 1 represents a linear kernel. A value greater than d will make the decision limit more complex and could lead to data overfitting.
  • Radial Basis function (RBF):
K ( x , y ) =   exp ( x   ·   y 2 2 σ 2 )
The value of this type of kernel decreases with distance and varies between zero and one; for this reason, it can be considered a measure of similarity.
  • Sigmoid:
K ( x ,   y ) =   tan h ( α   x · y δ )
There are two adjustable parameters in the sigmoid kernel, the alpha slope and the intercept constant δ . A common value for alpha is 1/N, where N is the size of the data.
For SVMs, both training and error evaluation are fundamental. For this reason, the data are divided into training and testing sets. The first group is used for algorithm training and contains the correct inputs and outputs, as supervised training is required. It usually consists of about 70% of the total data. The second is used for testing, and consists of a set of data containing 30% of the unused data, that is, to evaluate the accuracy of the neural network by calculating the prediction error chosen. This phase also allows the testing the algorithm; thus, it consists in the real use of the designed model.
One of the advantages of a support vector machine-based classifier is the ability to solve difficult and non-linear classification problems and ensure a high level of classification accuracy. As for solving simple problems, however, the accuracy is comparable to that of a classification technique based on a decision tree. Among the disadvantages are the high model creation time, although it is lower than that used by a neural network, and the non-interpretability of the model.

3. Results and Discussion

The measurements were carried out during May 2018. In the monitoring period, average wind speeds ranging between 9.0 m/s and 23.0 m/s were detected. During the measurements, the turbines were switched on and off according to the sequence shown in Table 1. During the measurements, the following operating conditions were monitored: all turbines off, all turbines on, one turbine off, and two turbines off. The monitoring period covered approximately 18 h of overall measurements. Figure 3 shows the time history of the sound pressure level for the operating conditions performed in the measurement session. The different operating conditions have been highlighted in the following paragraphs to better appreciate the differences in the measured levels from a visual point of view.
Analyzing Figure 3, it is possible to identify the operating conditions indicated above, at least in the first part of the measurements where the wind speed was kept at low levels. This indicates that the source characterizing the background noise is represented by the wind turbine under analysis. The background noise, in fact, undergoes a substantial decrease in the period in which the wind turbine is switched off. It is also possible to verify that under the same operating conditions, the recorded noise changes over time. This fact in the case of turbines in operation can be attributed to the different speeds of the blades, which generate different contributions to the environmental noise. On the contrary, in the conditions with the turbine off, this different trend of background noise confirms that the contributions to the environmental noise have a different origin.
A considerable increase in the noise detected by the instrument can be seen in the central part of Figure 3. By analyzing the measured values of the wind speed in this monitoring period (Table 1) it is possible to verify that these speeds undergo a significant increase. This helps us to understand that the increase in noise is due to the noise of the blades which, turning with greater speed due to the greater energy supplied by the wind, produce greater noise. However, analyzing the final part of the monitoring where all the turbines have been turned off, we can see that the background noise with the turbines off is much greater than that recorded in the initial part of the monitoring. This confirms the essential contribution of sources other than wind turbines to the background noise. Among these, the noise of the wind demonstrates a significant weight, impacting on trees or buildings and on the pavement and thus contributing to the background noise.
Figure 4 shows the average spectral levels in the one-third octave bands between 50 Hz and 5 kHz for the entire measurement session.
From the analysis of Figure 4, it is not possible to obtain useful information to characterize the sources. The noise is distributed over the entire frequency range, with values ranging from 40 dB to about 50 dB. There are no tonal components that can identify a specific source. In order to extract further information, it is necessary to individually analyze the portions of measurements relating to the various operating conditions.
The results of the measurements were subsequently elaborated in order to look for the characteristics that were able to discriminate between the different operating conditions of the wind turbines. According to the analysis of Figure 3, fourteen periods were identified (Table 1). Table 1 shows the durations and the acoustic measurements for the operating conditions performed in the measurement session. Based on the analysis of the results, it is possible to note that the different operating conditions are confirmed by the noise levels recorded by the sound level meter. The operating conditions with the turbines on returned the following values: 38.4, 40.2, 50.3, 43.7, and 45.8 dBA. The operating conditions with the turbines off, however, returned the following values: 24.9, 40.7, 50.8, and 36.7 dBA. These results show that the extreme operating conditions recorded noise values conditioned by the wind speed, demonstrating that the contribution of the wind to the background noise became significant under certain circumstances.
The selected periods do not show anomalous or occasional phenomena. Therefore, they can be used for setting the classification model. To build a classification model of the operating conditions of the wind turbine, the measured data were resampled by setting 1 s as the integration time. For each observation, the A-weighted equivalent sound pressure levels for each one-third octave band were calculated. 64,860 observations were collected, and 21 variables were selected. To characterize the three periods, we decided to use the average spectral levels in one-third octave bands between 50 Hz and 5 kHz as descriptors.
Figure 5 shows the average spectral levels in the third-octave bands between 50 Hz and 5 kHz for some of the observation periods identified. To facilitate the comparison, the periods of operation of the turbines have been placed side by side with periods of inactivity. From the analysis of the results shown in Figure 5, it is possible to note that the operating conditions are not easily identifiable. We can see that the conditions in which the turbines are switched off have lower values at high frequencies. However, not all frequencies contribute to the identification of the operating conditions. This indicates that not all frequencies are required for the model setup.
Analyzing the information contained in Table 1, we can confirm that in the four monitoring periods shown in Figure 5, the wind conditions remained almost constant. In fact, in all four measurements the wind speed remained approximately equal to 9.0 m/s (9.0–9.2). These boundary conditions allow us to be able to make a comparison between the different operating conditions of the wind farm. Figure 5a shows the average spectrum detected with all turbines on. A considerable energy contribution can be appreciated at both low and medium frequencies, with peaks recorded at 50 Hz and 125 Hz. The operating conditions with all turbines on showed the highest spectral values. Figure 5b shows the average spectrum detected with only one turbine off; in this case the energy contribution was also considerable, and no large differences were detected, at least through a visual analysis of the large differences with the operating conditions with all turbines on. Furthermore, there were peaks corresponding to the frequencies at 50 Hz and 125 Hz, but these showed lower values. Figure 5c shows the average spectrum detected with two turbines off; in this operating condition the energy contribution was significantly reduced. Once again, a peak at 50 Hz was recorded, although it was of a lower value, as well as a second peak at 200 Hz. Finally, Figure 5d shows the average spectrum detected with all turbines off; in this case, the energy contribution measured by the instrument was significantly decreased compared to the case where all turbines were on. Furthermore, it is possible to note that a peak at 50 Hz was confirmed, whereas the contribution to the medium frequencies that was recorded in the other three operating conditions was completely absent.
In addition to the extreme frequencies that have already been excluded from our analysis—having considered only the frequencies between 50 Hz and 5 Khz—there may be other frequencies that are not essential for the correct identification of the operating conditions. To identify the characteristics that are able to discriminate between the different operating conditions, a visual analysis was conducted through a wrapper for density lattice plots (Figure 6).
The bell curves show the distribution set at each frequency for the four identified operating conditions. As already indicated, extreme frequencies were left out of this analysis, which focused on the frequency range between 50 Hz and 5 kHz. In this frequency range, it can be observed that some discriminate these other labeled operating conditions better than others. For these components, it can be noted that the four operating conditions assume different distributions, but these are not easily identifiable, at least not with a purely visual approach. This is because in all cases the distributions overlap; this indicates that in these ranges of values it is not possible to use this feature to identify the operating conditions of the turbine. For other components, the distributions appear sufficiently distinct; this suggests that in these ranges of values it is possible to use this feature to identify the operating conditions of the wind farm.
To identify the components that best discriminate the four classification labels, a feature selection procedure was performed to identify which variables are most relevant to the model’s performance. The purpose of selecting features is to find features that allow researchers to better discriminate between different objects, and which are insensitive to translation, rotation, or scale problems. In other words, a reduction in predictors is applied, which reduces the complexity of the information to be processed and makes the system more efficient.
The Boruta algorithm iteratively compares the contribution of each feature to the final classification, with those of features created by randomly mixing the original ones (shadow features). Attributes that prove to be worse than shadow ones are progressively eliminated. Attributes that are significantly better than shadows are confirmed. Shadow features are recreated in each iteration. The algorithm stops when only confirmed attributes remain or when the maximum number of feature selection operations is reached. The purpose of this procedure is to select a subset of predictors from the starting one, which, while reducing the size (columns) of the dataset that retains the initial information, identified the most significant predictors compared to the others [35,36,37].
In Table 2, for each feature contained in the data frame, the mean, median, maximal, and minimal importance of the one-third frequency components are shown.
To facilitate the identification of the components that contribute the most to the discrimination between the operating conditions, in Table 2 the results have been sorted in descending order starting from the top (mean value). In this sorting, the column that shows the average value of importance, determined by the selection of characteristics, was used. It can be seen that for the first three components in the frequency range from 63 Hz to 400 Hz, the average of the importance of the variable is greater than 48. This indicates that these characteristics are the ones that provide the greatest contribution to the discrimination between operating conditions.
To obtain an overview and to better appreciate the results, a diagram is presented in Figure 7. Analyzing Figure 7 and Table 2, we can easily identify the variables that recorded the highest values of importance. If the first 10 returned variables (average > 46) are evaluated, it can be seen that they are within the 63-1600 Hz frequency range, with the addition of the 5 kHz frequency. To further reduce the calculation cost and avoid overfitting, we decided to further reduce the number of features selected to take only the first 10 among those that recorded the highest values on average.
The variables selected by Boruta’s method represent the input variables for a support vector machine-based model implemented to detect the operating conditions of the wind turbines. This model creates an SVM classifier to classify the wind turbine operating conditions.
The following SVM architecture was implemented:
  • Input data: The first 10 one-third frequency components returned by the feature selection method;
  • Radial basis function (RBF) kernel;
  • Multiclass classification (four classes).
In the training phase, the algorithms based on SVMs use labeled data; therefore a correspondence between the input values and the relative output values associated with them can be established. In our case, each occurrence related to a combination of one-third frequency components was associated with a value of the relative class. The training aims to identify a set of weights w, one for each characteristic, of which the linear combination predicts the value of the class. This discourse of weight regulation is common with the training of algorithms based on artificial neural networks. Compared to artificial neural networks in SVMs, the optimization of margin maximization is adopted in order to reduce the number of non-zero weights. The results obtained in this way correspond to the characteristics that count in the separation of classes through the identification of a hyperplane. These non-zero weights correspond to the support vectors because they identify the separation hyperplane [38,39].
Before proceeding with the model training, it is necessary to split the data. The sampling methodology used for the subdivision of data can have a significant effect on the quality of the subsets used for training and testing. Poor data splitting can result in inaccurate and highly variable model performance [40]. To evaluate the generalization capacity of the model, the data were divided into two randomly selected groups:
  • A training set equal to 70% of the available observations (45,402 samples). These records are used in the training phase to adjust the connection weights through an iterative procedure aimed at minimizing the classification error.
  • A test set equal to the remaining 30% of the available observations (19,458 samples). These records are used in the testing phase to evaluate the performance of the numerical simulation model. These are data that the model has never seen before, so they represent a test for the model.
In Table 3, a confusion matrix of the model is shown. The confusion matrix represents a valid tool for assessing the quality of the output returned by a classifier on the data set submitted as the input. The diagonal elements represent the number of points for which the expected label is equal to the true label, whereas the off-diagonal elements are those erroneously labeled by the classifier. The higher the diagonal values of the confusion matrix, the better the performance of the model, thus indicating many correct predictions.
The confusion matrix returned excellent results. The accuracy of the model was high (accuracy = 0.918), confirming that the model can identify the operating conditions of a wind turbine farm. From the analysis of Table 3, it can be verified that the model was able to correctly identify 17,866 observations out of 19,458. To complete the analysis of the results it is appropriate to evaluate how the forecast errors committed by the model were distributed. It can be seen that most errors involved the extreme classes, 0 (all towers off) and 1 (all towers on). Specifically, the model ranked 787 occurrences as 0 when the data were labeled as 1 (probably because wind noise in particular conditions becomes comparable with that produced by wind turbines). Furthermore, it classified 356 occurrences as 1 when instead the data were labeled as 2. This error may be due to the fact that the operating conditions 1 and 2 were very close, since in both these conditions only one turbine was missing. Of the errors mentioned, the first one has a greater weight, as a forecast of non-operation of the turbines when they are all in operation represents a situation that can lead to greater problems. In order to reduce these classification errors, it will be necessary to insert into the model the information related to the wind speeds detected during the monitoring phase.
To compare the classification results obtained using an SVM-based algorithm, the same labeled data were used to train a model based on artificial neural networks (ANNs). ANNs are a family of machine learning techniques composed of layers of interconnected neurons, with an architecture like that of biological neurons [41]. The functioning of an ANN mirrors that of the brain, based on several interconnected neurons working in parallel, which process an input, producing an output that will be sent to subsequent neurons. A very interesting aspect of the functioning of the brain is its ability to continuously vary the weights of the interconnections, making it possible to classify and generalize the stimuli received from the outside. In this way, it promotes adaptation by learning from past signals and consequently trains the model to facilitate learning by allowing new data to be processed more quickly and correctly [42,43].
A neural network is therefore a set of nodes arranged in layers, connected to each other by weights. The first layer is called the input layer and the last is the output layer, whereas the intermediate ones are defined as hidden layers, and are not accessible from the outside as all the characteristics of the complete network are stored in the matrices that define the weights [44,45]. The type of network determines the type of connections that are present between the nodes of different layers and between those of the same layer. The choice of the activation function determines the substantial difference with the corresponding biological neuron. In the latter, the sum of the incoming impulses is transmitted directly to the axons if the threshold is exceeded, essentially behaving like a linear regression model, approximating the distribution of data with a straight line. On the contrary, the use of a non-linear function allows for a better representation of the signals, without considering the fact that sometimes a linear regression is not usable [46,47].
The classifier was developed based on a feed-forward artificial multilayer neural network, with a single hidden layer and a multiclass output that returned the four operating conditions. Feedforward networks are those with the simplest architecture [48,49], being composed of an input layer, one or more hidden layers, and an output layer. Each neuron has its input parameters from the previous layer and no cross-connections are possible between nodes of the same layer or cycles in which the output is sent to previous layers. The flow of information then proceeds in only one direction and the output of each cycle is determined only by the current input. The softmax activation function is used with multinomial cross-entropy as the cost function [50,51,52,53].
Table 4 shows the classification results returned by the ANN-based classifier. It can be verified that the model was able to correctly identify 17,844 observations out of 19,458. The accuracy of the artificial neural network-based model was 0.917. We can see that most errors involved extreme the classes, 0 (all towers off) and 1 (all towers on). Specifically, the model ranked 694 occurrences as 0 when the data were labeled as 1, probably for the same reason as that in the SVM model, since wind noise in certain conditions is similar to that produced by wind turbines. It also classified 291 occurrences as 1 when the data were instead labeled as 2. This error may be because operating conditions 1 and 2 were very similar, since in both conditions only one turbine was not working.
Comparing the classification results obtained using an SVM-based algorithm and an ANN-based algorithm (Table 5), it appears that the two models provide comparable results, with a slight advantage observed in the SVM-based model.
The effect of the wind on the noise detected in the site of interest is strongly influenced by both the wind speed and its direction. The roughness of the ground exerts a great influence on the wind. The average wind speed decreases because of the ground, but at the same time the wind becomes turbulent. Atmospheric wind is a multi-correlated stochastic process in space and time. The average wind speed, due to the friction exerted by the ground, increases with height, according to an exponential or logarithmic profile. The type of soil therefore affects this profile in such a complex way that it is difficult to quantify the effect exactly. When the wind arrives in the vicinity of the building, noise is generated due to various physical phenomena (airborne noises, mechanical noises, structure-borne noises, and the Eigenmode excitation of elements). These contributions are added to those derived from the noise generated by wind turbines contributing to the background noise. The contribution derived from the wind can in some cases become comparable with that produced by wind turbines when the wind speed increases significantly; however, it is a broadband contribution. The wind direction also influences the contribution to the background noise, both in terms of noise produced by the turbines—which is propagated more effectively when the wind direction is in favor of the site of interest—and in the contribution generated by the impacts of the wind on the elements present in the vicinity of the building. To evaluate these contributions, it is necessary to carry out measurements of the speed and direction of the wind both in the vicinity of the turbine and in the vicinity of the site of interest.

4. Conclusions

The characterization of acoustic sources that contribute to ambient noise has long been a topic of study by researchers around the world. The difficulty in isolating a specific source is due to the considerable complexity of a sound made up of many contributions characterized by spectra of different frequencies. Each sound has a specific frequency spectrum, but when the spectra of different sources overlap it becomes difficult to distinguish between the different contributions. The availability of a tool that is capable of classifying the operating conditions of a wind farm could provide valid support in evaluating the contributions derived from various sources.
Nowadays, machine learning algorithms are widely used in various fields. In many cases, these algorithms have been used to address various problems related to wind turbines. In this study, measurements were mode of the noise emitted by wind turbines. The average spectral levels in the one-third octave bands were visually analyzed in order to identify the frequency range that was capable of discriminating between the operating conditions. A model was then adapted to evaluate the importance of variables in classifying wind turbine functioning classes. To reduce the number of predictor variables, a feature selection analysis was carried out based on the Boruta algorithm. In this way, only the most significant predictors were selected, which were capable of returning a response as close as possible to the expected one. Finally, machine learning-based models were developed, which analyzed the operating conditions of the wind turbine based on the sound pressure level values at specific frequencies. The high accuracy of the SVM model (accuracy = 0.918) suggests that the adoption of this tool could support the acoustic characterization of noise in environments close to wind turbines. Furthermore, the performance shown by the classifier led us to think that this technology could also be used in other fields. In addition, the classification errors committed by the model can be significantly reduced by adding the wind speeds detected during the monitoring phase to the system inputs, so the current model can be improved. This activity will be performed in future studies.

Author Contributions

Conceptualization, G.C., G.I., V.P.-R. and A.T.; data curation, G.C., G.I., V.P.-R. and A.T.; formal analysis, G.C., G.I., V.P.-R. and A.T.; investigation, G.C., G.I., V.P.-R. and A.T.; methodology, G.C., G.I., V.P.-R. and A.T.; resources G.C., G.I., V.P.-R. and A.T.; software, G.C.; supervision, G.C., G.I., V.P.-R. and A.T.; validation, G.C., G.I., V.P.-R. and A.T.; visualization, G.C., G.I., V.P.-R. and A.T.; writing—original draft, G.C., G.I., V.P.-R. and A.T.; writing—review and editing, G.C., G.I.,V.P.-R. and A.T. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Aleixandre-Tudó, J.L.; Castelló-Cogollos, L.; Aleixandre, J.L.; Aleixandre-Benavent, R. Renewable energies: Worldwide trends in research, funding and international collaboration. Renew. Energ. 2019, 139, 268–278. [Google Scholar] [CrossRef]
  2. Manzano-Agugliaro, F.; Alcayde, A.; Montoya, F.G.; Zapata-Sierra, A.; Gil, C. Scientific production of renewable energies worldwide: An overview. Renew. Sustain. Energy Rev. 2013, 18, 134–143. [Google Scholar] [CrossRef]
  3. Pacesila, M.; Burcea, S.G.; Colesca, S.E. Analysis of renewable energies in European Union. Renew. Sustain. Energy Rev. 2016, 56, 156–170. [Google Scholar] [CrossRef]
  4. Asdrubali, F.; Baldinelli, G.; D’Alessandro, F.; Scrucca, F. Life cycle assessment of electricity production from renewable energies: Review and results harmonization. Renew. Sustain. Energy Rev. 2015, 42, 1113–1122. [Google Scholar] [CrossRef]
  5. Mohtasham, J. Energy Procedia. Renew. Energ. 2015, 74, 1289–1297. [Google Scholar]
  6. Walker, N. Generating Wind Power; Crabtree Publishing Company: St Catharines, ON, USA, 2007. [Google Scholar]
  7. Sathyajith, M. Wind Energy: Fundamentals, Resource Analysis and Economics; Springer: Berlin/Heidelberg, Germany, 2006; pp. 1–9. [Google Scholar]
  8. Blaabjerg, F.; Ma, K. Wind Energy Systems. Proc. IEEE 2017, 105, 2116–2131. [Google Scholar]
  9. Jamieson, P. Innovation in Wind Turbine Design; John Wiley & Sons: Hoboken, NJ, USA, 2018. [Google Scholar]
  10. Saidur, R.; Rahim, N.; Islam, M.; Solangi, K. Environmental impact of wind energy. Renew. Sustain. Energy Rev. 2011, 15, 2423–2430. [Google Scholar] [CrossRef]
  11. Wagner, S.; Bareiß, R.; Guidati, G. Wind Turbine Noise; Springer Science and Business Media LLC: Berlin/Heidelberg, Germany, 1996. [Google Scholar]
  12. Anderson, J. Fundamentals of Aerodynamics, 6th ed.; McGraw-Hill: Columbus, OH, USA, 2017. [Google Scholar]
  13. Zhao, K.; Alimohammadi, S.; Okolo, P.N.; Kennedy, J.; Bennett, G.J. Aerodynamic noise reduction using du-al-jet planar air curtains. J. Sound Vib. 2018, 432, 192–212. [Google Scholar] [CrossRef]
  14. Liu, W. A review on wind turbine noise mechanism and de-noising techniques. Renew. Energy 2017, 108, 311–320. [Google Scholar] [CrossRef]
  15. Pedersen, E.; Waye, K.P.; Röösli, M.; Lörtscher, M.; Egger, M.; Pfluger, D.; Schreier, N.; Lörtscher, E.; Locher, P.; Spoerri, A.; et al. Wind turbine noise, annoyance and self-reported health and well-being in different living environments. Occup. Environ. Med. 2007, 64, 480–486. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Biggio, B.; Roli, F. Wild patterns: Ten years after the rise of adversarial machine learning. Pattern Recognit. 2018, 84, 317–331. [Google Scholar] [CrossRef] [Green Version]
  17. Nanda, S.; Zafari, F.; DeCusatis, C.; Wedaa, E.; Yang, B. Predicting network attack patterns in SDN using machine learning approach. In Proceedings of the 2016 IEEE Conference on Network Function Virtualization and Software Defined Networks (NFV-SDN), Palo Alto, CA, USA, 7–10 November 2016; pp. 167–172. [Google Scholar]
  18. Chou, J.-S.; Tran, D.-S. Forecasting energy consumption time series using machine learning techniques based on usage patterns of residential householders. Energy 2018, 165, 709–726. [Google Scholar] [CrossRef]
  19. Domun, Y.; Pedersen, L.J.; White, D.; Adeyemi, O.; Norton, T. Learning patterns from time-series data to discriminate predictions of tail-biting, fouling and diarrhoea in pigs. Comput. Electron. Agric. 2019, 163, 104878. [Google Scholar] [CrossRef]
  20. Voulodimos, A.; Doulamis, N.; Doulamis, A.; Protopapadakis, E. Deep Learning for Computer Vision: A Brief Review. Comput. Intell. Neurosci. 2018, 2018, 7068349. [Google Scholar] [CrossRef] [PubMed]
  21. Weinstein, B.G. A computer vision for animal ecology. J. Anim. Ecol. 2017, 87, 533–545. [Google Scholar] [CrossRef]
  22. Iannace, G.; Ciaburro, G.; Trematerra, A. Heating, ventilation, and air conditioning (HVAC) noise detection in open-plan offices using recursive partitioning. Buildings 2018, 8, 169. [Google Scholar] [CrossRef] [Green Version]
  23. Ciaburro, G.; Iannace, G. Improving Smart Cities Safety Using Sound Events Detection Based on Deep Neural Network Algorithms. Informatics 2020, 7, 23. [Google Scholar] [CrossRef]
  24. Graves, A. Supervised Sequence Labelling. Econometrics for Financial Applications; Springer Science and Business Media LLC: Berlin/Heidelberg, Germany, 2012; pp. 5–13. [Google Scholar]
  25. Fisher, D.H.; Pazzani, M.J.; Langley, P. Concept Formation: Knowledge and Experience in Unsupervised Learning; Fisher, D.H., Pazzani, M.J., Langley, P., Eds.; Morgan Kaufmann: San Francisco, CA, USA, 2014. [Google Scholar]
  26. Ciaburro, G. Sound Event Detection in Underground Parking Garage Using Convolutional Neural Network. Big Data Cogn. Comput. 2020, 4, 20. [Google Scholar] [CrossRef]
  27. Hansen, L.H.; Madsen, P.H.; Blaabjerg, F.; Christensen, H.; Lindhard, U.; Eskildsen, K. Generators and power electronics technology for wind turbines. In Proceedings of the IECON’01. 27th Annual Conference of the IEEE Industrial Electronics Society (Cat. No.37243), Denver, CO, USA, 29 November–2 December 2001; Volume 3, pp. 2000–2005. [Google Scholar]
  28. Doolan, C.J.; Moreau, D.J.; Brooks, L.A. Wind turbine noise mechanisms and some concepts for its control. Acoust. Aust. 2012, 40, 7–13. [Google Scholar]
  29. Doolan, C.J.; Moreau, D.J. An on-demand simultaneous annoyance and indoor noise recording technique. Acoust. Aust. 2013, 41, 141–145. [Google Scholar]
  30. Ljunggren, S. Expert group study on ‘Recommended practices for wind turbine testing and evaluation’, part 10. In Measurement of Noise Immersion from Wind Turbines at Noise Receptor Locations; Department of Building Science, The Royal Institute of Technology: Stockholm, Sweden, 1997. [Google Scholar]
  31. International Electrotechnical Commission. IEC 61400-11 Wind Turbine Generator Systems—Part 11: Acoustic Noise Measurement Techniques; International Electrotechnical Commission: Geneva, Switzerland, 2012. [Google Scholar]
  32. Awad, M.; Khanna, R. Support vector machines for classification. In Efficient Learning Machines; Apress: Berkeley, CA, USA, 2015; pp. 39–66. [Google Scholar]
  33. Lauer, F.; Bloch, G. Incorporating prior knowledge in support vector machines for classification: A review. Neurocomputing 2008, 71, 1578–1594. [Google Scholar] [CrossRef] [Green Version]
  34. Brereton, R.G.; Lloyd, G. Support Vector Machines for classification and regression. Analyst 2010, 135, 230–267. [Google Scholar] [CrossRef]
  35. Kursa, M.; Rudnicki, W. Feature Selection with theBorutaPackage. J. Stat. Softw. 2010, 36, 1–13. [Google Scholar] [CrossRef] [Green Version]
  36. Liaw, A.; Wiener, M. Classification and regression by randomForest. R News 2002, 2, 18–22. [Google Scholar]
  37. Breiman, L. Manual on Setting Up, Using, and Understanding Random Forests V3.1; Statistics Department University of California: Berkeley, CA, USA, 2002. [Google Scholar]
  38. Xu, J.; Han, J.; Nie, F.; Li, X. Multi-View Scaling Support Vector Machines for Classification and Feature Selection. IEEE Trans. Knowl. Data Eng. 2019, 32, 1419–1430. [Google Scholar] [CrossRef]
  39. Morais, C.; Lima, K.M. Comparing unfolded and two-dimensional discriminant analysis and support vector machines for classification of EEM data. Chemom. Intell. Lab. Syst. 2017, 170, 1–12. [Google Scholar] [CrossRef]
  40. Choubin, B.; Moradi, E.; Golshan, M.; Adamowski, J.; Hosseini, F.S.; Mosavi, A. An ensemble prediction of flood susceptibility using multivariate discriminant analysis, classification and regression trees, and support vector machines. Sci. Total. Environ. 2019, 651, 2087–2096. [Google Scholar] [CrossRef]
  41. Mashood, M.; Mukhtar, M.; Akyürek, E.; Kalganova, T.; Lesne, N. Neural Network based Control Method Implemented on Ambidextrous Robot Hand. Int. J. Autom. Smart Technol. 2017, 7, 27–32. [Google Scholar] [CrossRef]
  42. Ciaburro, G.; Iannace, G.; Passaro, J.; Bifulco, A.; Marano, D.; Guida, M.; Marulo, F.; Branda, F. Artificial neural network-based models for predicting the sound absorption coefficient of electrospun poly (vinyl pyrrolidone)/silica composite. Appl. Acoust. 2020, 169, 107472. [Google Scholar] [CrossRef]
  43. Saha, A.; Majumder, M. An Efficient Technique for Double Faults Detection and their Locations Identification in Digital Microfluidic Biochip. Int. J. Autom. Smart Technol. 2019, 9, 65–75. [Google Scholar]
  44. Ciaburro, G.; Iannace, G.; Ali, M.; Alabdulkarem, A.; Nuhait, A. An artificial neural network approach to modelling absorbent asphalts acoustic properties. J. King Saud Univ. Eng. Sci. 2021, 33, 213–220. [Google Scholar] [CrossRef]
  45. Abiodun, O.I.; Jantan, A.; Omolara, A.E.; Dada, K.V.; Mohamed, N.A.; Arshad, H. State-of-the-art in artificial neural network applications: A survey. Heliyon 2018, 4, e00938. [Google Scholar] [CrossRef] [Green Version]
  46. Iannace, G.; Ciaburro, G.; Trematerra, A. Modelling sound absorption properties of broom fibers using artificial neural networks. Appl. Acoust. 2020, 163, 107239. [Google Scholar] [CrossRef]
  47. Rao, G.S.; Kumari, G.V.; Rao, B.P. Image Compression Using Neural Network for Biomedical Applications. Adv. Intell. Syst. Comput. 2019, 11, 107–119. [Google Scholar] [CrossRef]
  48. Iannace, G.; Ciaburro, G.; Trematerra, A. Wind Turbine Noise Prediction Using Random Forest Regression. Machines 2019, 7, 69. [Google Scholar] [CrossRef] [Green Version]
  49. Aizenberg, I.; Moraga, C. Multilayer Feedforward Neural Network Based on Multi-valued Neurons (MLMVN) and a Backpropagation Learning Algorithm. Soft Comput. 2006, 11, 169–183. [Google Scholar] [CrossRef]
  50. Kramer, A.; Morgado-Dias, F. Applications of Artificial Neural Networks in Process Control Applications: A Review. In Proceedings of the 2018 International Conference on Biomedical Engineering and Applications (ICBEA), Amsterdam, The Netherland, 16–18 May 2018; pp. 1–6. [Google Scholar]
  51. Iannace, G.; Trematerra, A.; Ciaburro, G. Case study: Automated recognition of wind farm sound using artificial neural networks. Noise Control. Eng. J. 2020, 68, 157–167. [Google Scholar] [CrossRef]
  52. Romero, V.P.; Maffei, L.; Brambilla, G.; Ciaburro, G. Modelling the soundscape quality of urban waterfronts by artificial neural networks. Appl. Acoust. 2016, 111, 121–128. [Google Scholar] [CrossRef]
  53. Romero, V.P.; Maffei, L.; Brambilla, G.; Ciaburro, G. Acoustic, Visual and Spatial Indicators for the Description of the Soundscape of Waterfront Areas with and without Road Traffic Flow. Int. J. Environ. Res. Public Health 2016, 13, 934. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Map of the area where the wind farm is located, with indications of the position of the wind turbines and the measurement station. For each wind turbine, the distance that separates it from the measurement station has been added.
Figure 1. Map of the area where the wind farm is located, with indications of the position of the wind turbines and the measurement station. For each wind turbine, the distance that separates it from the measurement station has been added.
Applsci 11 06488 g001
Figure 2. Support vector machine (SVM)-based binary classification.
Figure 2. Support vector machine (SVM)-based binary classification.
Applsci 11 06488 g002
Figure 3. Time history of the sound pressure level measured in situ, in different operating conditions.
Figure 3. Time history of the sound pressure level measured in situ, in different operating conditions.
Applsci 11 06488 g003
Figure 4. Average spectral levels in one-third octave bands between 50 Hz and 5 kHz for the entire measurement session.
Figure 4. Average spectral levels in one-third octave bands between 50 Hz and 5 kHz for the entire measurement session.
Applsci 11 06488 g004
Figure 5. Average spectral levels in one-third octave bands between 50 Hz and 5 kHz for the four functioning conditions. (a) all turbines on; (b) one turbine off; (c) two turbines off; (d) all turbines off.
Figure 5. Average spectral levels in one-third octave bands between 50 Hz and 5 kHz for the four functioning conditions. (a) all turbines on; (b) one turbine off; (c) two turbines off; (d) all turbines off.
Applsci 11 06488 g005
Figure 6. Density lattice plots of average spectral levels in one-third octave band. 0 = all turbines off, 1 = all turbines on, 2 = one turbine off, 3 = two turbines off.
Figure 6. Density lattice plots of average spectral levels in one-third octave band. 0 = all turbines off, 1 = all turbines on, 2 = one turbine off, 3 = two turbines off.
Applsci 11 06488 g006
Figure 7. Attribute importance for all the features (one-third frequency components) determined by the Boruta algorithm. For each frequency, the values of the statistics shown in Table 2 are reported: mean, median, min, and max. In this way, it is possible to obtain a graphical comparison of the importance that each feature assumes in the classification of wind farm operation.
Figure 7. Attribute importance for all the features (one-third frequency components) determined by the Boruta algorithm. For each frequency, the values of the statistics shown in Table 2 are reported: mean, median, min, and max. In this way, it is possible to obtain a graphical comparison of the importance that each feature assumes in the classification of wind farm operation.
Applsci 11 06488 g007
Table 1. Summary of the acoustic measurements carried out in time intervals representative of different operating conditions. The measurements refer to several days of monitoring (Figure 3). 0 = all turbines off, 1 = all turbines on, 2 = one turbine off, 3 = two turbines off.
Table 1. Summary of the acoustic measurements carried out in time intervals representative of different operating conditions. The measurements refer to several days of monitoring (Figure 3). 0 = all turbines off, 1 = all turbines on, 2 = one turbine off, 3 = two turbines off.
Operating ConditionStart of Monitoring
(Hours)
Monitoring Periods (Hours:Minutes:Seconds)LeqA
(dBA)
Average Wind Speed
(m/s)
122:00:0002:00:0038,49.0
200:00:0000:20:0036,89.1
300:20:0000:20:0031.29.2
000:40:0000:20:0024.99.0
123:00:0001:00:0040.213.0
200:00:0000:20:0038.414.0
300:20:0000:20:0038.515.0
000:40:0000:20:0040.717.0
101:00:0002:40:0050.314.0
002:40:0001:50:0050.823.0
120:30:0002:30:0043.712.0
200:00:0000:30:0042.010.8
000:30:0000:30:0036.710.5
101:12:0004:01:0045.810.7
Table 2. Attribute statistics of the feature selection analysis related to the importance of the one-third frequency components, sorted by mean value (Figure 7).
Table 2. Attribute statistics of the feature selection analysis related to the importance of the one-third frequency components, sorted by mean value (Figure 7).
FrequencyMeanMedianMinMax
6379.7598678.7970475.0227486.74699
8052.2572552.6141848.5244355.3532
40048.5736149.2501345.0459949.92336
125048.552948.4937945.8825650.25572
500048.4359448.0252646.8555152.15254
63048.1746148.2775845.3175151.56099
12546.9164846.9890344.8929149.44468
160046.6934546.4456645.2381349.22559
5046.660446.3056744.255450.58229
25046.557446.6939642.4956749.51197
20042.8119642.2406940.7864846.74624
400040.6455240.1810938.2521442.71576
315039.2086439.2843636.4212941.18301
200037.0348937.0287835.6393838.51441
50036.580236.4627635.3455738.1025
31536.5084836.3980134.2132938.94105
250036.2673336.3472434.2624537.75657
80036.1896436.5087634.2222438.27648
100033.5629133.5376831.1933935.67666
10033.1427332.9038430.5255635.53637
16028.6553628.7280927.0635630.34784
Table 3. Confusion matrix for the support vector machine-based model.
Table 3. Confusion matrix for the support vector machine-based model.
Predictions
Target0123
03120483141
178713,3716119
21435680585
3119247570
Table 4. Confusion matrix for the artificial neural network-based model.
Table 4. Confusion matrix for the artificial neural network-based model.
Predictions
Target0123
030151483047
169413,450895
24929185367
31681214526
Table 5. Comparison between the accuracy returned by the SVM-based model and the ANN-based model.
Table 5. Comparison between the accuracy returned by the SVM-based model and the ANN-based model.
SVM-Based ModelANN-Based Model
Accuracy0.9180.917
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Ciaburro, G.; Iannace, G.; Puyana-Romero, V.; Trematerra, A. Machine Learning-Based Tools for Wind Turbine Acoustic Monitoring. Appl. Sci. 2021, 11, 6488. https://doi.org/10.3390/app11146488

AMA Style

Ciaburro G, Iannace G, Puyana-Romero V, Trematerra A. Machine Learning-Based Tools for Wind Turbine Acoustic Monitoring. Applied Sciences. 2021; 11(14):6488. https://doi.org/10.3390/app11146488

Chicago/Turabian Style

Ciaburro, Giuseppe, Gino Iannace, Virginia Puyana-Romero, and Amelia Trematerra. 2021. "Machine Learning-Based Tools for Wind Turbine Acoustic Monitoring" Applied Sciences 11, no. 14: 6488. https://doi.org/10.3390/app11146488

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop