Next Article in Journal
Robust Feedback Zoom Tracking for Digital Video Surveillance
Next Article in Special Issue
The Combined Application of Impinger System and Permeation Tube for the Generation of Volatile Organic Compound Standard Gas Mixtures at Varying Diluent Flow Rates
Previous Article in Journal
On the Use of Sensor Fusion to Reduce the Impact of Rotational and Additive Noise in Human Activity Recognition
Previous Article in Special Issue
Effect of Functional Group and Carbon Chain Length on the Odor Detection Threshold of Aliphatic Compounds
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Electronic Nose Based on Independent Component Analysis Combined with Partial Least Squares and Artificial Neural Networks for Wine Prediction

1
Sensory Systems Research Group, University of Extremadura, 06006 Badajoz, Spain
2
Industrial Applications of Artificial Intelligence Research Group, 06006 Badajoz, Spain
*
Author to whom correspondence should be addressed.
Sensors 2012, 12(6), 8055-8072; https://doi.org/10.3390/s120608055
Submission received: 10 April 2012 / Revised: 25 May 2012 / Accepted: 5 June 2012 / Published: 11 June 2012

Abstract

: The aim of this work is to propose an alternative way for wine classification and prediction based on an electronic nose (e-nose) combined with Independent Component Analysis (ICA) as a dimensionality reduction technique, Partial Least Squares (PLS) to predict sensorial descriptors and Artificial Neural Networks (ANNs) for classification purpose. A total of 26 wines from different regions, varieties and elaboration processes have been analyzed with an e-nose and tasted by a sensory panel. Successful results have been obtained in most cases for prediction and classification.

1. Introduction

The most important sensory characteristic in the study of the quality of wine is its aroma. Wine aroma has a great chemical complexity because it is produced by the simultaneous perception of many volatile compounds. This property contributes a 70% of weight in sensory panels with respect to texture and taste. Electronic nose (e-nose) technology is shown as a possibility for aroma profile analysis. Essentially the e-nose is an array of gas sensors with partially overlapped selectivities, a signal collecting unit and a pattern recognition software [1]. On the other hand, pattern analysis constitutes a really important building block in the development of gas sensor instruments given its ability of detecting, identifying and measuring volatile compounds. For these reasons this technology has been proposed as an artificial substitute of the human olfactory system. To successfully design a pattern analysis system for machine olfaction, a rigorous study is necessary on the diverse issues involved in processing multivariate data (see Figure 1): signal preprocessing and feature extraction, dimensionality reduction, classification, regression or clustering and validation [1,2]. A variety of Pattern Recognition and Machine Learning techniques is available for each module and only the preprocessing module is sensor dependent.

The first block in Figure 1 represents the e-nose hardware. After sensor signals have been acquired and saved into the computer, the first computational stage (i.e., signal preprocessing) starts and serves various purposes, including compensating for sensor drift, extracting descriptive parameters from sensor array response and preparing the feature vector for further processing. A dimensionality reduction stage projects this initial feature vector onto a lower dimensional space in order to avoid problems associated with high-dimensional, sparse datasets and redundancy. The resulting low-dimensional feature vector is further used to solve a given prediction problem, generally classification, regression or clustering.

The problem of identifying an unknown sample as one from a previously learned odorant set is addressed in classification tasks. In regression tasks, the aim is to predict a set of properties (e.g., concentration for an analyte in a complex mixture or the value of a sensory panel descriptor) from other variables. Finally, in clustering tasks the goal is to learn the structural relationship among different odorants. A final step, sometimes overlooked, is the selection of models and parameter setting and the estimation of true rates for a trained model by means of validation techniques.

2. Material and Methods

To perform the measurements of wine samples, a home-made and home-developed electronic nose was used. The e-nose and data processing system are described in this section.

2.1. Wine Samples

A total of 26 wines (19 red and 7 white) from different regions were tasted by a sensory panel and analyzed by the electronic nose for the experiment. A list of the wine samples is shown in Table 1. Wine samples were acquired in specialized shops and supplied by collaborator wine cellars.

2.2. Sensory Panel

A group of 30 people with previous experience in wine analysis were trained in recognizing 45 aromatic compounds from wine. In this training, all compounds were presented to the panelists at their threshold concentration in water and in test wines [3]. Wines were presented in random order to the panelists. No information was given to the assessors about the origin of the samples and no colored lighting assessed panelists. The wine tasting took place in an air-conditioned room (21 °C) with isolated booths. Judges assessed the aroma using a tasting evaluation sheet that included 10 sensory descriptors (herbaceous, fruity, flower-like, spicy, vegetal, phenolic, microbiologic, chemical, oxidation and woody-like) and 5 quality parameters (wine quality, aromatic intensity, alteration, persistence and off-flavours). The different terms were evaluated in a scale from 1 to 5 (1, null, very weak; 2, weak; 3, medium; 4, strong; 5, very strong). The average of all the panelists was calculated to build the prediction models. All the sensory evaluations were realized under Spanish Standardization Rules (UNE) [4].

2.3. Electronic Nose

An array of 16 thin film tin oxide sensors was prepared by RF sputtering onto alumina and doped with chromium and indium. Sensor thickness varied between 200 and 800 nm. Some sensors were doped with Chromium and Indium, either on the surface or in an intermediate layer. Array composition is described in Figure 2(a). The array was placed in a 24 cm3 stainless steel cell with a heater and a thermocouple. The operating temperature of sensors is controlled to 250 °C with a PID temperature controller. Two sampling methods were available for analysis: purge and trap and static headspace followed by a dynamic injection. In purge and trap, a Tekmar 3100 was used [5]. In headspace, 10 mL of solution were kept in a 50 mL Dreschel bottle at 30 °C for 30 minutes followed by 20 minutes carrying the volatile compounds to the sensor cell. The carrier gas was high purity nitrogen (99.9998%) at a constant flow of 200 mL/min [6]. In Figure 3, a block diagram of the measuring system is shown.

The resistance of the sensors was measured with a Keithley 2001 8 1/2 digits digital multimeter (DMM) with a Keithley 7001 40-channels multiplexer connected to the personal computer through a GPIB interface. More details of the system can be found in [7]. Responses of the individual sensors are defined relative to the minimum resistance to a 12% (V/V) solution of ethanol in deionised water for all measurements (calibration solution). The response of the sensors is low, so no special requirements are needed in instrumentation. Two measurements are made each minute, and it is enough to obtain the response curve profile for data analysis. Figure 2(b) shows the typical transient responses of four chemoresistive sensors of the array, operating at 250 °C, exposed towards the headspace of one of the wine samples (Pioz-W1). The response of the sensors corresponds to several pulses of 20 minutes of exposition to the tested wine flavour followed by a pure nitrogen purge for 40 minutes. These responses were stored in hard disk and processed later by the data processing system for prediction and classification purposes.

2.4. Data Processing System

This section shows the system designed for the analysis of e-nose responses of the wine samples under test. The aim of this system is to perform the classification of the wine samples and the prediction of the sensory panel attributes from the e-nose responses. The data collected were analyzed by means of pattern recognition techniques using commercial software packages.

Matlab® [8] was used for linear methods like Independent Component Analysis (ICA) [9,10] and non-linear methods based on Artificial Neural Networks (ANNs) like backpropagation and probabilistic neural networks. Partial Least Squares Regression was performed with Unscrambler® [11].

2.4.1. Preprocessing and Feature Extraction

The aim of feature extraction is to find a low-dimensional mapping f : x ∈ ℜNy ∈ ℜM(M < N) that preserves most of the information in the original feature vector x. In our case, a previous optimization of the method of feature extraction and selection were performed [12]; the responses of the individual sensors were defined relative to the minimum resistance to 12% (V/V) of ethanol for all the measurements:

r = R wine R calibration
where Rwine was the minimum resistance of the sensor in the measurement of wine and Rcalibration was the minimum resistance of the sensor in a solution of 12% of ethanol. This calibration was performed once a week, to eliminate the sensors drift [2]. Data were centered and scaled for further analysis.

2.4.2. Dimensionality Reduction and Classification

The traditional method for reduction in dimensionality and visualization of multivariate measurement data is Principal Component Analysis (PCA) [13]. This work introduces the higher order statistical method called Independent Component Analysis (ICA) [12] as an alternative method of data processing for an e-nose data. Moreover, ICA is an appropriate approach for signal processing as it is demonstrated in [14].

ICA is a statistical and computational technique for revealing hidden factors that underlie sets of random variables or measurements. This technique defines a generative model or the observed multivariate data, which is typically given as a large database of samples. In the model, the data variables are assumed to be linear or nonlinear mixtures of some unknown latent variables, and the mixing system is also unknown. The aim of ICA is the decomposition for multivariate signals in statistically independent component contributions with minimum loss of information [15]. ICA can be seen as an extension to principal component analysis. While PCA is an effective method for data compression, ICA is an effective method for the extraction of independent features [16,17]. It is much more powerful technique, even, capable of finding the underlying factors or sources when classic methods fail completely.

The ICA bilinear model can be written:

X = A S T + E
where X (i × j) is the original experimental raw data matrix and A (i × n) and ST (n × j) are the so-called mixing and source matrices, respectively, in ICA notation. In addition, E (i × j) is the error matrix. The i and j index are respectively the numbers of rows and columns of the data matrix X, and n the number of components included in the bilinear decomposition of Equation (1). ICA algorithms try to find “unmixing” matrix W according to:
S ^ T = WX
where SˆT is the estimation of the source matrix ST. Then, when W is the pseudoinverse of the mixing matrix A, the estimated source signal will be equal to the original source signal ST (provided that error tends to zero):
S ^ T = WX = WA S T = S T

The two main assumptions of ICA are:

  • The mixing vectors in A are linearly independent.

  • The components in ST are mutually statistically independent.

On the other hand regression problems constitute a more challenging domain for e-nose instruments [2]. The goal of regression is to establish a predictive model from a set of independent variables (e.g., gas sensor responses) to another set of continuous dependent variables.

Partial Least Squares regression (PLS) combines the properties of multiple linear regression and principal component analysis to produce a technique that is able to accept collinear data and separate the sample noise in order to make linear combinations in the dependent concentration matrix. PLS is the “gold standard” in chemometrics due to its ability to handle collinear data and reduce the number of required calibration samples [18].

As opposed to PCR, which extracts the “latent variables” from the directions of maximum variance in the sensor matrix Y (the eigenvectors of YTY), PLS finds the directions of maximum correlation between the sensor response matrix and the calibration mixtures matrix (Y and C) in a sequential fashion. The first PLS latent variable (t = Yω) is obtained by projecting Y along the eigenvector ω corresponding to the largest eigenvalue of YTCCTY [19]. To find the second and subsequent latent variables, the current PLS latent variable is deflated by its ordinary least squares prediction (a simple approach to regression is to assume that the dependent variables can be predicted from a linear combination of the sensor responses) and the eigen-analysis is repeated. A stopping point for the sequential expansion is determined through cross-validation. Published works on multicomponent analysis using gas sensor arrays and PLS may be found in references [2023].

2.5. Artificial Neural Networks

The discrimination of the tested wines belonging to the same class has been tackled with a pattern recognizer based on ANN providing nonlinearity in the multivariate classification performance. Two types of neural networks have been tested for the classification of the wine aroma.

Also, Leave-One-Out (LOO) cross validation was applied to check the performance of the network [24]. LOO consists in training N distinct nets (where N is the number of measurements) by using N1 training vectors, while the validation of the trained network is carried out by using the remaining vector, excluded from the training set. This procedure is repeated N times until all vectors are validated [25].

2.5.1. Backpropagation Networks

Backpropagation algorithm is the most suitable learning rule for multi-layer perceptrons. It involves two stages: firstly, a feedforward stage where the exterior input information on the input nodes is propagated forward in order to compute the output information indicators at the output unit, and secondly, a backward phase where alterations to the connection weights are adjusted based on the differences between the computed and the actual indications at the output units [26,27].

The information processing that it carry out is basically an approximation of a mapping or function:

f : D R n R m
by means of training over a set of example mapping, yk = f(yk), where k = 1, 2, 3…, N

The backpropagation network architecture used has a learning rate of 0.01 and training times less than 3 minutes. Also, it is formed by three layers: the input layer has 16 neurons corresponding to the 16 sensors, a variable number of neurons with tansig function for the hidden layer, and nine neurons with hardlim function for the output layer, the same number of existing classes.

2.5.2. Probabilistic Networks

PNN is a kind of feedforward neural network. The original PNN structure is a direct neural network implementation of Parzen nonparametric Probability Density Function (PDF) estimation and Bayes classification rule [28,29]. The standard training procedure of PNN requires a single pass over all the patterns of the training set [29].

If XRd is a d-dimensional pattern vectors and its associated class is i ∈ (S1, S2, S3, …, Sk), where k is the number of possible classes, and a posteriori probability Pr(Si|x), that is from class Si, is by Bayes' rule:

P r ( S i | x ) = P r ( x | S i ) P r ( S i ) P ( x )
where Pr(x|Si), with i = 1, 2, 3, …, k is a priori pdf of the pattern in classes to be separated. And Pr(Si) with i = 1, 2, 3, …, k are a priori probabilities of the classes. P(x) is assumed to be a constant.

The decision rule is to select class Si for with Pr(Si|x) is maximum. This will happen if for all ji

P ( x | S i ) P r ( S i ) > P ( x | S i ) P r ( S j )

A Probabilistic Neural Network (PNN) composed of three layers, with radial basis transfer functions in the hidden layer and a competitive one in the output [2], was used for classification purposes.

3. Results and Discussion

Wine samples were tasted by the sensory panel and analyzed by the electronic nose. Only headspace sampling method have been used in electronic nose because of the purge and trap did not offer good prediction ability due to the concentration and elimination of several chemical compounds in the trap [30]. Matlab® program was used for preprocessing and feature extraction. This program automatically processes the measurements files and generates the output vectors for dimensionality reduction. Besides, the program uses the neural networks toolbox and it performs the dimensionality reduction using PCA, LDA or ICA and classification with several types of Neural Networks (BackPropagation and Probabilistic Neural Networks). Finally it generates the corresponding plots and the confusion matrix obtained with the validation of ANNs. A summary of the results obtained in these analyses is shown in the following paragraphs.

3.1. Sensory Panel Responses

Taste of the samples was performed by the panel. The data obtained were processed and the average of all the panelists was calculated for each wine and descriptor. The sensory attributes obtained for young white, young red and oak aged wines are shown in Table 2. In young white wines the flower and fruity aromas predominate, whereas in young red wines the fruity and spicy aromas predominate, and the fruity, spicy and woody-like aromas predominate in oak aged wines.

3.2. E-Nose Response

As mentioned in Section 2.3, Figure 2 represents the typical transient responses of four chemoresistive sensors, operating at 250 °C, exposed towards the headspace of the blank wine. The response of the sensors corresponds to several pulses of 20 minutes of exposition to the sample of wine, followed by a pure nitrogen purge for 40 minutes. When electrovalves are switched on and wine aroma is carried to the sensors cell, the resistance of the sensors decrease and when electrovalves are switched off, the response increase to the equilibrium values.

Figures 46 show the radial plot of the 16 sensors responses to the three sets of samples measured: white, red and aged wines respectively. It can be noticed that the sensors give different signal for each different sample, illustrating the discrimination capabilities of the array.

No relation among sensors response and aromatic profile of each wine can be established according to Figures 46 and a classification of samples cannot be performed. The variation in response intensity is due to different headspace composition of the wines. Each sample has its characteristic organic volatile compounds profile.

3.3. Dimensionality Reduction and Classification

For a better visualization of the data and to eliminate redundancy, dimensionality reduction techniques are generally used. In this case ICA was carried out using signals corresponding to five repeated exposures collected in different days. The plot of the two first Independent Components for the measurements of young white, young red and aged red wines are shown in Figures 7, 8 and 9 respectively. As shown in Figure 7, the clusters corresponding to each wine are well separated except some partial overlapping between samples of W7 and W5. The young white wines are clearly separated among them. Figure 8 shows the ICA plot for young red wines, where a partial overlapping among some samples of R1, R2 and R9 can be observed. The clusters of the other wines are well separated. Figure 9 shows the ICA plot for aged red wines. In this figure there are several overlapping zones between some wines due to the similarity of the wines. Some of these zones correspond to measurements of the same wine but aged in a different type of oak barrel, e.g., M5–M6 and M9–M10. On the contrary, other wine samples, such as M1–M2, M3–M4 and M7–M8 are well separated.

A regression model to estimate sensory panel indicators from electronic nose was built by partial least squares regression (PLS). Models were cross-validated by the leave-one-out method. Preprocessing, modelling and validation were performed with Unscrambler®. Parameters used to evaluate the models prediction ability were: root mean square prediction error and correlation coefficient between real and predicted Y variables. All variables were normalized prior to the analysis.

3.3.1. Sensory Data versus E-Nose Data

The correlation coefficients and mean squared error values calculated for all the parameters in this extraction technique are listed in Table 3.

These results show that good correlation coefficients and mean squared error values are obtained except for woody-like parameter. Nevertheless, results could be improved by analyzing more wine samples with sensory panel and e-nose in order to train the model and obtain more accuracy in the predictions. Models were built in order to predict the sensorial descriptors and they exhibit good prediction ability, with correlations coefficients from 0.50 to 0.95 and mean squared error values between 4.31 ×10−4 and 1.29 ×10−3. Figure 10 shows the relationship between estimated (y-axis) and real values of sensorial descriptors (x-axis) based on the PLS model calculated from e-nose responses using headspace method. Validation points are shown in figures obtained with cross-validation. As shown in Figure 10, there is a close relationship between sensory panel attributes from sensory evaluation and those estimated by the PLS model. Therefore, a good prediction of these variables could be performed from the e-nose measurements.

3.4. Artificial Neural Networks

Two types of neural networks were used for classification: Backpropagation and Probabilistic Neural Networks. In general the output of the neural network is given as a confusion matrix. From the confusion matrix a success rate is defined as the number of correct classified measures in each class over the total number of measures in this class.

3.4.1. Backpropagation Networks

In the training of feedforward networks, several numbers of neurons in the hidden layer have been tested. The optimal number turned out to be 14 neurons. The network was trained with the data obtained with the wine samples measurements. Leave-one-out (LOO) validation was performed in order to check the performance of the network. The classification success was 97% for young white wines, 87% for young red wines and 84% for aged red wines. As an example, the confusion matrix obtained for white wines is shown in Table 4.

3.4.2. Probabilistic Networks

Probabilistic neural network were trained with the same data than backpropagation networks. Leave-one-out (LOO) validation was also performed. The classification success was 94% for young white wines, 84% for young red wines and 82% for aged red wines.

4. Conclusions and Future Works

An electronic nose has been designed in this work for prediction and classification of wine samples. This e-nose consist of a gas sensor array combined with Partial Least Squared and a pattern recognition system based on Independent Component Analysis for dimensionality reduction. Results obtained through wine headspace technique have been used for predicting the sensory panel descriptors of the wines analyzed. In general, good results have been obtained for the correlation between sensory panel and e-nose due to the similarity between this sampling system and the natural process of taste in humans. This is a consequence to the global response of receptors (human nose and chemical sensors) to the whole aroma, allowing to predict wine quality parameters and sensory descriptors of sensory panel.

The signals recorded from the multisensor array exposed to the aroma of the wine samples have also been used for classifying these samples through ICA as explorative techniques and dimensionality reduction combined with a pattern recognizer based on ANN. Results show classification success ratios higher than 87% for all types of wine.

Finally, it is important to remark that, results could be improved by measuring more wine samples in order to increase the database and to train the networks and model for classification and prediction. Samples should be selected in order to present a wide range in the parameters to predict.

Acknowledgments

This work has been supported by the Regional Government of Extremadura, through the European Regional Development Funds (GR10097) and the Spanish Ministry of Science and Innovation, through the project NAMIRIS (TEC2010-21357-C05-02).

References

  1. Pearce, T.; Schiffman, S.; Nagle, H.; Gardner, J.W. Handbook of Machine Olfaction; Wiley-VCH: Berlin, Germany, 2006. [Google Scholar]
  2. Gutiérrez-Osuna, R. Pattern analysis for machine olfaction: A review. IEEE Sens. J. 2002, 2, 189–202. [Google Scholar]
  3. Kotseridis, Y.; Razungles, A.; Bertrand, A.; Baumes, R. Differentiation of the aromas of Merlot and Cabernet Sauvignon wines using sensory and instrumental analysis. J. Agric. Food Chem. 2000, 48, 5383–5388. [Google Scholar]
  4. Aenor. 2010. Available online: http://www.en.aenor.es/aenor/aenor/perfil/perfil.asp (accessed on 7 June 2012).
  5. Santos, J.; Arroyo, T.; Aleixandre, M.; Lozano, J.; Sayago, I.; Garcia, M.; Fernandez, M.J.; Ares, L.; Gutierrez, J.; Cabellos, J.M.; et al. A comparative study of sensor array and GC–MS: Application to Madrid wines characterization. Sens. Actuators B Chem. 2004, 102, 299–307. [Google Scholar]
  6. Santos, J.; Lozano, J.; Alexaindre, M.; Sayago, I.; Fernández, M.J.; Arés, L.; Guitiérrez, J.; Horrillo, M.C. Discrimination of different aromatic compounds in water, ethanol and wine with a thin film sensor array. Sens. Actuators B Chem. 2004, 103, 98–103. [Google Scholar]
  7. Lozano, J.; Santos, J.; Horrillo, M. Classification of white wine aromas with an electronic nose. Talanta 2005, 67, 610–616. [Google Scholar]
  8. Matlab®. 1984. Available online: http://www.mathworks.com/products/matlab/index.html (accessed on 7 June 2012).
  9. Kermit, M.; Tomic, O. Independent component analysis applied on gas sensor array measurement data. IEEE Sens. J. 2003, 3, 218–228. [Google Scholar]
  10. Bermejo, S. Independent component analysis for solid-state chemical sensor arrays. J. Appl. Intell. 2006, 24, 61–73. [Google Scholar]
  11. Unscrambler®. 1987. Available online: http://www.camo.com/rt/Products/Unscrambler/unscrambler.html (accessed on 7 June 2012).
  12. Jutten, C.; Herault, J. Blind separation of sources, part I: An adaptive algorithm based on neuromimetic architecture. Signal Process. 91, 24, 1–10. [Google Scholar]
  13. Duda, R.; Hart, P.; Stork, D. Pattern classification. J. Classif. 2007, 24, 115–117. [Google Scholar]
  14. Parastar, H.; Jalali-Heravi, M.; Tauler, R. Is independent component analysis appropriate for multivariate resolution in analytical chemistry? Trends Anal. Chem. 2012, 31, 134–143. [Google Scholar]
  15. Comon, P. Independent component analysis, a new concept? Signal Process 1994, 36, 287–314. [Google Scholar]
  16. Hyvarinen, A.; Oja, E. Independent component analysis: Algorithms and applications. Neural Netw. 2000, 13, 411–430. [Google Scholar]
  17. Lathauwer, L.D.; Moor, B.D. An introduction to independent component analysis. J. Chemom. 2000, 14, 123–149. [Google Scholar]
  18. Geladi, P.; Kowalski, B. Partial least-squares: A tutorial. Anal. Chim. Acta 1986, 185, 1–17. [Google Scholar]
  19. Lindgren, F.; Geladi, P.; Wold, S. The Kernel algorithm for PLS. J. Chemom. 1986, 7, 45–59. [Google Scholar]
  20. Hierlemann, A.; Weimar, U.; Kraus, G.; Schweizer-Berberich, M.; Göpel, W. Polymer-based sensor arrays and multicomponent analysis for the detection of hazardous organic vapors in the environment. Sens. Actuators B Chem. 1995, 26, 126–134. [Google Scholar]
  21. Sundgren, H.; Winquistn, F. Artificial neural networks and gas sensor arrays: Quantification of individual components in a gas mixture. Meas. Sci. Technol. 1991, 2, 464–469. [Google Scholar]
  22. Wang, X.; Carey, W.; Yee, S.S. Monolithic thin-film metal-oxide gas-sensor arrays with application to monitoring of organic vapors. Sens. Actuators B Chem. 1995, 28, 63–70. [Google Scholar]
  23. Carey, W.; Yee, S.S. Calibration of nonlinear solid-state sensor arrays using multivariate regression techniques. Sens. Actuators B Chem. 1992, 9, 113–122. [Google Scholar]
  24. Vandeginste, B.G.M.; Massart, D.L.; Buydens, L.M.C.; de Jong, S.; Lewi, P.J.; Smeyers-Verbeke, J. Handbook of Chemometrics & Qualimetrics: Part B; Elsevier: New York, NY, USA, 1998; pp. 238–239. [Google Scholar]
  25. Bishop, C.M. Neural Networks for Pattern Recognition; Oxford University Press: Oxford, UK, 1999. [Google Scholar]
  26. Masters, T. Practical Neural Network Recipes in C++; Academic Press: SanDiego, CA, USA, 1993. [Google Scholar]
  27. Eberhart, R.; Dobbins, R. Neural Network PC Tools: A Practical Guide; Academic Press: Waltham, MA, USA, 1990. [Google Scholar]
  28. Specht, D.F. Probabilistic neural network. Neural Netw. 1990, 3, 109–118. [Google Scholar]
  29. Parzen, E. On the estimation of a probability density function and mode. Ann. Math. Stat. 1962, 3, 1065–1076. [Google Scholar]
  30. Lozano, J.; Santos, J.P.; Arroyo, T.; Aznar, M.; Cabellos, J.M.; Gil, M.; Horrillos, M.C. Correlating e-nose responses to wine sensorial descriptors and gas chromatography-mass spectrometry profiles using partial least squares regression analysis. Sens. Actuators B Chem. 2007, 127, 267–276. [Google Scholar]
Figure 1. Stages of typical signal processing in an electronic nose.
Figure 1. Stages of typical signal processing in an electronic nose.
Sensors 12 08055f1 1024
Figure 2. Typical transient response and doping characteristics of the gas sensors. (a) Sensor array composition; (b) Typical transient response of four sensors of the array.
Figure 2. Typical transient response and doping characteristics of the gas sensors. (a) Sensor array composition; (b) Typical transient response of four sensors of the array.
Sensors 12 08055f2 1024
Figure 3. Measurement setup: 1. Nitrogen bottle; 2. Mass Flowmeter Controller; 3. Dreschel bottle with sample; 4. Thermostatic bath; 5. Tekmar 3100 purge and trap; 6. Sensors cell; 7. Digital multimeter; 8. Multiplexer; 9. Computer.
Figure 3. Measurement setup: 1. Nitrogen bottle; 2. Mass Flowmeter Controller; 3. Dreschel bottle with sample; 4. Thermostatic bath; 5. Tekmar 3100 purge and trap; 6. Sensors cell; 7. Digital multimeter; 8. Multiplexer; 9. Computer.
Sensors 12 08055f3 1024
Figure 4. Polar plots of the variation in sensor resistance for the white wines.
Figure 4. Polar plots of the variation in sensor resistance for the white wines.
Sensors 12 08055f4 1024
Figure 5. Polar plots of the variation in sensor resistance for the red wines.
Figure 5. Polar plots of the variation in sensor resistance for the red wines.
Sensors 12 08055f5 1024
Figure 6. Polar pSlots of the variation in sensor resistance for the aged red wines.
Figure 6. Polar pSlots of the variation in sensor resistance for the aged red wines.
Sensors 12 08055f6 1024
Figure 7. ICA plot of the young white wine measurements.
Figure 7. ICA plot of the young white wine measurements.
Sensors 12 08055f7 1024
Figure 8. ICA plot of the young red wine measurements.
Figure 8. ICA plot of the young red wine measurements.
Sensors 12 08055f8 1024
Figure 9. ICA plot of the aged red wine measurements.
Figure 9. ICA plot of the aged red wine measurements.
Sensors 12 08055f9 1024
Figure 10. Sensory panel attributes estimated by PLS model (y) vs. real values (x).
Figure 10. Sensory panel attributes estimated by PLS model (y) vs. real values (x).
Sensors 12 08055f10 1024
Table 1. Wine samples tasted by the sensory panel and measured by the e-nose.
Table 1. Wine samples tasted by the sensory panel and measured by the e-nose.
Num.WineType of wine
W1Pioz (Guadalajara, Spain)Young white
W2Viña Esmeralda (Miguel Torres, Spain)Young white
W3Castillo de San Diego (B. Barbadillo, Spain)Young white
W4Marqués de Riscal (Rioja, Spain)Young white
W5Viñas del Vero (Somontano, Spain)Young white
W6Don Álvaro de Luna Albillo 2001 (Madrid, Spain)Young white
W7Albariño 2003 (B. Martín Códax, Spain)Young white
R1Puerta de Alcalá 2003 (Vinos Jeromín, Madrid, Spain)Young red
R2Cillar de Silos (Bodegas de la Villa, Spain)Young red
R3Castillo de Olite (Bodegas Artesanas, Spain)Young red
R4Cabernet Sauvignon Encin 2003 (Madrid, Spain)Young red
R5Ruchel Mencía 2002 (B. Majlu, Spain)Young red
R6Monasterio de las Viñas (Cariñena, Spain)Young red
R7Castillo de Aguarón (Carineña, Spain)Young red
R8Don Luciano (B. García Carrión, Spain)Young red
R9Sliven Merlot 2002 (Bulgary)Young red
M1Pablo Morate American Oak Barrel (Madrid, Spain)Oak Barrel aged
M2Pablo Morate French Oak Barrel (Madrid, Spain)Oak Barrel aged
M3Carlos Gosalbez American Oak Barrel 2002 (Madrid, Spain)Oak Barrel aged
M4Carlos Gosalbez French Oak Barrel 2002 (Madrid, Spain)Oak Barrel aged
M5Andrés Morate American Oak Barrel (Madrid, Spain)Oak Barrel aged
M6Andrés Morate French Oak Barrel (Madrid, Spain)Oak Barrel aged
M7Luis Saavedra American Oak Barrel 2002 (Madrid, Spain)Oak Barrel aged
M8Luis Saavedra French Oak Barrel 2002 (Madrid, Spain)Oak Barrel aged
M9Carlos Gosalbez American Oak Barrel 2003 (Madrid, Spain)Oak Barrel aged
M10Carlos Gosalbez French Oak Barrel 2003 (Madrid, Spain)Oak Barrel aged
Table 2. Wine samples measured by the e-nose and tasted by the sensory panel.
Table 2. Wine samples measured by the e-nose and tasted by the sensory panel.
HerbaceousFlower-likeFruitySpicyVegetalPhenolicMicrobiologicChemicalOxidationWoody-likeAlterationPersistenceAverage intensityOff-flavoursWine quality
Pioz1.632.292.591.711.821.411.351.531.591.181.352.532.591.862.37
Viña Esmeralda1.693.473.591.411.411.181.291.241.121.061.123.533.761.23.71
Castillo de San Diego1.562.332.061.51.331.281.391.281.3311.52.332.591.482.44
Marqués de Riscal23.282.171.611.891.281.171.44111.443.063.851.593.44
Malvar Encín 20031.383.291.941.181.471.241.241.351.241.061.242.943.061.353.08
Viñas del Vero1.612.52.331.781.671.51.781.391.171.781.612.723.371.82.8
Don Álvaro de Luna1.412.111.672.171.391.561.441.51.563.111.52.833.521.573.13
Martín Codax1.53.5631.221.561.171.061.061.171.111.283.223.871.243.52
Puerta de Alcalá 20031.411.322.272.051.551.451.321.231.141.231.182.452.761.362.52
Cillar de Silos1.411.52.641.771.821.271.411.231.091.411.323.093.351.673.06
Garnacha Encín 20031.551.452.091.551.411.321.181.51.411.181.552.52.641.682.45
Castillo de Olite1.411.452.452.321.681.551.271.321.821.911.452.773.031.672.73
Cabernet Sauvignon Encín 20031.381.441.881.311.381.561.631.691.751.061.942.53.062.212.29
Ruchel Mencía 20021.631.191.8121.561.561.251.131.561.51.382.382.881.632.52
Monasterio de las Viñas1.631.632.381.811.941.381.441.061.191.311.442.5631.382.67
Castillo de Aguarón1.561.252.381.691.751.311.311.131.131.131.062.562.881.272.67
Don Luciano1.381.52.381.441.811.311.251.131.131.441.132.6931.252.85
Sliven Merlot 20021.691.442.312.51.561.751.251.251.251.751.062.813.291.312.88
Pablo Morate American Oak Barrel1.51.191.942.631.381.341.381.251.193.51.192.813.551.233.23
Pablo Morate French Oak Barrel1.311.251.882.51.381.441.311.131.192.8112.943.11.173.04
Carlos Gosalbez American Oak Barrel 20021.251.312.382.251.561.751.691.441.062.881.53.253.441.753.06
Carlos Gosalbez French Oak Barrel 20021.381.382.312.561.561.51.561.441.062.941.313.343.311.273.32
Andrés Morate American Oak Barrel1.221.282.392.111.171.331.441.281.062.861.113.193.341.193.31
Andrés Morate French Oak Barrel1.331.52.222.221.441.891.391.391.173.141.333.313.461.153.44
Luis Saavedra American Oak Barrel 20021.61.32.42.231.431.81.41.471.22.871.23.133.461.43.03
Luis Saavedra French Oak Barrel 20021.61.472.272.331.471.471.431.571.293.231.233.073.271.223.17
Carlos Gosalbez American Oak Barrel 20031.471.532.82.41.41.61.731.271.072.971.23.63.561.243.56
Carlos Gosalbez French Oak Barrel 20031.431.62.672.171.531.471.61.41.072.931.133.53.341.333.42
Table 3. Correlation coefficients and mean squared error values between the sensory panel descriptors and e-nose variables in validation.
Table 3. Correlation coefficients and mean squared error values between the sensory panel descriptors and e-nose variables in validation.
Sensory panelCorrelation CoefficientMean Squared Error
Alteration0.9196300.000560
Aromatic intensity0.9232760.001061
Chemical0.9228970.000431
Flower-like0.9527480.001094
Fruity0.8790960.001142
Herbaceous0.9530790.000479
Microbiologic0.8822640.000499
Off-flavours0.9159630.000692
Oxidation0.9320350.000552
Persistence0.9402710.000701
Phenolic0.9085090.000464
Quality0.9313050.000852
Spicy0.7998930.000744
Vegetal0.9593480.000489
Woody-like0.5087330.001286
Table 4. Confusion matrix obtained in validation of network with white young wines.
Table 4. Confusion matrix obtained in validation of network with white young wines.
W1W2W3W4W5W6W7
W15000000
W20500000
W30050000
W40005000
W50000410
W60000050
W70000005

Share and Cite

MDPI and ACS Style

Aguilera, T.; Lozano, J.; Paredes, J.A.; Álvarez, F.J.; Suárez, J.I. Electronic Nose Based on Independent Component Analysis Combined with Partial Least Squares and Artificial Neural Networks for Wine Prediction. Sensors 2012, 12, 8055-8072. https://doi.org/10.3390/s120608055

AMA Style

Aguilera T, Lozano J, Paredes JA, Álvarez FJ, Suárez JI. Electronic Nose Based on Independent Component Analysis Combined with Partial Least Squares and Artificial Neural Networks for Wine Prediction. Sensors. 2012; 12(6):8055-8072. https://doi.org/10.3390/s120608055

Chicago/Turabian Style

Aguilera, Teodoro, Jesús Lozano, José A. Paredes, Fernando J. Álvarez, and José I. Suárez. 2012. "Electronic Nose Based on Independent Component Analysis Combined with Partial Least Squares and Artificial Neural Networks for Wine Prediction" Sensors 12, no. 6: 8055-8072. https://doi.org/10.3390/s120608055

APA Style

Aguilera, T., Lozano, J., Paredes, J. A., Álvarez, F. J., & Suárez, J. I. (2012). Electronic Nose Based on Independent Component Analysis Combined with Partial Least Squares and Artificial Neural Networks for Wine Prediction. Sensors, 12(6), 8055-8072. https://doi.org/10.3390/s120608055

Article Metrics

Back to TopTop