Next Article in Journal
Dual Mode NOx Sensor: Measuring Both the Accumulated Amount and Instantaneous Level at Low Concentrations
Previous Article in Journal
Artificial Neural Network for Location Estimation in Wireless Communication Systems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Pattern Classification Using an Olfactory Model with PCA Feature Selection in Electronic Noses: Study and Application

1
College of Computer Science & Information Engineering, Zhejiang Gongshang University, Hangzhou 310018, China
2
College of Informatics and Electronics, Zhejiang University of Science and Technology, Hangzhou 310018, China
*
Author to whom correspondence should be addressed.
Sensors 2012, 12(3), 2818-2830; https://doi.org/10.3390/s120302818
Submission received: 26 December 2011 / Revised: 12 February 2012 / Accepted: 24 February 2012 / Published: 1 March 2012
(This article belongs to the Section Chemical Sensors)

Abstract

: Biologically-inspired models and algorithms are considered as promising sensor array signal processing methods for electronic noses. Feature selection is one of the most important issues for developing robust pattern recognition models in machine learning. This paper describes an investigation into the classification performance of a bionic olfactory model with the increase of the dimensions of input feature vector (outer factor) as well as its parallel channels (inner factor). The principal component analysis technique was applied for feature selection and dimension reduction. Two data sets of three classes of wine derived from different cultivars and five classes of green tea derived from five different provinces of China were used for experiments. In the former case the results showed that the average correct classification rate increased as more principal components were put in to feature vector. In the latter case the results showed that sufficient parallel channels should be reserved in the model to avoid pattern space crowding. We concluded that 6∼8 channels of the model with principal component feature vector values of at least 90% cumulative variance is adequate for a classification task of 3∼5 pattern classes considering the trade-off between time consumption and classification rate.

1. Introduction

Some bionic analytical instruments such as electronic nose (eNose) and electronic tongue (eTongue) generally consist of an array of cross-sensitive chemical sensors and an appropriate pattern recognition (PARC) method for automatically detecting and discriminating target analytes [13]. The PARC methods have become a critical component in the successful implementation of such human sense-mimicking systems [4]. Many PARC methods have been introduced into e-noses, falling into three general categories: multivariate statistical approaches [5], artificial neural network (ANN) approaches [6,7], and neuromorphic approaches [8,9]. Among those methods, biologically-inspired models and algorithms are just beginning to emerge, but are considered as promising sensor array signal processing methods for future eNose development.

Besides, feature selection or feature extraction is also an important issue for developing robust PARC models in machine learning. The outputs of sensor arrays are usually time series. Some features should be extracted to represent the original signals for PARC systems. In some cases, not all elements in feature vector obtained from the preprocessing stage are essential for classification due its high-dimensionality and redundancy [10]. For example, hundreds of variables per measurement (m/z fragments) are available as “pseudo sensor” in a mass-spectrometry-based electronic nose (MS-based eNose) [11,12]. And a significant number of them are noisy or highly correlated. Thus feature selection techniques, which finding a small set of parameters that somehow represent the overall information, have become an unavoidable preprocessing step for developing robust PARC models. The obvious advantages of those are reduction in classifier training and testing time, and overcoming the curse of dimensionality, as well as improving classifier accuracy sometimes. Many feature selection techniques are used for eNose applications, such as curve fitting [13], discrete wavelet transform [14], principal component analysis (PCA) [15,16], linear discriminant analysis[17] and so on.

In our previous work [1820], we introduced an olfactory neural network called the KIII model for pattern recognition in electronic noses. The distributed open-ended structure of the KIII model is compatible of any dimension of input vector; however, more running time is costly because six more second-order ordinary differential equations will be added into the system when one channel of the model is extended, so for a given classification task it is always advisable to reduce the dimensions of input vectors as much as possible. This paper is concerned with the relationship between classification performance of the KIII model and the outer “data” factor, i.e., the dimension of input feature vector, as well as the inner “structure” factor, i.e., the amount of its corresponding parallel channels. The PCA technique is taken as example for dimension reduction and selection for two data sets (three classes of wine and five classes of green tea, respectively). The study may be helpful to determine the input feature vector in high-dimensional data (e.g., MS-based eNose data) and its parallel channels as applying the KIII model for pattern recognition.

2. Materials and Methods

2.1. KIII Olfactory Model

Over the last decades of study on mammalian olfactory system, Freeman and his colleagues [2123] developed an olfactory model entitled KIII, which appropriately describes the whole olfactory pathway from sensory neuron to olfactory cortex, as well as offering a probable interpretation of principle of biological olfaction in the concept of mesoscopic neurodynamics [24]. The KIII model is a massively parallel architecture with multiple layers coupled with both feed-forward and feedback loops through distributed delay lines, as shown in Figure 1. External stimulus signals from receptors (R) transmit to periglomerular cells (P) and then to the olfactory bulb (OB) layer via the primary olfactory nerve (PON) in parallel. The OB layer consists of a set of mutually coupled neural oscillators, each being formed by two granule cells (G) and two mitral cells (M). Then the total output of all M1 nodes transmits via a lateral olfactory tract (LOT) to the anterior olfactory nucleus (AON) and prepyriform cortex (PC) layer, which provides the final output of the system to other parts of the brain from deep pyramidal cells (C), meanwhile back to the OB and AON layers [8].

The dynamics of every node can be described by a 2nd-order ordinary differential equation (ODE) as follows:

1 a b [ x i ( t ) + ( a + b ) x i ( t ) + a b x i ( t ) ] = j i N [ W ij Q ( x j ( t ) , q j ) ] + I i ( t )
where xi(t) and xj(t) respectively represent the state variable of the i-th and the j-th node connected with each other, and Wij indicates the connection weight from j to i. Ii(t) is an external input signal to the i-th node. The parameters a and b reflect two rate constants. Q(·) is a sigmoid function derived from the Hodgkin-Huxley equation and evaluated by experiments [25]:
Q ( x j ( t ) , q ) = { q ( 1 exp ( ( exp ( x ( t ) / q   )   ) 1 ) , x ( t ) > x 0 1 , x ( t ) < x 0 x 0 = ln ( 1 q ln ( 1 + 1 / q ) )

The dynamics of the whole KIII system can be mathematically expressed by a set of ODEs. The parameters of Wij in the KIII model were determined by a set of reliable parameter optimization algorithms to ensure the model output electroencephalogram-like waveform as observed in olfactory systems [26]. For stability and robustness of the system, a low-level Gaussian noise was imported into R and AON to simulate peripheral and central sources of noise in olfactory systems, respectively [27]. Theoretical studies showed many dynamics behaviors in the model such as limit cycle, fixed point, quasi-periodic oscillation and chaos [2227]. The optimized KIII model is not only considered as a mathematical model for neurobiology study, but also of broad interested for its good pattern learning and recognition potential.

2.2. Learning and Classification for KIII Model

The operation of learning and memorizing in the KIII model can be described as follows. The system with no stimuli is in a high dimensional state of spatially coherent basal activity, presenting a chaotic global attractor in phase space; while with external stimulus the system will soon turn to a γ-range of quasi-periodic burst, presenting a local basin of an attractor wing in phase space. Using some learning rules to adjust lateral connection weights in the OB layer, the model is able to remember a number of odor patterns. When a familiar odor presents again, its spatiotemporal pattern will soon transmit to a corresponding local attractor. Figure 2(a) shows the trajectory with M1 node and G1 node of the same channel in phase space without any stimuli (peripheral noise and central noise still exist). Figure 2(b) shows the trajectory before stimulation (blue curve of Note 1), during stimulation (black curve of Note 2) and after stimulation (black curve of Note 3). It can be seen that the system state changes quickly as stimuli change.

Figure 3 shows pattern reproduce and retrieval of a 16-channle KIII model with simple Hebbian learning rule. Every subplot presents an amplitude modulated pattern (AM) output of each M1 node in the OB layer. Figure 3(a) is pattern output of learned pattern P = [0 1 0 0 0 0 0 1 0 0 0 0 1 0 0 0]. The consistent channel position and amplitude with pattern P demonstrate a memory recall. Figure 3(b) is pattern output of an incomplete version P’ = [0 1 0 0 0 0 0 0 0 0 0 0 1 0 0 0]. Compared with pattern P, there is an AM burst in input-absence channel M1(8) (green curve), demonstrating a good pattern recovery property.

Some learning rules were proposed for the KIII model, such as Hebbian reinforcement learning rule, global habituation rule, anti-Hebbian learning rule and local habituation rule [8,23,28,29]. In this paper a training algorithm combined both Hebbian reinforcement learning rule and global habituation rule is applied to adjust the lateral connection weights of M1 nodes in the OB layer, as (3):

If         M 1 ( i ) > M m       and     M 1 ( j ) > M m       then If         W ij < h Heb           then W ij = R Heb * W ij else       W ij = h Heb else       W ij = h hab * W ij

When the activities of the i-th and the j-th M1 node are larger than the mean activity of all M1 node Mm in the OB layer, their connection weights are strengthened by RHeb (RHeb > 1); otherwise their connection weights decrease at the habituation rate hhab (hhab < 1) and eventually diminish asymptotically toward zero after several training cycles. hHeb is for value restriction of connection weights. The activity is measured by the standard deviation of AM burst in M1 node. From the view of artificial neural network, it is an unsupervised learning method and the goal is to make the connection weight space converge for small enough.

At the end of learning, the connection weights are fixed and the cluster centroids of every pattern are determined. While inputting a new sample from testing set, the Euclidean distances from the corresponding activity vector to those training pattern cluster centroids are calculated, and the minimum distance determines the classification.

All model implementations in this paper were carried out in MATLAB 7.5 (The MathWorks® Inc., Natick, MA, USA) on a Lenovo computer (Pentium® Dual CPU 1.86 GHz and RAM 2 GB) running Windows XP (Microsoft® Corp., Redmond, WA, USA).

2.3. E-nose Instrumentation

A homemade electronic nose system has been developed for data acquisition, as illustrated in Figure 4. Eight MOS Taguchi-type gas sensors (TGS832, TGS880, TGS800, TGS822, TGS826, TGS825, TGS816 and TGS812) were purchased from Figaro Engineering Inc. (Osaka, Japan). They are arranged in two PCB boards with corresponding voltage divider circuits, and then fixed in a sealed gas chamber with a volume of 315 mL. Two micro-bumps (0.5 L/min) are connected to the chamber for aroma sampling and exhausting. The measurement are controlled, monitored and recorded by homemade Delphi software in a computer which connected to a microprocessor-based circuit (MSP430 with a 12-bit ADC) via a RS-232 cable.

3. Results and Discussion

3.1. Case I: Wine Classification

To demonstrate the classification performance of the KIII model with respect to dimension varying of input feature vector, a wine data set provided by UCI Machine Learning Repository [30] was used in this paper. These data are the results of a chemical analysis of wines grown in the same region in Italy but derived from three different cultivars. The analysis determined the quantities of 13 constituents found in each of the three types of wines. All data are 13-dimension with three classes defined by the three cultivars. The data sets are very large, with 59, 71 and 48 samples per class labeled as Class 1, Class 2 and Class 3. The detail on the data set can be found in [30].

Selecting an appropriate set of features which is optimal for a given classification task is one of the most important issues in machine learning. Many techniques such as principal component analysis and independent component analysis (ICA) produce a mapping between the original feature space to a lower dimensional feature space, and are usually proposed for dimension reduction and feature selection.

In this paper, the PCA technique is used for feature selection. The aim is to pick out patterns in multivariate data and reduce the dimensionality of the input vector without a significant loss of information. PCA can also help to get an overall view of these data through giving an appropriate visual representation with fewer dimensions. Figure 5 shows the 3D-PCA plot of three classes of wine. Examining the PCA plot, the first three principal components (PCs) account for 66.5% of the variance of the data. There is not a good clustering of three classes in 3D plot. Some samples of one class drift into feature spaces of other classes. It implies that the first three principal components do not represent sufficient information for the data set. The some other of principal components should be put in to preserve more of the relevant information of the original data. There are some guidelines for determining the optimal amount of PCs [31,32]. In the following part of this section, more than three principal components will present to the KIII model as feature vector. And the correct classification rates are calculated to make sure how many principal components are sufficient.

Five samples in each class of wine data set were randomly chosen for training set and the others were used for testing. The first n principal components are selected as input feature vector for n-channel KIII model. The parameters of the KIII model in this paper are hhab = 0.5487, hHeb = 0.0395, RHeb = 1.5 and the others are from [27]. The 4th-order Runge-Kutta method with a fixed step of one (1,200 steps total) was applied for numerical integration of the ODEs. The average correct classification rates of ten trials are presented in Table 1, as well as in Figure 6. From the plot, it is very clear that the classification performance of the KIII model gradually become better and better as PC numbers increasing. Figure 7 shows the average correct classification rates of three classes with the increasing of cumulative variance percentage (CVP). Roughly speaking, more than 85% of CVP would represent sufficient information of the data set, and could get 90% of correct classification rate. However, the running time of KIII model implementation increased correspondingly. Considering the trade-off between time consumption and classification rate, the optimal dimension of feature vector is about seven in this classification task.

3.2. Case II: Tea Classification

Five brands of commercial green tea derived from five different provinces of China, i.e., Anhui, Henan, Hubei, Sichuan and Zhejiang (labeled as Class 1 to Class 5 in turn), represent different cultivars and tea-making process technology. For each brands of green tea, 22 measurements were made through the customized e-nose instrumentation described in Section 2.3. Every measurement lasted more than 400 sec with a sampling rate of 20 Sa/sec, and yielded a large data matrix. PCA was also used here to investigate how the response vectors from sensor arrays cluster and reduce the dimensionality of the raw data. Figure 8 shows the 2D-PCA plot of five classes of green tea. Examining the PCA plot, the first two principal components account for 97.26% of the variance of the data-set (PC1 and PC2 accounted for 55.83% and 41.43% of the variance, whereas PC3 and PC4 accounted for 1.96% and 0.53%, respectively). Despite the slight overlap of two clusters, we can easily observe an excellent discrimination of the five kinds of green tea.

According to the experience of Section 3.1, in this case, only the first two principal components are sufficient as selected features to train KIII model. The verification experiment was carried out as follows. Five samples in each class of tea data set were randomly chosen for training set and the others were used for testing. The first two principal components were selected as input feature vector for a 2-channel KIII model. The parameters of the model and the implementing method were the same as described above. The average correct classification rates of ten trials are presented in Table 2. Exceeding our expectations, the average classification rates of five classes are only 9.88%. Some classes are completely confused with others. When more than two principal components presented to the KIII model, the average classification rates rose strikingly from 18.35% to 87.88%, and then slowly and steadily at about 95.06%. It suggests that CVP is not a unique factor to effect KIII model’s classification performance.

As described in Section 2.2, the KIII model acts as an associative memory capable of storing previously trained patterns by means of Hebbian rule at OB layer, and recovering incomplete or corrupted pattern. From a neurodynamics viewpoint, there is a global chaotic attractor composed of a central part and multiple wings in the KIII system, and the functions above work through the transition back and forth between the central part and one wing or between the wings. When parallel channels of the KIII model reduce, there will not enough state space to represent that kind of attractor, i.e., different pattern classes. That is to say, pattern volume of the KIII model is related to the amount of its parallel channels. It is concluded that six channels is adequate for the classification task of five pattern classes in this case.

4. Conclusions

In this paper, the classification performance of a bionic olfactory model called KIII with respect to outer factor (the dimension of input feature vector) and inner factor (the amount of its parallel channels) was investigated, through recognition of three classes of wine derived from three different cultivars and five classes of green tea derived from five different provinces of China. In the first case, the PCA plot implied that the first three principal components do not represent sufficient information for the database, so more PCs were put in to feature vector and the corresponding classification rates were calculated. The results showed that classification performance of the KIII model gradually becomes better and better as PC numbers increase. Considering the trade-off between time consumption and classification rate, a PC feature of seven (about 90% of cumulative variance percentage) is an appropriate dimension for this special task. In the second case, the first two PCs (about 97% of cumulative variance percentage) were selected as input feature vector for a 2-channel KIII model. The poor results suggested that cumulative variance percentage is not the only factor to affect the KIII model’s classification performance. The results showed that classification performance of the KIII model gradually become better and better as its parallel channels extend, so we concluded that 6∼8 channels of the model with principal component feature vector values of at least 90% cumulative variance is adequate for a classification task of 3∼5 pattern classes considering the trade-off between time and classification rate. The study may be helpful to determine the input feature dimension and its parallel channels as applying the KIII model for pattern recognition (not limited to eNose applications). Future works will be addressed on analyzing higher-dimensional MS-based electronic nose data using the model.

Acknowledgments

This work was supported by Zhejiang Provincial Natural Science Foundation of China under Grant No. Y1110074 and the Scientific Research Fund of Zhejiang Provincial Education Department under Grant No. Y201010012. It was also financially supported in part by the Research Fund of the Science and Technology Department of Zhejiang Province (2008C14100) and Commonweal Technology Research Foundation of Zhejiang Province (2011C23068). The authors would like to thank UCI Machine Learning Repository for providing the wine data set.

References

  1. Vlasov, Y.; Legin, A.; Rudnitskaya, A. Electronic tongues and their analytical application. Anal. Bioanal. Chem 2002, 373, 136–146. [Google Scholar]
  2. Röck, F.; Barsan, N.; Weimar, U. Electronic nose: Current status and future trends. Chem. Rev 2008, 108, 705–725. [Google Scholar]
  3. Baldwin, E.A.; Bai, J.; Plotto, A.; Dea, S. Electronic noses and tongues: Applications for the food and pharmaceutical industries. Sensors 2011, 11, 4744–4766. [Google Scholar]
  4. Valle, M. Bioinspired sensor systems. Sensors 2011, 11, 10180–10186. [Google Scholar]
  5. Huang, Y.; Li, F.; Xia, Y.; Chen, K. Scent profiling of Cymbidium ensifolium by electronic nose. Sci. Hortic 2011, 128, 306–310. [Google Scholar]
  6. Tudu, B.; Jana, A.; Metla, A.; Ghosh, D.; Bhattacharyya, N.; Bandyopadhyay, R. Electronic nose for black tea quality evaluation by an incremental RBF network. Sens. Actuators B 2009, 138, 90–95. [Google Scholar]
  7. Bucak, İ.Ö.; Karlık, B. Hazardous odor recognition by CMAC based neural networks. Sensors 2009, 9, 7308–7319. [Google Scholar]
  8. Li, G.; Fu, J.; Zhang, J.; Zheng, J. Progress in bionic information processing techniques for an electronic nose based on olfactory models. Chin. Sci. Bull 2009, 54, 521–534. [Google Scholar]
  9. Marco, S.; Gutierrez-Gálvez, A. Recent developments in the application of biologically inspired computation to chemical sensing. AIP Conf. Proc 2009, 1137, 151–154. [Google Scholar]
  10. Gutierrez-Osuna, R. Pattern analysis for machine olfaction: A review. IEEE Sens. J 2002, 2, 189–202. [Google Scholar]
  11. Llobet, E.; Gualdrón, O.; Vinaixa, M.; El-Barbri, N.; Brezmes, J.; Vilanova, X.; Bouchikhi, B.; Gómez, R.; Carrasco, J.A.; Correig, X. Efficient feature selection for mass spectrometry based electronic nose applications. Chemometrics Intell. Lab. Syst 2007, 85, 253–261. [Google Scholar]
  12. Ampuero, S.; Bosset, J.O. The electronic nose applied to dairy products: A review. Sens. Actuators B-Chem 2003, 94, 1–12. [Google Scholar]
  13. Carmel, L.; Levy, S.; Lancet, D.; Harel, D. A feature extraction method for chemical sensors in electronic noses. Sens. Actuators B-Chem 2003, 93, 67–76. [Google Scholar]
  14. Distante, C.; Leo, M.; Siciliano, P.; Persaud, K.C. On the study of feature extraction methods for an electronic nose. Sens. Actuators B-Chem 2002, 87, 274–288. [Google Scholar]
  15. Balasubramanian, S.; Panigrahi, S.; Logue, C.M.; Gu, H.; Marchello, M. Neural networks-integrated metal oxide-based artificial olfactory system for meat spoilage identification. J. Food Eng 2009, 91, 91–98. [Google Scholar]
  16. Yu, H.; Wang, Y.; Wang, J. Identification of tea storage times by linear discrimination analysis and back-propagation neural network techniques based on the eigenvalues of principal components analysis of e-Nose sensor signals. Sensors 2009, 9, 8073–8082. [Google Scholar]
  17. Hines, E.L.; Boilot, P.; Gardner, J.W.; Gongora, M.A. Pattern analysis for electronic noses. In Handbook of Machine Olfaction: Electronic Nose Technology; Pearce, T.C., Schiffman, S.S., Nagle, H.T., Gardner, J.W., Eds.; Wiley-VCH: Weinheim, Germany, 2002; pp. 133–160. [Google Scholar]
  18. Fu, J.; Yang, X.; Yang, X.; Li, G.; Freeman, W.J. Application of biologically modeled chaotic neural network to pattern recognition in artificial olfaction. Conf. Proc. IEEE Eng. Med. Biol. Soc 2005, 5, 4666–4669. [Google Scholar]
  19. Fu, J.; Li, G.; Qin, Y.; Freeman, W.J. A pattern recognition method for electronic noses based on an olfactory neural network. Sens. Actuators B-Chem 2007, 125, 489–497. [Google Scholar]
  20. Fu, J.; Li, G.; Freeman, W.J. Pattern classification method for electronic noses based on olfactory neural network using time series. Chin. J. Sens. Actuators 2007, 20, 1958–1962, in Chinese.. [Google Scholar]
  21. Freeman, W.J. Mass Action in the Nervous System; Academic Press: New York, NY, USA, 1975. [Google Scholar]
  22. Yao, Y.; Freeman, W.J. Model of biological pattern recognition with spatially chaotic dynamics. Neural Netw 1990, 3, 153–170. [Google Scholar]
  23. Kozma, R.; Freeman, W.J. Chaotic resonance—Methods and applications for robust classification of noise and variable patterns. Int. J. Bifurcation Chaos 2001, 11, 1607–1629. [Google Scholar]
  24. Freeman, W.J. Neurodynamics: An Exploration of Mesoscopic Brain Dynamics; Springer-Verlag: London, UK, 2000. [Google Scholar]
  25. Freeman, W.J. Nonlinear gain mediating cortical stimulus-response relations. Biol. Cybern 1979, 33, 237–247. [Google Scholar]
  26. Chang, H.; Freeman, W.J. Parameter optimization in models of the olfactory neural system. Neural Netw 1996, 9, 1–14. [Google Scholar]
  27. Chang, H.; Freeman, W.J.; Burke, B. Biologically modeled noise stabilizing neurodynamics for pattern recognition. Int. J. Bifurcation Chaos 1998, 8, 321–345. [Google Scholar]
  28. Gutierrez-Galvez, A.; Gutierrez-Osuna, R. Increasing the separability of chemosensor array patterns with Hebbian/anti-Hebbian learning. Sens. Actuators B-Chem 2006, 116, 29–35. [Google Scholar]
  29. Gutierrez-Osuna, R.; Gutierrez-Galvez, A. Habituation in the KIII olfactory model with chemical sensor arrays. IEEE Trans. Neural. Netw 2003, 14, 1565–1568. [Google Scholar]
  30. Frank, A.; Asuncion, A. UCI Machine Learning Repository; School of Information and Computer Science, University of California: Irvine, CA, USA, 2010. Available online: http://archive.ics.uci.edu/ml (accessed on 7 May 2011).
  31. Peres-Neto, P.R.; Jackson, D.A.; Somers, K.M. How many principal components? Stopping rules for determining the number of non-trivial axes revisited. Comput. Stat. Data Anal 2005, 49, 974–997. [Google Scholar]
  32. Dray, S. On the number of principal components: A test of dimensionality based on measurements of similarity between matrices. Comput. Stat. Data Anal 2008, 52, 2228–2237. [Google Scholar]
Figure 1. The topological structure of the KIII olfactory model [19].Notes: The lateral connection weights of M1 nodes (red arrows) are adjustable for learning, while the others (black arrows) are usually fixed for system stability.
Figure 1. The topological structure of the KIII olfactory model [19].Notes: The lateral connection weights of M1 nodes (red arrows) are adjustable for learning, while the others (black arrows) are usually fixed for system stability.
Sensors 12 02818f1 1024
Figure 2. The trajectory in phase-space. (a) Before; (b) During stimulation.
Figure 2. The trajectory in phase-space. (a) Before; (b) During stimulation.
Sensors 12 02818f2 1024
Figure 3. Binary experiments for (a) Pattern reproduce; (b) Pattern retrieval in a 16-channle KIII model with simple Hebbian learning rule.
Figure 3. Binary experiments for (a) Pattern reproduce; (b) Pattern retrieval in a 16-channle KIII model with simple Hebbian learning rule.
Sensors 12 02818f3 1024
Figure 4. Photo of the experimental set-up with the customized electronic nose system.
Figure 4. Photo of the experimental set-up with the customized electronic nose system.
Sensors 12 02818f4 1024
Figure 5. The 3D-PCA plot of three classes of wine derived from three different cultivars.Note: Symbols: O-Class 1; *-Class 2 and □-Class 3.
Figure 5. The 3D-PCA plot of three classes of wine derived from three different cultivars.Note: Symbols: O-Class 1; *-Class 2 and □-Class 3.
Sensors 12 02818f5 1024
Figure 6. The correct classification rate of three classes with respect to PC numbers.
Figure 6. The correct classification rate of three classes with respect to PC numbers.
Sensors 12 02818f6 1024
Figure 7. The average correct classification rate with respect to cumulative variance percentage of principal components.
Figure 7. The average correct classification rate with respect to cumulative variance percentage of principal components.
Sensors 12 02818f7 1024
Figure 8. The 2D-PCA plot of five classes of green tea derived from five different provinces of China.Note: Symbols: O-class 1; *-class 2; □-class 3; ♦-class 4; x-class 5.
Figure 8. The 2D-PCA plot of five classes of green tea derived from five different provinces of China.Note: Symbols: O-class 1; *-class 2; □-class 3; ♦-class 4; x-class 5.
Sensors 12 02818f8 1024
Table 1. The classification performance of the KIII model with the increase of the dimension of input feature vector.
Table 1. The classification performance of the KIII model with the increase of the dimension of input feature vector.
PCs456789
Class 179.1985.4595.1495.5698.3498.82
Class 273.0887.8293.2996.9695.9196.73
Class 381.2584.3592.0393.8392.0794.14
Average77.8485.8793.4995.4595.4496.57
CVP73.6080.1685.1089.3492.0294.24

Note: all classification rate and PCA values in the table are expressed as percentages.

Table 2. The classification performance of the KIII model for five kinds of green tea with the increase of its parallel channels.
Table 2. The classification performance of the KIII model for five kinds of green tea with the increase of its parallel channels.
PCs234567
Class 110.5914.1230.5937.0687.0695.88
Class 210.0018.2425.8835.2986.4794.12
Class 37.0615.2927.6544.1292.9493.53
Class 412.3521.7627.0641.7687.0694.12
Class 59.4122.3534.7148.2485.8897.65

Average9.8818.3529.1841.2987.8895.06

CVP97.2699.2299.7599.9199.9699.99

Note: all classification rate and PCA values in the table are expressed as percentages.

Share and Cite

MDPI and ACS Style

Fu, J.; Huang, C.; Xing, J.; Zheng, J. Pattern Classification Using an Olfactory Model with PCA Feature Selection in Electronic Noses: Study and Application. Sensors 2012, 12, 2818-2830. https://doi.org/10.3390/s120302818

AMA Style

Fu J, Huang C, Xing J, Zheng J. Pattern Classification Using an Olfactory Model with PCA Feature Selection in Electronic Noses: Study and Application. Sensors. 2012; 12(3):2818-2830. https://doi.org/10.3390/s120302818

Chicago/Turabian Style

Fu, Jun, Canqin Huang, Jianguo Xing, and Junbao Zheng. 2012. "Pattern Classification Using an Olfactory Model with PCA Feature Selection in Electronic Noses: Study and Application" Sensors 12, no. 3: 2818-2830. https://doi.org/10.3390/s120302818

Article Metrics

Back to TopTop