Next Article in Journal
Interfacial Microstructure and Properties of Clad Rebar Prepared by Clean-Interface Assembly and Vacuum Hot-Rolling
Next Article in Special Issue
Code-Switching in Automatic Speech Recognition: The Issues and Future Directions
Previous Article in Journal
Special Issue on Emerging Technologies in Food Science: Advances in Microencapsulation
Previous Article in Special Issue
Target Speaker Extraction by Fusing Voiceprint Features
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Speech Emotion Recognition Using a Dual-Channel Complementary Spectrogram and the CNN-SSAE Neutral Network

1
College of Information and Computer, Taiyuan University of Technology, Jinzhong 030600, China
2
Department of Physics and Electronic Engineering, Yuncheng University, Yuncheng 044000, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2022, 12(19), 9518; https://doi.org/10.3390/app12199518
Submission received: 16 August 2022 / Revised: 12 September 2022 / Accepted: 19 September 2022 / Published: 22 September 2022
(This article belongs to the Special Issue Advances in Speech and Language Processing)

Abstract

:

Featured Application

Emotion recognition is the computer’s automatic recognition of the emotional state of input speech. It is a hot research field, resulting from the mutual infiltration and interweaving of phonetics, psychology, digital signal processing, pattern recognition, and artificial intelligence. At present, speech emotion recognition has been widely used in the fields of intelligent signal processing, smart medical care, business intelligence, assistant lie detection, criminal investigation, the service industry, self-driving cars, voice assistants of smartphones, and human psychoanalysis, etc.

Abstract

In the background of artificial intelligence, the realization of smooth communication between people and machines has become the goal pursued by people. Mel spectrograms is a common method used in speech emotion recognition, focusing on the low-frequency part of speech. In contrast, the inverse Mel (IMel) spectrogram, which focuses on the high-frequency part, is proposed to comprehensively analyze emotions. Because the convolutional neural network-stacked sparse autoencoder (CNN-SSAE) can extract deep optimized features, the Mel-IMel dual-channel complementary structure is proposed. In the first channel, a CNN is used to extract the low-frequency information of the Mel spectrogram. The other channel extracts the high-frequency information of the IMel spectrogram. This information is transmitted into an SSAE to reduce the number of dimensions, and obtain the optimized information. Experimental results show that the highest recognition rates achieved on the EMO-DB, SAVEE, and RAVDESS datasets were 94.79%, 88.96%, and 83.18%, respectively. The conclusions are that the recognition rate of the two spectrograms was higher than that of each of the single spectrograms, which proves that the two spectrograms are complementary. The SSAE followed the CNN to get the optimized information, and the recognition rate was further improved, which proves the effectiveness of the CNN-SSAE network.

Graphical Abstract

1. Introduction

Emotion recognition has become an important research field in neuroscience, computer science, cognitive science, and medicine [1,2]. Although facial expressions [3], body movements [4], gestures [5], and EEG signals [6] can provide a good representation of human emotions, many researchers extract raw emotions from speech signals because speech is a fast and normal means of communication between humans [7]. Speech emotion recognition (SER) has wide application prospects in modern intelligent systems, such as self-driving cars, voice assistants of smartphones, human psychoanalysis, and medical services [8,9]. Despite the many applications, SER is a challenging assignment because emotions are not objective. There is no consensus on how to classify or measure emotions [10].
SER usually consists of four parts: preprocessing, feature extraction, feature optimization or selection, and classification [11]. Feature extraction is very important for SER. Speech features include acoustic features and spectrographic features [12], whereas acoustic features include prosodic features [13], spectral features [14], and quality features [9,15]. They are basic features, called low-level descriptors (LLD) [16]. Prosodic features and quality features are limited to the time domain and reflect less frequency-domain information [17,18]. Spectral features are limited to the frequency domain, and reflect less time-domain information. They are unable to explore the underlying information of the speech. A popular method is to extract the features through a neural network. A typical application is learning from standard spectrograms and log-mel spectrograms using CNN [19]. Spectrograms are a widely used two-dimensional representation of speech signals: they contain information in both the frequency domain and time domain, so they can more comprehensively reflect emotion-related information [20].
The spectrogram has become an active research topic in recent years, but it suffers from two problems. The first is how to extract information from spectrograms more efficiently and reliably without losing any information. Most studies have been based on standard or Mel spectrograms, each of which has its shortcomings. The standard spectrogram can be obtained directly as the logarithm of the spectral energy, ignoring the characteristics of human hearing [21]. In contrast, the Mel spectrogram is obtained by calculating the logarithm of the spectral energy through the Mel filter bank. Mel filters are designed according to the auditory characteristics of the human ears, paying more attention to the low-frequency part of speech [22]. Low-arousal emotions in speech contain more unvoiced and trailing sounds, which are reflected more in the high-frequency part. To capture low-arousal emotions, the IMel spectrogram, which focuses on high-frequency components, is proposed in this paper. To generate the IMel spectrogram, IMel filter banks replace the Mel filter banks used to generate the Mel spectrogram. The Mel spectrogram amplifies low-frequency information, the IMel spectrogram amplifies high-frequency information, and the two spectrograms complement each other when they are combined.
The second problem is the optimization and selection of the deep features of the spectrogram. Compared with traditional methods, deep learning technology for SER have the advantages of detecting complicated structures and features, and extracting and tuning features automatically. There are many techniques, Recurrent Neural Networks (RNNs), Long Short-Term Memory (LSTM), and Deep Neural Networks (DNNs) are very effective for speech classification [16]. However, Neural Networks (CNNs) are suitable for video and image, and provide efficient results [10]. CNN can provide state-of-the-art accuracy for classification [6]. Because of their strong expressiveness, we choose CNN to extract high-level and deep feature information from the spectrogram [23]. For example, in reference [24], time-frequency information in the context of the Mel spectrogram was analyzed using a CNN.
In addition to the above problems, The CNN deep features may contain redundancy, which may reduce recognition accuracy. To solve this problem, researchers have used various methods to optimize or select features. Traditional optimization or dimension-reduction methods include principal component analysis (PCA) [25] and linear discriminant analysis (LDA) [26], which are linearly dependent, and are widely used for SER. However, the non-linear feature selection algorithms perform better than linear feature selection algorithms [2]. For example, adversarial auto-encoder (AE), which is a linear feature selection algorithm, reduced the dimensions, optimized the features, and improved accuracy compared to PCA and LDA [27]. In this paper, a stacked sparse autoencoder (SSAE) is used to refine the deep features of a CNN: that is, multiple sparse autoencoders (SAEs) are superimposed to form a deep model structure with a strong ability to compute complex functions. The features extracted by multiple layers are more vivid and representative than those extracted by an SAE [28].
Different spectrograms can provide different types of speech emotion information. Instead of just using only a single type of information, we implemented a fusion strategy to combine multiple information sources. The Mel spectrogram and IMel spectrogram of the preprocessed speech are sent to two channel structures CNN separately, forming a dual-channel complementary structure. A CNN is used to extract one-dimensional features from the two-dimensional spectrograms of each of the two channels. The first channel extracts the deep features of the Mel spectrogram and highlights the low-frequency information. The second channel extracts the deep features of the IMel spectrogram and highlights the high-frequency information. The spliced features are then sent to an SSAE nonlinear structure to perform feature optimization. Finally, a softmax layer is used to identify the output.
To summarize, the main improvements of this paper are as follows.
(1)
The IMel spectrogram, which is the complement of the Mel spectrogram, that focuses on the high-frequency part of the signal, is proposed. It can supplement the Mel spectrogram focusing on the low-frequency part of the signal.
(2)
The CNN-SSAE network, to extract deep optimized features of the spectrogram, is proposed. The deep features extracted by CNN often contain redundancy information that affects the recognition. In order to obtain the essential information of the features, this paper adopts SSAE to learn more abstract and representative features.
(3)
To improve the recognition rate, a dual-channel structure based on the Mel and IMel spectrograms is proposed. A single channel based on the Mel spectrogram is incomplete, and two channels are complementary.
The structure of this paper is as follows. Section 2 introduces the related research. In Section 3, the architecture of SER is described. Section 4 describes the generation of the IMel spectrogram, and provides theoretical analysis. Section 5 presents the CNN-SSAE neural network for extracting deep optimized features. Section 6 introduces the datasets, features, experimental settings, and evaluation scenarios. Section 7 presents the conclusions and discusses the future research directions of SER.

2. Related Work

2.1. Development of Acoustic Features

Mel-frequency cepstral coefficients (MFCCs) are the most widely used perceptual incentive features. Their improved forms [29], such as log frequency power coefficients [30] and linear prediction cepstral coefficients [31], have been used in SER. In addition to these basic features, researchers have constructed many new acoustic features for SER. In reference [32], inspired by the AM-FM modulation model, the features of modulation spectra and modulation frequency were extracted and combined with cepstral features. In reference [33], three features were selected: the MFCC feature of the typical human auditory model, the feature of Zero Crossings with Maximal Teager Energy Operator (ZCMT) of the uncompensated human auditory model, and the feature of Glottal Compensation to Zero Crossings with Maximal Teager Energy Operator (GCZCMT) feature. In reference to [34], a new set of harmonic features was proposed. Reference [35] proposed weighted spectral features based on Local Hu moments (HuWSF), which is based on Local Hu moments. However, these features do not reflect the relationship between the frequency domain and time domain at the same time, and therefore many researchers have turned their attention to spectrogram.

2.2. Development of the Spectrogram

Previous studies of the spectrogram can be divided into three situations:
(1)
Improving the recognition network. Reference [36] inputs the standard spectrogram into a deep CNN with a rectangular kernel for recognition, reference [37] inputs the standard spectrogram into a CNN with an improved pooling strategy for recognition, and reference [38] inputs the Mel spectrogram into a visual attention CNN for recognition.
(2)
Processing the spectrogram to improve the recognition effect better. In references [39,40], log-Mels, deltas, and delta-deltas were used as the inputs of the structure in the hybrid system. In reference [41], the K-means clusters after extracting 88 viterbilt symbols from each frame signal were sent to a three-dimensional CNN by the standard spectrogram of the key frames. In reference [42], the multiscale standard spectrogram was processed by a Chirplet filter, and the local direction, time information, intensity, and frequency comparison features were sent to a support vector machine (SVM).
(3)
Combining acoustic features with the spectrogram. Reference [43] combined the IS09 feature in openSMILE [44] with the Mel spectrogram and sent it to the model of the attention-LSTM- attention model. Reference [45] proposed combining the acoustic features and spectrogram features to improve the recognition rate.
To summarize, these studies mainly used the standard spectrogram or Mel spectrogram to construct SER models. To represent emotional information in a more comprehensive manner, we propose the IMel spectrogram, which pays more attention to high-frequency information.

2.3. Development of Feature Optimization

Overall, high dimensionality has a substantial impact on classification accuracy and efficiency. To ensure higher precision and shorter calculation time, the number of feature dimensions should be reduced. In reference [46], three-dimensionality reduction methods—LDA, PCA, and PCA + LDA—were used to reduce the number of dimensions of the features. In reference [34], the sequential floating forward selection algorithm, which is an iterative method for finding features that are close to the optimal features, was proposed. In reference [47], a feature selection method, particle swarm optimization-assisted biogeography-based optimization, was proposed. In reference [48], an optimal discriminative dimensionality-reduction matrix was used in the QPSO algorithm to reduce the number of dimensions.
In recent years, in deep learning, the SAE has also been applied to research feature dimensions. The multilevel structure of the SAE is suitable for the fusion and compression of information in input data and can extract necessary features. The SSAE, which consists of several SAE units, is used to extract the key information of the features. The following results prove the feasibility of using the SSAE to optimize the features. In reference [49], to identify distinguishing features of nuclei, an SSAE was used to learn high-level features from pixel intensities. In reference [50], an SSAE was used to extract high-level features from fast and noncell samples in the training set. In reference [51], a method of feature representation for multicell Pap smear images, using an SSAE was proposed. Reference [52] used a trained SSAE to extract high-level features from image patches. Therefore, an SSAE can capture advanced features by learning from low-level features. Advanced high-level features can differentiate between various emotions.
To combine the advantages of the CNN and SSAE to extract deep optimized features, a CNN-SSAE neural network method is proposed in this paper.

3. Architecture of Speech Emotion Recognition

The dual-channel structure for SER proposed in this paper is shown in Figure 1, The method comprises the following steps:
(1)
Preprocessing, which mainly includes endpoint detection, framing, and windowing, is performed.
(2)
The Mel spectrogram and IMel spectrogram, which are complementary, are generated.
(3)
The CNN-SSAE deep optimized features from the Mel spectrograms and IMel spectrograms are extracted and spliced.
(4)
The spliced features are sent to the softmax layer for identification.

4. IMel Spectrogram

This section describes the IMel spectrogram and its application to SER. First, the generation process and basic theory of the IMel spectrogram are introduced. Second, the performance of the IMel spectrogram is analyzed and compared with that of the Mel spectrogram.

4.1. Generation of the Mel Spectrogram

The commonly used Mel spectrogram contains the features of human auditory perception. It is based on the Mel frequency domain, in which linear frequencies are mapped to the Mel frequency. The equation below shows the relation:
F M e l = 2595 ln ( 1 + 700 )
where f l i n is the frequency, expressed on a linear scale, and the unit is Hz; f M e l is the frequency on the Mel scale, and the unit is Mel. In Figure 2a, f M e l and f l i n are logarithms, and the resulting Mel filter banks are shown in Figure 2b. The resolution at the low frequency is higher than that at the high frequency, which is consistent with the characteristics of human hearing perception, and the Mel spectrogram based on the Mel filter banks has the same property. Therefore, the Mel spectrogram has been widely used and has achieved good results in the field of speech recognition. However, when applied to SER, it has certain limitations.
Figure 3 shows Russell’s two-dimensional expression of emotional space [53]. The abscissa is valence and the ordinate is arousal. Arousal indicates the intensity of emotions: a higher numerical value corresponds to a higher degree of emotional stimulation. The speech signals for high-aroused emotions, such as happiness and anger, include more quasi-periodic and low-frequency parts. The speech signals for Low-arousal emotions, such as sadness and tiredness, include more nonperiodic high-frequency parts and tail parts. The Mel spectrogram has a high resolution in the low-frequency part and is therefore suitable for high-arousal emotions. However, for some emotions with low arousal, more attention should be given to the high-frequency parts to achieve better SER. Therefore, in this paper, the IMel frequency domain is proposed and the IMel spectrogram, which is complementary to the Mel spectrogram and is generated by IMel filter banks, is obtained.

4.2. Generation of the IMel Spectrogram

The IMel spectrogram is generated by using an improved IMel filter bank, and the transformation relation is as follows:
f I M e l = 2595 × exp ( 1 + f l i n / 700 )
f I M e l is expressed in the IMel frequency domain, which is opposite to the Mel frequency domain. Figure 4a shows the exponential relationship between f l i n and f I M e l in the IMel frequency domain. As shown in Figure 4b, the high-frequency part of the IMel filter banks has a narrow bandwidth and high resolution, thus enhancing the influence of high-frequency signals. Therefore, the IMel spectrograms generated using the IMel filter banks have the same characteristics.
Figure 5a,b show the same person saying the same sentence “The dog is sitting by the door” with different emotions. In Figure 5a, anger is intense and quasi-periodic, with higher energy and more low-frequency parts. The speech in Figure 5a has the Mel spectrogram shown in Figure 5c, with clear stripes, highlighting the emotional information of the high-arousal signal. In Figure 5b, the speech is sad, weak, and low in energy, the range of the ordinate is [−0.1,0.1], which is ten times smaller than the range of angry speech. The unvoiced segment “S” in Figure 5b, marked by the ellipse, is the key point of detection. The Mel spectrogram in Figure 5d and the IMel spectrogram in Figure 5f corresponding to “S” are displayed in the ellipses. The IMel spectrogram amplifies the role of the high-frequency part and highlights the emotional information of the low-arousal signal. This paper combines the Mel spectrogram highlighting the low-frequency part with the IMel spectrogram highlighting the high-frequency part to apply their complementary effects to SER.
As shown in Figure 6, the speech signal is preprocessed, and then the energy obtained after a fast Fourier transform (FFT) is calculated. After passing through the Mel filter banks and the IMel filter banks separately, the logarithm is calculated, and finally the Mel spectrogram and the IMel spectrogram are obtained. The steps performed to generate the IMel spectrogram are as follows:
(1)
The speech signal x s ( n ) = [ s 1 ( n ) , s 2 , ( n ) , , s i ( n ) , , s l ( n ) ] is framed, where s i represents each frame and l denotes the number of frames.
(2)
Windows s i ( n ) = s i ( n ) × w ( n ) are added, where the window function ω ( n ) is a Hamming window and n is the window length.
(3)
Each frame of the signal s i ( n ) is transformed by FFT, X i ( r ) = n = 0 N 1 s i ( n ) × e j 2 π r / N , r = 0 , 1 , N 1 and the energy S i ( r ) = | X i ( r ) | 2 is obtained. is the number of points of the FFT.
(4)
f l and f h are converted from a linear scale to the IMel scale and are expressed as f l and f h :
f l = 2595 × exp ( 1 + f l / 700 )
f h = 2595 × exp ( 1 + f h / 700 )
where f l and f h are the lowest frequency and the highest frequency of the filter frequency respectively.
(5)
The center frequency of each filter in the filter banks is Calculated:
f c j = f l + j × ( f u f l ) / ( M + 1 )
where j = 1 , 2 , , M 1 , M is the number of filters.
(6)
The calculated center frequency of each filter is converted from the IMel scale to a linear scale:
f c j = 700 × ( ln ( f c j / 2595 ) 1 )
(7)
The calculated center frequency on the linear scale is rounded to the nearest point of the FFT:
f = ( N F F T + 1 ) × ( f c j / F s )
where N F F T represents the scale of the decomposition and F s represents the sampling frequency.
(8)
Where H j ( m ) represents the transfer function of the IMel filter bank, the output O j ( m ) = S i ( r ) × H j ( m ) is computed and the logarithm is calculated as L O j ( m ) = lnO j ( m ) , where J is the number of filters, and the sample of each filter is m . Finally, groups of L O j are arranged vertically to form the IMel spectrogram of each frame, which is represented as S F i .
(9)
The features of all frames S F i are combined horizontally, and the IMel spectrogram is obtained.

5. CNN-SSAE Neural Network

As shown in Figure 1, the CNN-SSAE neural network proposed in this paper consists of a CNN and an SSAE. First, the spectrogram is taken as input, and the deep features are obtained by the CNN, Second, the deep optimized features are obtained by SSAE dimension reduction. Finally, the recognition results are classified through the softmax layer.

5.1. Convolutional Neural Network

The network structure of the CNN is shown in Figure 7 and the parameters of it are presented in Table 1. The CNN can transform the two-dimensional spectrogram into one-dimensional deep features to capture the emotional information of the speech. The characteristics of the CNN proposed in this paper are as follows: (1) After adding the batch normalization (BN) layer to the convolutional layers Con1, Con2, Con3 and Con4, the structure of the whole network is regularized to prevent overfitting, allow a higher learning rate, greatly improve the training efficiency, and solve the vanishing gradient problem. (2) The Global Average Pooling (GAP) layer is used instead of the fully connected layer. In the fully connected layer, a sub-region of each feature subgraph is set for averaging, the sub-regions are slid, and finally, all the average values are connected in series. In contrast, using GAP, each feature subgraph obtains an average value; this increases the emotional information somewhat, reduces the number of parameters, realizes model compression, and improves efficiency.
The deep features of the spectrogram obtained by the CNN may be redundant. Therefore, an SSAE is adopted to further optimize the deep features, reduce the number of dimensions, and improve the recognition rate.

5.2. Stacked Sparse Autoencoder

Multiple SAEs are stacked together to form an SSAE. It is suitable for the fusion and compression of information in input data and can extract necessary features. The SAE is usually used for data compression and fusion. The name refers to the sparse restrictions that are added to each hidden layer. The previous layer contains more hidden cells than the latter layer, and the hidden features learned in the latter layer are more abstract. This structure is similar to the working state of the human brain.
The SAE1 shown in Figure 8 is composed of an input layer, a hidden layer and an a output layer. The input data x1 are mapped to a hidden laye1; this is called the encoding process. The hidden layer1 is remapped to the reconstructed data y1; this is called the decoding process. This process should minimize the error function E from x 1 to y1:
E = n = 1 N x n y n 2 + β j = 1 M K L ( ρ ρ ^ j )
where n denotes the number of samples, j denotes the number of hidden layers, ρ denotes the target value of the average activation degree, and ρ ^ j denotes the average activation degree of the j-th layer.
ρ ^ j = 1 N n = 1 N f ( W j x + b j )
K L ( ρ ρ ^ j ) means K L distance (Kullback-Leibler divergence):
K L ( ρ ρ ^ j ) = ρ log ( ρ ρ j ) + ( 1 - ρ ) log ( 1 ρ 1 ρ ^ j )
When training the network, the parameters should be constantly adjusted to minimum β . A value of ρ closer to 0 corresponds to a smaller average activation degree of the middle layer ρ ^ j . In addition, it is necessary to constantly adjust the connection weight W j and bias b j to minimum the error function E .
In the SSAE pre-training show in Figure 8, the weights and biases of each layer are W j and b j , the adjusted parameters are δ j and δ j , and the weights and biases after fine-tuning are W j * = W j + δ j and b j * = b j + δ j . SSAE is formed by stacking multiple layers. First, the 128-dimensional deep features of the spectrogram are as input layer of S A E 1 , whose hidden layer1 output is the 118-dimensional vector h ( 1 ) , where h ( 1 ) = s i g m o i d ( b 1 * + W 1 * x ) . Second, hidden layer1 is regarded as the input of S A E 2 , whose hidden layer2 output is the 108-dimensional vector h ( 2 ) , h ( 2 ) = s i g m o i d ( b 2 * +   W 2 * h ( 1 ) ) . Third, hidden layer2 is regarded as the input of S A E 3 , whose hidden layer output is the 98-dimensional vector h ( 3 ) , where h ( 3 ) = s i g m o i d ( b 3 * + W 3 * h ( 2 ) ) . Finally, the deep optimized features h ( 3 ) are fed into the softmax layer for recognition.

6. Experiment

6.1. Experimental Datasets

To verify the effectiveness of our proposed spectrogram and dual-channel complementary model, we tested it on three widely used datasets: the Berlin dataset of German emotional speech (EMO-DB) [54], the Surrey Audio-Visual Expressed Emotion dataset (SAVEE) [55], and the Ryerson audiovisual dataset of emotional speech and song (RAVDESS) [56].
The EMO-DB database was created by researchers at the University of Berlin and is an emotional database in Germany. This database contains 535 utterances produced by 10 professional actors, including 5 males and 5 females), who produced 49, 58, 43, 38, 55, 35, 61, 69, 56, and 71 utterances respectively. These data are made up of seven different emotions: happiness, neutral, anger, boredom, disgust, sadness, and fear. The Sampling rate is 48 kHz (compressed to 16 kHz), and 16-bit quantization is adopted. It belongs to acted, and discrete type, access type is open, and modalities is audio.
The SAVEE database records the data of four native English-speaking males (DC, JE, JK, KL), graduate students and researchers from Surrey University, aged from 27 to 31. It contains a total of 480 utterances, each of which produces 120 utterances. The average length of each utterance is 4 s with a sampling rate of 44.1 kHz. This database consists of seven different emotions: happiness, sadness, anger, disgust, fear, neutral and surprise. It belongs to acted, and discrete type, access type is free, and modalities is audio/visual.
The RAVDESS database is an English emotional dataset, which consists of 1440 utterances produced by 24 actors, including 12 males and 12 females, who expressed eight different emotions: happiness, sadness, surprise, anger, calmness, disgust, fearfulness, and neutral. The database is widely used for emotional song and speech recognition, with a sampling rate of 48 kHz. It belongs to evoked and discrete types; access type is free and modalities are audio/visual.
In the above databases, emotion is expressed in different languages and cultures; for example, pronunciation characteristics of German and English differ; however, they have some common characteristics. Low-arousal emotions, such as sadness and disgust, have low energy and the speed of speech is generally slow. High-arousal emotions, such as happiness and anger, have high energy and the emotions are expressed strongly. For all datasets, the Mel spectrogram and IMel spectrogram are always complimentary.

6.2. Experimental Features

To verify that the Mel spectrograms and the IMel spectrograms are complementarity and to assess the effectiveness of the CNN-SSAE, six experimental features were designed.
(1)
Mel: The deep features of the Mel spectrogram were extracted by the CNN and sent to an SVM for recognition.
(2)
IMel: The deep features of the IMel spectrogram were extracted by the CNN and sent to an SVM for recognition.
(3)
Mel + IMel: The Mel deep features and IMel deep features were combined, and then send to an SVM for recognition.
(4)
MelSSAE: The Mel deep features were optimized and identified by the SSAE.
(5)
IMelSSAE: The IMel deep features were optimized and recognized by the SSAE.
(6)
(Mel + IMel) SSAE: The Mel deep features and IMel deep features were combined and then optimized and identified by the SSAE.

6.3. Parameter Settings

We chose the Hamming window to have a length of 25 ms and a shift of 10 ms and set the length of the FFT to 512. In the CNN-SSAE models, the batch size, learning rate of RMSProp, and dropout rate were set to 32, 0.001, and 0.1–0.2, respectively. To ensure the comprehensiveness of the experiment, fivefold cross-validation was adopted and the average value of five test results was taken to ensure the correctness of the experimental results.

6.4. Experimental Scenarios

To ensure that the experiment was comprehensive, we employed three experimental scenarios to analyze the results: speaker independent (SI), speaker-dependent (SD), and gender-dependent (GD) [57,58].
(1)
SD: The samples were randomly divided into two groups: the training set, containing 80%, and the test set, containing the remaining 20%.
(2)
SI: The samples were divided into two groups according to the subjects. The test set was composed of all the samples spoken by one subject, and the training set contained samples spoken by the remaining subjects.
(3)
GD: GD consists of two scenarios, GD-male and GD-female, according to gender. In the GD-male scenario, the samples were divided into two groups, with male data as the training set and female data as the test set. Conversely, in the GD-female scenario, female data were used as the training set, and male data were used as the test set.

6.5. Evaluation Indexes

Two evaluation indexes are used to measure the SER performance in this paper. The unweighted accuracy (UA) is the average value of recall for each class, and the weighted accuracy (WA) is the number of samples correctly classified divided by the total number of samples.
W A = k = 1 K t u r e p o s i t i v e s k k = 1 K t o t a l p o s i t i v e s k
U A = 1 K k = 1 K t u r e p o s i t i v e s k t o t a l p o s i t i v e s k
where the number of emotional categories is represented as K , and the actual number of samples with this emotion is expressed as t o t a l p o s i t i v e s k , t r u e p o s i t i v e s k is the number of samples correctly classified for emotion category k .

6.6. Experimental Results and Analysis

6.6.1. Analysis of the IMel Spectrogram

(1) Factors influencing the spectrogram
How to get the most suitable spectrogram is the implementation challenge, because in the process of spectrogram implementation, the FFT length is 256, the window length is 256, the window shift is 128, and many other parameters have an impact on the results, such as the number M of filters and the frequency range R.
The number M of filters in the filter bank has an important influence on the display of the spectrogram. Figure 9a shows the filter banks with 20 filters, Figure 9b shows the Mel spectrogram corresponding to Figure 9a,c shows a filter bank with 40 filters, and Figure 9d shows the Mel spectrogram corresponding to Figure 9c. It can be observed that the spectrogram of Figure 9d is more detailed and obvious than that of Figure 9b, and is therefore more suitable for emotion detection.
The vertical axis of the spectrogram indicates frequency and the horizontal axis indicates time. In this paper, R represents the frequency range of the spectrogram and R max represents the maximum frequency. Both R and R max have an important impact on recognition. According to the Nyquist principle, the maximum frequency f m is equal to half of the sampling frequency f s . For example, the sampling frequency of the SAVEE dataset is 44,200 Hz and the maximum frequency is 22,100 Hz. As shown in Figure 10a,c, the frequency ranges of the Mel filter banks and IMel filter banks are 0–4000 Hz, and Figure 10b,d show the corresponding Mel spectrogram and IMel spectrogram. In Figure 10e,g, the frequency ranges of the Mel filter and IMel filter are both 0–8000 Hz, and Figure 10f,h show the corresponding Mel spectrogram and IMel spectrogram. In Figure 10i,k, the frequency ranges of the Mel filter and IMel filter are both 0–21,000 Hz, and Figure 10j,l show the corresponding Mel spectrogram and IMel spectrogram. It can be observed that the low-frequency decomposition of the Mel spectrogram in Figure 10j is not sufficiently detailed, whereas the attention in Figure 10l is mostly in the invalid region. Therefore, the following work later section of this paper is to set the frequency range at 0–4000 Hz or 0–8000 Hz.
(2) Complementarity analysis of Mel spectrogram and IMel spectrogram
To prove that the Mel spectrogram and IMel spectrogram are complementary, the correlation analysis method is used in this paper. For two related variables, the correlation coefficient (whose value is between −1 and 1) indicates the degree of correlation. A greater absolute value of the correlation coefficient indicates more obvious emotional information and a brighter reflection in the image. As shown in Figure 11a–c, a correlation analysis of Mel deep features, IMel deep features, and Mel + IMel deep features of 480 sentences in the SAVEE database was conducted and expressed in the form of images. The abscissa and ordinate represent the 480 samples, The color of anger is bright in Figure 11a, dark in Figure 11b, and brighter in Figure 11c. The color of sadness is dark in Figure 11a, bright in Figure 11b, and brighter in Figure 11c. This shows that the Mel deep features and the IMel deep features are complementary, Moreover, the Mel + IMel deep features can maintain the advantage of each of them and can enhance the overall advantage, thereby playing a role in enhancing emotional information.

6.6.2. Analysis of Recognition Accuracy

(1) Recognition results of each experimental scenarios
The sampling frequency of EMO-DB is 16,000 Hz, and that of SAVEE and RAVDESS is 44,200 Hz. Although the sampling frequencies are different, the frequency range of the speech signal is 200–3400 Hz. For all types of datasets, the energy is mainly concentrated in the range of 0–4000 Hz and there is some energy in the range of 4000–8000 Hz in some cases. From the above analysis, R max = 4000 Hz and R max = 8000 Hz were compared, In addition, to assess the influence of the number of filters M in the filter bank, M = 40, M = 60, and M = 80 were compared.
Table 2 shows that when R max = 8000 Hz, M = 60 and E = WA, the highest recognition rate of the EMO-DB database was 94.79 + 1.77%, that of the SAVEE database was 88.96 + 3.56%, and that of the RAVDESS database was 83.18 + 2.31%. This indicates that the frequency range of 0–8000 Hz reflects more comprehensive information. When the number M of filters was equal to 60, the decomposition of the spectrogram was more appropriate. Therefore, the following experiments for SI, GD and various types of emotion analysis were conducted with R max = 8000 Hz and M = 60.
As seen from Table 2, Table 3, Table 4 and Table 5:
(1)
The recognition accuracy of the Mel feature was higher than that of IMel, and the effect of using IMel alone was not as good as that of using Mel alone.
(2)
MelSSAE, IMelSSAE, and (Mel + IMel) achieved higher recognition accuracy than Mel, IMel, and Mel + IMel, respectively, thereby proving the effectiveness of the SSAE in dimension reduction.
(3)
The recognition accuracy of Mel + IMel was not necessarily higher than that of Mel or IMel, because directly splicing the two may contain redundancy and affect the recognition accuracy. However, after SSAE optimization, (Mel + IMel) SSAE improved the recognition accuracy of Mel, IMel, and Mel + IMel, thereby proving that Mel and IMel are complementary in deep features.
(4)
The standard deviation value is the average value that describes the distance that each recognition result deviates from the average. The values in the SI environment (Table 2) were less than the values in the SD and GD environments (Table 3, Table 4 and Table 5). This is because SI took all samples of one subject as the test set and other people’s samples as the training set, to reflect the similarity between individuals and groups. For example, GD-male took samples of all male subjects as the training set and samples of one female subject as the test set. To discover the similarity between different genders, there were large differences between the different genders, so the standard deviation of the recognition results was greater in the SI environment.
Although the experimental standards are different, the experimental results of SI, SD, and GD show the similar trends mentioned above; the results therefore prove the general applicability of this algorithm in this paper.
(2) Recognition accuracy of various emotions
Table 6 presents a comparison between the emotion recognition rate of the Mel feature and that of the IMel feature when = 8000 Hz and M = 60 in the three datasets. Figure 11 is a line chart corresponding to the recognition accuracy shown in Table 6. It can be observed that for high-arousal emotions, such as anger and happiness, the recognition accuracy of the Mel features was higher than that of the IMel features. For anger, the recognition accuracies of the Mel features were 94.83%, 80.17%, and 80.82%, respectively, and those of the IMel features were 88.71%, 71.67%, and 68.79%, respectively. These results indicate that the Mel spectrogram is more suitable for high-arousal emotions than the IMel spectrogram. For low-arousal emotions, such as sadness and disgust, the recognition accuracy of the IMel features was higher than that of the Mel features. For sadness, the recognition accuracies of the Mel features were 75.87%, 70.54%, and 66.87%, respectively, and those of the IMel features were 84.68%, 79.35%, and 70.96%, respectively. These results indicate that the IMel spectrogram is more suitable for low-arousal and high-frequency emotions than the Mel spectrogram.
(3) ROC curves
In Figure 12, the horizontal coordinate is the false Positive Rate, and the vertical coordinate is the ture positive Rate. AUC refers to the area under the ROC curve. Figure 12a presents the ROC curves of the system for ‘anger’, ‘anxiety’, ‘boredom’, ’disgust’, ‘happiness’, ‘neutral’, ‘sadness’ obtained by our proposed approach to the DBM-DB dataset, the red dashed line is the mean ROC value of each emotion. As can be seen from the figure, the system of this paper has good performance.
(4) The performance of the proposed system
Although the algorithm proposed in this paper has two CNN channels to extract the deep features, and doubles the number of parameters in terms of complexity, we can build two projects in Pycharm for Mel spectrogram and IMel spectrogram at the same time, and therefore there is no change in running time. The dimension of the single channel is 128, that of double channel is 256, and whether the input is 128 or 256 dimensions has little effect on the running time of SSAE.
(5) Comparison of various papers
Table 7 lists the results of the comparison between this paper and related papers published in recent years. The listed papers all use spectrograms to detect emotion. It can be observed that our method of combining the Mel spectrogram and IMel spectrogram is better than methods that use only a Mel spectrogram or a standard spectrogram.

7. Conclusions

In this paper, the frequency domain of IMel was proposed, and an IMel spectrogram that can highlight the high-frequency part was formed, which makes up for the shortage of Mel spectrograms that only highlight the low-frequency part. Then, a CNN-SSAE deep optimization network was presented. In this network, the two-dimensional spectrogram was sent to the CNN to obtain one-dimensional deep features, and then the internal redundant features were compressed and optimized by SSAE. The EMO-DB, SAVEE, and RAVDESS speech databases were used for experiments to verify the complementarity of the Mel spectrogram and IMel spectrogram and the effectiveness of CNN-SSAE in optimizing deep features. According to the current research situation, Although the proposed results obtained better recognition results, However, it also has limitations. The IMel filter is only the inverse process of the Mel filter, highlighting the high-frequency part. In the next step, wavelet algorithm, Hilbert–Huang transform, etc., can be used to highlight different frequency components of the layered signal; therefore, the next step will be to realize new spectrograms to obtain better recognition results. At present, the research tends to be multi-modal, combining the results of spectrogram with EEG features or kinematic features. In addition, the next step will be to extract better features through a deep optimized network, and obtain better recognition results.

Author Contributions

Conceptualization, J.L. and X.Z.; methodology, X.Z.; software, L.H.; validation, X.Z., F.L. and Y.S.; formal analysis, J.L.; investigation, S.D.; resources, X.Z. and S.D.; data curation, Y.S. and F.L.; writing—original draft preparation, J.L. and X.Z.; writing—review and editing, J.L., L.H. and S.D.; visualization, Y.S.; supervision, F.L.; project administration, J.L.; funding acquisition, L.H. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Nature Science Foundation of China under Grant 62271342, in part by “Project 1331” Quality Enhancement and Efficiency Construction Plan National First-class Major Construction Project of Electronic Science and Technology, in part by the National Natural Science Foundation of China Youth Science Foundation under Grant 12004275, in part by the Natural Science Foundation of Shanxi Province, China, under Grant 201901D111096, in part by Research Project Supported by Shanxi Scholarship Council of China HGKY2019025.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Yildirim, S.; Kaya, Y.; Kili, F. A modified feature selection method based on metaheuristic algorithms for speech emotion recognition. Appl. Acoust. 2021, 173, 107721. [Google Scholar] [CrossRef]
  2. Fahad, M.S.; Ashish, R.; Jainath, Y.; Akshay, D. A survey of speech emotion recognition in natural environment science direct. Digit. Signal Process. 2021, 110, 102951. [Google Scholar] [CrossRef]
  3. Wang, S.H.; Phillips, P.; Dong, Z.C.; Zhang, Y.D. Intelligent facial emotion recognition based on stationary wavelet entropy and Jaya algorithm. Neurocomputing 2018, 272, 668–676. [Google Scholar] [CrossRef]
  4. Gunes, H.; Piccardi, M. Bi-modal emotion recognition from expressive face and body gestures. J. Netw. Comput. Appl. 2007, 30, 1334–1345. [Google Scholar] [CrossRef]
  5. Noroozi, F.; Corneanu, C.A.; Kaminska, D.; Sapinski, T.; Escalera, S.; Anbarjafari, G. Survey on emotional body gesture recognition. IEEE Trans. Affect. Comput. 2021, 12, 505–523. [Google Scholar] [CrossRef]
  6. Islam, M.R.; Moni, M.A.; Islam, M.M.; Rashed-Al-Mahfuz, M.; Islam, M.S.; Hasan, M.K.; Hossain, S.; Ahmad, M.; Uddin, S.; Azad, A.; et al. Emotion recognition from EEG signal focusing on deep learning and shallow learning techniques. IEEE Access 2021, 9, 94601–94624. [Google Scholar] [CrossRef]
  7. Abbaschian, B.J.; Sierra-Sosa, D.; Elmaghraby, A. Deep learning techniques for speech emotion recognition from databases to models. Sensors 2021, 21, 1249. [Google Scholar] [CrossRef] [PubMed]
  8. Zhang, H.; Huang, H.; Han, H. A novel heterogeneous parallel convolution bi-LSTM for speech emotion recognition. Appl. Sci. 2021, 11, 9897. [Google Scholar] [CrossRef]
  9. El Ayadi, M.; Kamel, M.S.; Karray, F. Survey on speech emotion recognition: Features, classification schemes, and databases. Pattern Recognit. 2011, 44, 572–587. [Google Scholar] [CrossRef]
  10. Akçay, M.B.; Oğuz, K. Speech emotion recognition: Emotional models, databases, features, preprocessing methods, supporting modalities, and classifiers. Speech Commun. 2020, 116, 56–76. [Google Scholar] [CrossRef]
  11. Cheng, L.; Zong, Y.; Zheng, W.; Li, Y.; Tang, C.; Schuller, B.W. Domain Invariant Feature Learning for Speaker-Independent Speech Emotion Recognition. IEEE/ACM Trans. Audio Speech Lang. Process. 2022, 30, 2217–2230. [Google Scholar]
  12. Ozer, I. Pseudo-colored rate map representation for speech emotion recognition. Biomed. Signal Process. Control 2021, 66, 102502. [Google Scholar] [CrossRef]
  13. Prasomphan, S. Detecting human emotion via speech recognition by using speech spectrogram. In Proceedings of the 2015 IEEE International Conference on Data Science and Advanced Analytics (DSAA), Paris, France, 19–21 October 2015; pp. 1–10. [Google Scholar]
  14. Jiang, P.; Fu, H.; Tao, H.; Lei, P.; Zhao, L. Parallelized convolutional recurrent neural network with spectral features for speech emotion recognition. IEEE Access 2019, 7, 90368–90377. [Google Scholar] [CrossRef]
  15. Farooq, M.; Hussain, F.; Baloch, N.K.; Raja, F.R.; Yu, H.; Zikria, Y.B. Impact of feature selection algorithm on speech emotion recognition using deep convolutional neural network. Sensors 2020, 20, 6008. [Google Scholar] [CrossRef]
  16. Khalil, R.A.; Jones, E.; Babar, M.I.; Jan, T.; Zafar, M.H.; Alhussain, T. Speech emotion recognition using deep learning techniques: A review. IEEE Access 2019, 7, 117327–117345. [Google Scholar] [CrossRef]
  17. Chen, L.F.; Su, W.J.; Feng, Y.; Wu, M.; Hirota, K. Two-layer fuzzy multiple random forest for speech emotion recognition in human-robot interaction. Inf. Sci. 2019, 509, 150–163. [Google Scholar] [CrossRef]
  18. Zhang, S.Q.; Zhang, S.L.; Huang, T.J.; Huang, T.J.; Gao, W. Speech Emotion Recognition Using Deep Convolutional Neural Network and Discriminant Temporal Pyramid Matching. IEEE Trans. Multimed. 2017, 20, 1576–1590. [Google Scholar] [CrossRef]
  19. Lieskovská, E.; Jakubec, M.; Jarina, R.; Chmulík, M. A review on speech emotion recognition using deep learning and attention mechanism. Electronics 2021, 10, 1163. [Google Scholar] [CrossRef]
  20. Sugan, N.; Srinivas, N.S.S.; Kumar, L.S.; Nath, M.K.; Kanhe, A. Speech emotion recognition using cepstral features extracted with novel triangular filter banks based on bark and erb frequency scales. Biomed. Signal Process. Control 2020, 104, 102763. [Google Scholar]
  21. Zheng, C.S.; Tan, Z.H.; Peng, R.H.; Li, X.D. Guided spectrogram filtering for speech dereverberation. Appl. Acoust. 2018, 134, 154–159. [Google Scholar] [CrossRef]
  22. Liu, Z.T.; Xie, Q.; Wu, M.; Cao, W.H.; Mei, Y.; Mao, J.W. Speech emotion recognition based on formant characteristics feature extraction and phoneme type convergence. Inf. Sci. 2021, 563, 309–325. [Google Scholar] [CrossRef]
  23. Satt, A.; Rozenberg, S.; Hoory, R. Efficient Emotion Recognition from Speech Using Deep Learning on Spectrograms. In Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH, Stockholm, Sweden, 20–24 August 2017. [Google Scholar]
  24. Yao, Z.W.; Wang, Z.H.; Liu, W.H.; Liu, Y.Q. Speech emotion recognition using fusion of three multi-task learning-based classifiers: HSF-DNN, MS-CNN and LLD-RNN. Speech Commun. 2020, 120, 11–19. [Google Scholar] [CrossRef]
  25. Daneshfar, F.; Kabudian, S.J. Speech emotion recognition using discriminative dimension reduction by employing a modified quantumbehaved particle swarm optimization algorithm. Multimed. Tools Appl. 2020, 79, 1261–1289. [Google Scholar] [CrossRef]
  26. Yuan, J.; Chen, L.; Fan, T.; Jia, J. Dimension reduction of speech emotion feature based on weighted linear discriminate analysis. Image Process. Pattern Recognit. 2015, 8, 299–308. [Google Scholar]
  27. Sahu, S.; Gupta, R.; Sivaraman, G.; AbdAlmageed, W.; Espy-Wilson, C. Adversarial auto-encoders for speech based emotion recognition. arXiv 2018, arXiv:1806.02146. [Google Scholar]
  28. Mao, Q.; Dong, M.; Huang, Z.; Zhan, Y. Learning salient features for speech emotion recognition using convolutional neural networks. IEEE Trans. Multimedia. 2014, 16, 2203–2213. [Google Scholar] [CrossRef]
  29. Ancilin, J.; Milton, A. Improved speech emotion recognition with Mel frequency magnitude coefficient. Appl. Acoust. 2021, 179, 108046. [Google Scholar] [CrossRef]
  30. Nwe, T.L.; Foo, S.W.; Silva, L.C.D. Speech emotion recognition using hidden markov models. Speech Commun. 2003, 41, 603–623. [Google Scholar] [CrossRef]
  31. Diana, T.B.; Meshia, C.O.; Wang, F.N.; Jiang, D.M.; Verhelst, W.; Sahli, H. Hierarchical sparse coding framework for speech emotion recognition. Speech Commun. 2018, 99, 80–89. [Google Scholar]
  32. Kerkeni, L.; Serrestou, Y.; Raoof, K.; Mbarki, M.; Mahjoub, M.A.; Cleder, C. Automatic speech emotion recognition using an optimal combination of features based on EMD-TKEO. Speech Commun. 2019, 114, 22–35. [Google Scholar] [CrossRef]
  33. Sun, Y.; Zhang, X.Y. Characteristics of human auditory model based on compensation of glottal features in speech emotion recognition. Future Gener. Comput. Syst. 2018, 81, 291–296. [Google Scholar]
  34. Yang, B.; Lugger, M. Emotion recognition from speech signals using new harmony features. Signal Process. 2010, 99, 1415–1423. [Google Scholar] [CrossRef]
  35. Sun, Y.X.; Wen, G.H.; Wang, J.B. Weighted spectral features based on local Hu moments for speech emotion recognition. Biomed. Signal Process. Control 2015, 18, 80–90. [Google Scholar] [CrossRef]
  36. Badshah, A.; Rahim, N.; Ullah, N.; Ahmad, J. Deep features based speech emotion recognition for smart affective services. Multimed. Tools Appl. 2019, 78, 5571–5589. [Google Scholar] [CrossRef]
  37. Anvarjon, T.; Mustaqeem; Kwon, S. Deep-Net: A Lightweight CNN-Based Speech Emotion Recognition System Using Deep Frequency Features. Sensors 2020, 20, 5212. [Google Scholar] [CrossRef]
  38. Minji, S.; Myungho, K. Fusing Visual Attention CNN and Bag of Visual Words for Cross-Corpus Speech Emotion Recognition. Sensors 2020, 20, 5559. [Google Scholar]
  39. Chen, M.Y.; He, X.J.; Yang, J.; Zhang, H. 3-D Convolutional Recurrent Neural Networks with Attention Model for Speech Emotion Recognition. IEEE Signal Process. Lett. 2018, 25, 1440–1444. [Google Scholar] [CrossRef]
  40. Liu, D.; Shen, T.S.; Luo, Z.L.; Zhao, D.X.; Guo, S.J. Underwater target recognition using convolutional recurrent neural networks with 3-D Mel-spectrogram and data augmentation. Appl. Acoust. 2021, 178, 107989. [Google Scholar] [CrossRef]
  41. Hajarolasvadi, N.; Demirel, H. 3D CNN-Based Speech Emotion Recognition Using K-Means Clustering and Spectrograms. Entropy 2019, 21, 479. [Google Scholar] [CrossRef]
  42. Zhang, X.; Song, P.; Cha, C. Time frequency atomic auditory attention model for cross database speech emotion recognition. J. Southeast Univ. 2016, 4, 11–16. [Google Scholar]
  43. Yu, Y.; Kim, Y. Attention-LSTM-Attention Model for Speech Emotion Recognition and Analysis of IEMOCAP Database. Electronics 2020, 9, 713. [Google Scholar] [CrossRef]
  44. Eyben, F.; Wöllmer, M.; Schuller, B. Opensmile: The munich versatile and fast open-source audio feature extractor. In Proceedings of the 18th ACM international conference on Multimedia, Firenze, Italy, 25–29 October 2010; pp. 1459–1462. [Google Scholar]
  45. Ozseven, T. Investigation of the effect of spectrogram images and different texture analysis methods on speech emotion recognition. Appl. Acoust. 2018, 142, 70–77. [Google Scholar] [CrossRef]
  46. Liu, Z.T.; Xie, Q.; Wu, M.; Cao, W.H.; Mei, Y.; Mao, J.W. Speech emotion recognition based on an improved brain emotion learning model. Neurocomputing 2018, 309, 145–156. [Google Scholar] [CrossRef]
  47. Yogesh, C.K.; Hariharanb, M.; Ngadirana, R.; Adomb, A.H.; Yaacobc, S.; Berkai, C.; Polat, K. A new hybrid pso assisted biogeography-based optimization for emotion and stress recognition from speech signal. Expert Syst. Appl. 2017, 69, 149–158. [Google Scholar]
  48. Daneshfar, F.; Kabudian, S.J.; Neekabadi, A. Speech emotion recognition using hybrid spectral-prosodic features of speech signal/glottal waveform, metaheuristic-based dimensionality reduction, and Gaussian elliptical basis function network classifier. Appl. Acoust. 2020, 166, 107360. [Google Scholar] [CrossRef]
  49. Xu, J.; Xiang, L.; Liu, Q.S.; Gilmore, H.; Wu, J.Z.; Tang, J.H.; Madabhushi, A. Stacked Sparse Autoencoder (SSAE) for Nuclei Detection on Breast Cancer Histopathology Images. IEEE Trans. Med. Imaging 2015, 35, 119–130. [Google Scholar] [CrossRef]
  50. Tang, Q.L.; Fang, Q.; Xia, X.F.; Yang, J.R. Breast pathology image cell identification based on stacked sparse autoencoder and holistically-nested structure. J. South-Cent. Univ. Natl. Nat. Sci. Ed. 2019, 3, 397–403. [Google Scholar]
  51. Mufidah, R.; Wasito, I.; Hanifah, N.; Faturrahman, M.; Ghaisani, F.D. Automatic nucleus detection of pap smear images using stacked sparse autoencoder (ssae). In Proceedings of the International Conference on Algorithms Computing and Systems, Jeju Island Republic of Korea, 10–13 August 2017; pp. 9–13. [Google Scholar]
  52. Li, S.Q.; Jiang, H.Y.; Bai, J.; Liu, Y.; Yao, Y.D. Stacked sparse autoencoder and case-based postprocessing method for nucleus detection. Neurocomputing 2019, 24, 494–508. [Google Scholar] [CrossRef]
  53. Quan, X.L.; Zeng, Z.G.; Jiang, J.H.; Zhang, Y.Q.; LV, B.L.; Wu, D.R. Physiological signals based affective computing: A systematic review. Acta Autom. Sin. 2021, 8, 1769–1784. [Google Scholar]
  54. Burkhardt, F.; Paeschke, A.; Rolfes, M.; Sendlmeier, W.F. A database of german emotional speech; INTERSPEECH 2005—Eurospeech. In Proceedings of the 9th European Conference on Speech Communication and Technology, Lisbon, Portugal, 4–8 September 2005. [Google Scholar]
  55. Jackson, P.J.B.; Haq, S.U. Surrey Audio-Visual Expressed Emotion (Savee) Database; University of Surrey: Guildford, UK, 2014. [Google Scholar]
  56. Livingstone, S.R.; Russo, F.A. The ryerson audio-visual database of emotional speech and song (ravdess): A dynamic, multimodal set of facial and vocal expressions in north American English. PLoS ONE 2018, 13, e0196391. [Google Scholar] [CrossRef]
  57. Yogesh, C.K.; Hariharan, M.; Ngadiran, R.; Adom, A.H.; Yaacob, S.; Polat, K. Hybrid bbo pso and higher order spectral features for emotion and stress recognition from natural speech. Appl. Soft Comput. 2017, 56, 217–232. [Google Scholar]
  58. Wang, K.X.; Su, G.X.; Liu, L.; Wang, S. Wavelet packet analysis for speaker-independent emotion recognition. Neurocomputing 2020, 398, 257–264. [Google Scholar] [CrossRef]
Figure 1. Recognition network. The first channel extracts optimized features from the Mel spectrogram to highlight the low-frequency information, and the other channel extracts optimized features from the IMel spectrogram to highlight high-frequency information.
Figure 1. Recognition network. The first channel extracts optimized features from the Mel spectrogram to highlight the low-frequency information, and the other channel extracts optimized features from the IMel spectrogram to highlight high-frequency information.
Applsci 12 09518 g001
Figure 2. (a) The relationship between flin and fMel. (b) Mel filter banks.
Figure 2. (a) The relationship between flin and fMel. (b) Mel filter banks.
Applsci 12 09518 g002
Figure 3. Russell’s two-dimensional emotion spatial representation.
Figure 3. Russell’s two-dimensional emotion spatial representation.
Applsci 12 09518 g003
Figure 4. (a) The relationship between f l i n and f I M e l . (b) The IMel filter banks.
Figure 4. (a) The relationship between f l i n and f I M e l . (b) The IMel filter banks.
Applsci 12 09518 g004
Figure 5. (a) Speech signal of anger. (b) Speech signal of sad. (c) Mel spectrogram of anger. (d) Mel.spectrogram of sad. (e) IMel spectrogram of anger. (f) IMel spectrogram of sad.
Figure 5. (a) Speech signal of anger. (b) Speech signal of sad. (c) Mel spectrogram of anger. (d) Mel.spectrogram of sad. (e) IMel spectrogram of anger. (f) IMel spectrogram of sad.
Applsci 12 09518 g005
Figure 6. The generation of Mel spectrogram and IMel spectrogram.
Figure 6. The generation of Mel spectrogram and IMel spectrogram.
Applsci 12 09518 g006
Figure 7. The structure of the CNN.
Figure 7. The structure of the CNN.
Applsci 12 09518 g007
Figure 8. The structure of the SSAE.
Figure 8. The structure of the SSAE.
Applsci 12 09518 g008
Figure 9. (a) Mel filter banks M = 20. (b) Mel spectrogram M = 20. (c) Mel filter banks M = 40. (d) Mel.spectrogram M = 40.
Figure 9. (a) Mel filter banks M = 20. (b) Mel spectrogram M = 20. (c) Mel filter banks M = 40. (d) Mel.spectrogram M = 40.
Applsci 12 09518 g009
Figure 10. (a) Mel filter banks R: 0–4000. (b) Mel spectrogram R: 0–4000. (c) IMel filter banks R:0–4000. (d) IMel spectrogram R: 0–4000. (e) Mel filter banks R: 0–8000. (f) Mel spectrogram R: 0–8000. (g) IMel filter banks R: 0–8000. (h) IMel spectrogram R: 0–8000. (i) Mel filter banks R: 0–21,000. (j) Mel spectrogram R: 0–21,000. (k) IMel filter banks R: 0–21,000. (l) IMel spectrogram R: 0–21,000.
Figure 10. (a) Mel filter banks R: 0–4000. (b) Mel spectrogram R: 0–4000. (c) IMel filter banks R:0–4000. (d) IMel spectrogram R: 0–4000. (e) Mel filter banks R: 0–8000. (f) Mel spectrogram R: 0–8000. (g) IMel filter banks R: 0–8000. (h) IMel spectrogram R: 0–8000. (i) Mel filter banks R: 0–21,000. (j) Mel spectrogram R: 0–21,000. (k) IMel filter banks R: 0–21,000. (l) IMel spectrogram R: 0–21,000.
Applsci 12 09518 g010
Figure 11. (a) Correlation coefficient diagram of the Mel deep features. (b) Correlation coefficient diagram of the IMel deep features. (c) Correlation coefficient diagram of the Mel + IMel deep features.
Figure 11. (a) Correlation coefficient diagram of the Mel deep features. (b) Correlation coefficient diagram of the IMel deep features. (c) Correlation coefficient diagram of the Mel + IMel deep features.
Applsci 12 09518 g011
Figure 12. (a) ROC curve of the system for EMO-DB. (b) ROC curve of the system for SAVEE. (c) ROC curve of the system for RAVDESS.
Figure 12. (a) ROC curve of the system for EMO-DB. (b) ROC curve of the system for SAVEE. (c) ROC curve of the system for RAVDESS.
Applsci 12 09518 g012
Table 1. CNN network architecture and parameter.
Table 1. CNN network architecture and parameter.
LayersPaddingKernel SizeStridesOutput Shape
Input3-channel color picture[80,504,3]
Convolutional layer Con1SAME[1,6,3,32][1,1][80,504,32]
Pooling layer Pool1SAME[1,4][1,4][80,126,32]
Convolutional layer Con2SAME[4,1,32,64][1,1][80,126,64]
Pooling layer Pool2SAME[2,1][2,1][40,126,64]
Convolutional layer Con3SAME[3,3,64,96][1,1][40,126,96]
Convolutional layer Con4VALID[40,2,96,128][1,1][1,125,128]
Pooling layer Pool3Reshape the tensor into a single dimension[125,128]
GAPGlobal Average Pooling[128]
Table 2. SI Recognition Accuracy ± Standard Deviations (%).
Table 2. SI Recognition Accuracy ± Standard Deviations (%).
Database R max MEMelIMelMel + IMelMelSSAEIMelSSAE(Mel + IMel) SSAE
EMO-DB400040WA83.14 ± 2.5878.41 ± 4.1984.14 ± 2.5885.42 ± 2.2086.46 ± 2.6390.63 ± 1.86
UA80.21 ± 2.9674.53 ± 5.3080.74 ± 2.2187.50 ± 2.6285.29 ± 3.0189.34 ± 1.90
60WA77.06 ± 2.6174.53 ± 4.1981.36 ± 2.5887.50 ± 2.2082.29 ± 2.6389.58 ± 1.86
UA75.36 ± 3.5672.33 ± 4.8180.14 ± 2.6886.12 ± 2.8983.89 ± 2.8788.03 ± 1.56
80WA85.77 ± 5.4780.59 ± 3.1887.94 ± 3.4490.63 ± 1.2183.33 ± 3.4191.67 ± 2.68
UA84.90 ± 3.3879.02 ± 3.2181.03 ± 4.2486.81 ± 1.1981.78 ± 2.7787.89 ± 2.37
800040WA83.09 ± 3.4575.54 ± 2.2984.92 ± 1.7887.14 ± 2.3083.33 ± 1.8990.63 ± 2.73
UA81.49 ± 3.7374.03 ± 3.0883.79 ± 1.8484.31 ± 2.1282.51 ± 1.5389.09 ± 2.16
60WA89.49 ± 1.7879.54 ± 2.2988.07 ± 2.6692.78 ± 3.2487.58 ± 2.3394.79 ± 1.77
UA88.50 ± 1.9874.03 ± 3.0884.53 ± 2.6691.09 ± 2.9886.47 ± 2.6793.32 ± 1.98
80WA85.85 ± 1.8781.61 ± 5.5284.31 ± 2.3685.54 ± 2.6684.38 ± 3.2293.75 ± 2.30
UA83.63 ± 2.7779.78 ± 4.1983.58 ± 2.2386.41 ± 3.1283.60 ± 3.0592.68 ± 2.45
SAVEE400040WA81.25 ± 4.1768.96 ± 3.7180.63 ± 2.4083.08 ± 2.5479.56 ± 3.6084.41 ± 2.98
UA79.05 ± 4.9065.71 ± 3.9278.21 ± 2.4482.54 ± 2.7878.11 ± 3.2983.72 ± 2.46
60WA74.79 ± 5.6369.59 ± 0.0579.79 ± 2.4078.82 ± 3.2479.03 ± 1.9283.66 ± 2.14
UA73.45 ± 5.9965.28 ± 0.0677.50 ± 2.3277.21 ± 2.5578.56 ± 2.0182.23 ± 2.46
80WA78.13 ± 3.6177.08 ± 2.8579.17 ± 4.3684.46 ± 3.4181.34 ± 3.7487.72 ± 2.61
UA75.60 ± 4.0174.64 ± 2.7576.90 ± 4.7583.02 ± 3.5680.20 ± 2.6686.51 ± 2.96
800040WA72.48 ± 2.5571.79 ± 3.5675.25 ± 3.6178.04 ± 3.4172.83 ± 3.7685.85 ± 3.51
UA74.17 ± 2.6163.45 ± 4.4778.97 ± 3.8376.98 ± 3.2571.56 ± 3.0884.41 ± 3.12
60WA75.61 ± 3.9473.30 ± 4.1277.48 ± 4.8583.25 ± 3.8981.17 ± 3.6388.96 ± 3.56
UA67.14 ± 4.0762.08 ± 3.6071.58 ± 4.8082.11 ± 3.2580.01 ± 3.7087.54 ± 3.38
80WA73.71 ± 2.5572.09 ± 3.9475.35 ± 4.9281.48 ± 2.9881.58 ± 3.5587.67 ± 3.10
UA65.12 ± 2.4462.98 ± 4.5075.00 ± 4.0780.30 ± 3.0280.24 ± 3.1585.68 ± 2.94
RAVDESS400040WA73.93 ± 1.9264.85 ± 1.5575.71 ± 2.5475.98 ± 2.3870.74 ± 2.6279.62 ± 1.75
UA72.11 ± 1.8562.59 ± 1.6074.65 ± 2.2174.85 ± 2.6169.55 ± 2.1578.54 ± 1.66
60WA76.29 ± 1.1965.55 ± 3.5378.72 ± 2.8977.57 ± 3.2573.83 ± 3.5481.31 ± 2.38
UA75.22 ± 1.2964.53 ± 4.4877.58 ± 2.4576.30 ± 3.1272.63 ± 3.1980.47 ± 2.15
80WA72.94 ± 1.5666.81 ± 3.5775.32 ± 3.1078.50 ± 0.8472.89 ± 1.7282.24 ± 2.37
UA70.64 ± 2.1765.67 ± 3.9969.89 ± 0.5377.67 ± 0.9871.16 ± 1.3881.28 ± 2.63
800040WA71.31 ± 2.6968.05 ± 4.0173.42 ± 3.0276.13 ± 3.2471.34 ± 2.7480.70 ± 1.91
UA70.88 ± 2.6365.27 ± 4.2072.02 ± 2.2175.69 ± 3.0470.29 ± 2.3579.65 ± 1.84
60WA74.01 ± 2.2271.16 ± 1.5777.01 ± 3.1080.24 ± 2.5881.17 ± 3.6383.18 ± 2.31
UA67.14 ± 4.0762.08 ± 3.6071.58 ± 4.8082.11 ± 3.2575.70 ± 1.9883.01 ± 3.01
80WA71.77 ± 2.7370.53 ± 2.9174.16 ± 1.6179.37 ± 1.8978.50 ± 2.3482.24 ± 2.27
UA70.83 ± 3.4969.36 ± 3.4972.96 ± 1.5378.56 ± 1.6577.68 ± 2.0781.66 ± 2.39
R-the frequency range, M-the number of filters, E-Evaluation indexes.
Table 3. SD Recognition Accuracy ± Standard Deviations (%).
Table 3. SD Recognition Accuracy ± Standard Deviations (%).
DatabaseEMelIMelMel + IMelMelSSAEIMelSSAE(Mel + IMel) SSAE
EMO-DBWA83.53 ± 5.4267.63 ± 8.8680.78 ± 5.1789.10 ± 4.6474.97 ± 6.8292.52 ± 3.90
UA83.60 ± 5.8666.16 ± 9.7279.06 ± 5.5587.24 ± 6.7272.31 ± 8.2491.39 ± 3.80
SAVEEWA68.17 ± 7.8054.88 ± 5.5468.38 ± 9.8774.42 ± 10.5961.53 ± 3.8076.08 ± 7.56
UA66.35 ± 8.9051.71 ± 7.0165.76 ± 10.5671.68 ± 8.6157.51 ± 4.5573.11 ± 8.89
RAVDESSWA68.83 ± 7.8153.35 ± 10.7567.91 ± 8.6073.33 ± 5.4556.72 ± 8.1574.86 ± 5.79
UA67.00 ± 8.6752.96 ± 9.9565.70 ± 8.40570.81 ± 7.7755.09 ± 8.7471.55 ± 5.65
E-Evaluation indexes.
Table 4. GD-male Recognition Accuracy ± Standard Deviations (%).
Table 4. GD-male Recognition Accuracy ± Standard Deviations (%).
DatabaseEMelIMelMel + IMelMelSSAEIMelSSAE(Mel + IMel) SSAE
EMO-DBWA77.02 ± 3.9259.77 ± 11.2277.94 ± 4.9072.44 ± 8.7769.76 ± 9.6872.37 ± 7.02
UA77.41 ± 6.0560.02 ± 13.1277.16 ± 7.5471.95 ± 10.1568.80 ± 10.8769.21 ± 8.66
RAVDESSWA65.77 ± 9.6349.97 ± 8.7665.20 ± 7.9270.05 ± 5.6855.05 ± 6.1473.64 ± 6.00
UA66.23 ± 10.5351.97 ± 10.7766.23 ± 8.4466.16 ± 5.3352.56 ± 5.8269.10 ± 5.53
E-Evaluation indexes.
Table 5. GD-female Recognition Accuracy ± Standard Deviations (%).
Table 5. GD-female Recognition Accuracy ± Standard Deviations (%).
DatabaseEMelIMelMel + IMelMelSSAEIMelSSAE(Mel + IMel) SSAE
EMO-DBWA80.77 ± 7.4563.53 ± 9.5078.49 ± 9.8684.61 ± 7.2976.63 ± 8.1784.68 ± 5.43
UA80.49 ± 4.5061.60 ± 10.0474.65 ± 4.8882.14 ± 6.3073.31 ± 7.6781.42 ± 4.45
RAVDESSWA68.06 ± 7.2448.20 ± 7.1987.36 ± 7.7071.53 ± 6.3752.50 ± 7.2671.94± 8.22
UA66.93 ± 7.8847.92 ± 6.9766.53 ± 7.9067.71 ± 6.6149.74 ± 7.2667.71 ± 7.58
E-Evaluation indexes.
Table 6. Recognition Accuracy of Various Emotions (%).
Table 6. Recognition Accuracy of Various Emotions (%).
EmotionFeaturesAngerHappinessNeutralDisgustSadnessAnxietyFearfulSurpriseBoredomCalmAverage
EMO-DBMel94.8378.7189.9668.5275.8779.56--88.16-82.23
IMel88.7158.6782.2976.5684.6880.50--81.56-79.56
SAVEEMel80.1776.9276.6268.0870.54-78.6778.33--75.61
IMel71.6769.6771.3273.4579.35-75.9771.67--73.30
RAVDESSMel80.8277.8771.0071.0366.87-78.7668.14-77.5874.01
IMel68.7971.2367.4775.5870.96-69.6774.93-70.6471.16
Table 7. Comparison of various papers in recent years.
Table 7. Comparison of various papers in recent years.
DatabaseLiteratureYearsSpectrogramAccuracy (%)
EMO-DBTursunov et al. [8]2020Standard spectrogram92.02
Minji et al. [9]2020Mel spectrogram86.92
Zhang et al. [12]2016Standard spectrogram88.40
Our Method-Mel and IMel spectrogram94.79
SAVEEMinji et al. [9]2020Mel spectrogram75.00
Noushin et al. [11]2019Mel spectrogram81.05
Ozseven et al. [20]2018Standard spectrogram75.90
Our Method-Mel and IMel spectrogram88.96
RAVDESSZeng et al. [21]2019Standard spectrogram64.48
Jalal et al. [22]2019Mel spectrogram68.10
Mustaqeem et al. [23]2019Standard spectrogram79.50
Our Method-Mel and IMel spectrogram83.18
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Li, J.; Zhang, X.; Huang, L.; Li, F.; Duan, S.; Sun, Y. Speech Emotion Recognition Using a Dual-Channel Complementary Spectrogram and the CNN-SSAE Neutral Network. Appl. Sci. 2022, 12, 9518. https://doi.org/10.3390/app12199518

AMA Style

Li J, Zhang X, Huang L, Li F, Duan S, Sun Y. Speech Emotion Recognition Using a Dual-Channel Complementary Spectrogram and the CNN-SSAE Neutral Network. Applied Sciences. 2022; 12(19):9518. https://doi.org/10.3390/app12199518

Chicago/Turabian Style

Li, Juan, Xueying Zhang, Lixia Huang, Fenglian Li, Shufei Duan, and Ying Sun. 2022. "Speech Emotion Recognition Using a Dual-Channel Complementary Spectrogram and the CNN-SSAE Neutral Network" Applied Sciences 12, no. 19: 9518. https://doi.org/10.3390/app12199518

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop