Next Article in Journal
Panic Attack, Severe Hypophosphatemia and Rhabdomyolysis in the Setting of a Motor Functional Neurological Disorder
Previous Article in Journal
Sensor-Based Rehabilitation in Neurological Diseases: A Bibliometric Analysis of Research Trends
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Semi-Supervised Speech Deception Detection Algorithm Combining Acoustic Statistical Features and Time-Frequency Two-Dimensional Features

1
Key Laboratory of Food Information Processing and Control, Ministry of Education, Henan University of Technology, Zhengzhou 450001, China
2
Henan Engineering Laboratory of Grain IOT Technology, Henan University of Technology, Zhengzhou 450001, China
*
Author to whom correspondence should be addressed.
Brain Sci. 2023, 13(5), 725; https://doi.org/10.3390/brainsci13050725
Submission received: 9 March 2023 / Revised: 23 April 2023 / Accepted: 24 April 2023 / Published: 26 April 2023
(This article belongs to the Special Issue Neural Network for Speech and Gesture Semantics)

Abstract

:
Human lying is influenced by cognitive neural mechanisms in the brain, and conducting research on lie detection in speech can help to reveal the cognitive mechanisms of the human brain. Inappropriate deception detection features can easily lead to dimension disaster and make the generalization ability of the widely used semi-supervised speech deception detection model worse. Because of this, this paper proposes a semi-supervised speech deception detection algorithm combining acoustic statistical features and time-frequency two-dimensional features. Firstly, a hybrid semi-supervised neural network based on a semi-supervised autoencoder network (AE) and a mean-teacher network is established. Secondly, the static artificial statistical features are input into the semi-supervised AE to extract more robust advanced features, and the three-dimensional (3D) mel-spectrum features are input into the mean-teacher network to obtain features rich in time-frequency two-dimensional information. Finally, a consistency regularization method is introduced after feature fusion, effectively reducing the occurrence of over-fitting and improving the generalization ability of the model. This paper carries out experiments on the self-built corpus for deception detection. The experimental results show that the highest recognition accuracy of the algorithm proposed in this paper is 68.62% which is 1.2% higher than the baseline system and effectively improves the detection accuracy.

1. Introduction

Lying is a common phenomenon in life [1] and is considered to be a high-level executive control behavior that causes activity in the amygdala, insula, and prefrontal regions of the brain [2], which in turn leads to changes in speech parameters such as frequency during speaking. There has been an attempt to improve the recognition rate of lie detection with advanced techniques. Deception detection techniques have been applied to criminal investigations [3], psychotherapy [4,5], children’s education [6], and national security [7] with some success. Traditional deception detection methods require contact with the human body, which may bring psychological burdens and interfere with the results of deception detection [8]. Aldert [9] also pointed out that the application of medical devices to collect physiological and brain signals may make these signals invasive and inconvenient to use, while speech signals can produce better results. Compared with traditional deception detection methods, speech deception detection methods have the advantages of easy access to data, absence of time and space constraints, and high concealment. Therefore, deception detection using speech has a strong theoretical and practical restudy value for the study of cognitive brain science [10].
Early relevant studies have confirmed that some acoustic features in speech are related to deception [11]. Ekman et al. [12] collected and analyzed the subjects’ impressions of some TV clips, and found that the fundamental frequency part of lies was higher than that of truth. Lying and stress are always related. Kaliappan and Hansen et al. [13,14] found that some acoustic parameters related to lying, such as resonance peak, Bark energy characteristics, and MFCC, changed with the alteration of pressure level. DePaulo et al. [15] meta-analyzed 158 features proposed by previous polygraph research work and selected 23 speech and speech-related features with significant expressions. The study found that lies showed less detailed expressions, repetitive utterances, more content, shorter expression lengths, and incoherent speech utterances compared to the truth. The research team of Purdue University in the United States used the amplitude modulation model and frequency modulation model to conduct speech deception detection research, proving that Teager energy-related features could distinguish truth from lie [16]. In addition, some relevant scholars considered combining multiple features for deception detection. Researchers at Columbia University considered combining acoustic features, prosodic features, and lexical features for research on lie detection in speech [17]. In 2013, Kirchhuebel et al. used the acoustic and temporal features of speech to study the effects of different conversation modes on deception detection from three aspects: emotional arousal/stress [18], cognitive load, and ultra control. Some scholars classify acoustic features into prosody features and spectral-based correlation analysis features [19]. Speech prosody refers to the vocal modulations that accompany speech and comprises variations in fundamental frequency, duration, and energy. In recent years, speech prosody has been recognized in several disciplines, including psycholinguistics, as a bridge between speech acts and mental disorders [20], and therefore has great research value in revealing the brain mechanisms behind speech communication. Spectral-based features can reflect the connection between speech tract shape and speech behavior [21]. The cochlea of the human ear is the key to forming hearing, which can convert speech signals into neural pulses and send them to the auditory area of the brain, generating hearing. The basilar membrane of the cochlea is equivalent to a nonlinear filter bank, and its movement frequencies are converted into nerve impulses by outer hair cells and inner hair cells. The mel-frequency cepstrum coefficient (MFCC) [22] is a feature parameter discovered based on this auditory mechanism, which is in nonlinear correspondence with frequency and has been widely used in the fields of speech emotion recognition and deception detection. Research has shown that early extraction and analysis of acoustic parameters affect the differentiation of early ERP responses, while stimuli caused by acoustic characteristics in the early stages can affect brain cognition in the later stages [23]. The nervous system encodes these evolving acoustic parameters to obtain a clear representation of different speech patterns, further enabling the brain to clearly distinguish between lies and truth. With the development of deep learning technology, researchers extracted deep features through deep neural networks and applied them to speech deception detection research. Xie et al. [24] combined spectral features that exploit the orthogonality and translational invariance of Hu moments with deep learning methods and used deep confidence networks for their experiments, achieving extremely high recognition results. Liang et al. [25] extracted speech-depth features using convolutional long and short-time memory networks, and achieved good recognition results on the self-built deception detection database.
Although the above scholars have made many achievements in the field of deception detection, the data-driven deep neural network is extremely dependent on large-scale labeled high-quality speech data, and the problem of insufficient data has become a key problem restricting the development of the field of voice-based lie detection [26]. The supervised model is the most common machine learning model, which is widely used in the field of speech deception detection and has achieved high recognition accuracy. When the amount of labeled data is insufficient, the improvement in lie detection accuracy by supervised models can appear to be inadequate. Unsupervised models, which can discover the intrinsic structure of data and are often used for data mining, may be particularly useful in cases where labeled data is not available or when there is a need to identify new patterns in speech. Due to the limitation of data volume, the application of unsupervised models in the field of lie detection in speech has yet to be further investigated. Semi-supervised learning is a learning method that combines supervised and unsupervised learning. Semi-supervised models learn the local features of a small amount of labeled data and the overall distribution of a large amount of unlabeled data to obtain acceptable or even better recognition results. Semi-supervised models offer a promising approach for lie detection in speech and other tasks. Tarvainen et al. [27] proposed a method of averaging the weights of the mean-teacher model, combined with the consistent regularization method, by adding perturbed data points to push the decision boundary to the appropriate location, improving the generalization of the model, and significantly improving the learning speed and classification accuracy of the network. Liu et al. [28] added a pseudo-label generation module under the framework of the classic domain confrontation network and reduced the impact of pseudo-label noise and the error rate of prediction results by introducing the mean-teacher model. In the field of speech deception detection, Fu H et al. [29] proposed a speech deception detection model based on a semi-supervised denoising autoencoder network (DAE), which achieved good results using only a small amount of labeled data. Due to the limitations of traditional acoustic features, the trained network representation ability is insufficient, and it is difficult to achieve high recognition accuracy. Su et al. [30] trained the BILSTM network and SVM models separately and further fused the classification results using a decision-level score fusion scheme to integrate all developed models. Fang et al. [31] proposed a speech deception detection strategy combining the semi-supervised method and the full-supervised method, and constructed a hybrid model combining semi-supervised DAE and fully supervised LSTM network, effectively improving the accuracy of semi-supervised speech deception detection. Although the above research has made some achievements, it ignores the exploration of the multifeature deception detection algorithm under a fully semi-supervised framework. Improper fusion of features can easily lead to poor generalization ability of the semi-supervised model.
Inspired by the feature fusion method and semi-supervised learning, this paper proposes a semi-supervised speech deception detection algorithm that integrates acoustic statistical features and time-frequency two-dimensional features to solve the problems in the research of speech deception detection, aiming to suppress the dimension disaster caused by multiclass feature fusion and obtain features with more favorable information in the semi-supervised learning environment. Firstly, the proposed algorithm employs a hybrid network composed of a semi-supervised AE network and a mean-teacher model network to extract the fusion features of deception detection, with the aid of the mean-teacher model to extract spectral features rich in time-frequency information, and applies a semi-supervised AE network to extract low-dimensional, high-level acoustic statistical features. Secondly, the consistency regularization method is introduced, and the dropout method is added to improve the generalization ability of the model and suppress the over-fitting phenomenon. Finally, the fusion features are input into the softmax classifier for classification, and the model is optimized by using a dynamically adjusted weighted sum of the cross-entropy loss of labeled data, the consistency regularization of unlabeled data, and the reconstruction loss of the AE network.

2. Materials and Methods

2.1. System Model

The proposed model framework is shown in Figure 1. The model applied the semi-supervised AE to obtain the depth acoustic statistical features, used the semi-supervised mean-teacher model based on the CNN network to extract the depth time-frequency two-dimensional features, and then employed the consistent regularization methods to constrain the fusion of the output features of the two semi-supervised networks to suppress the model over-fitting. Each module is described as follows.

2.2. Speech Feature Extraction

2.2.1. Three-Dimensional Mel-Spectrum Feature

Lying can cause time-frequency changes in speech, and the mel spectrum has been proven to be rich in time-frequency features [32,33]. In this paper, 64 sets of filters, a 25 ms Hamming window, and a 10 ms overlap were used to obtain the features of the mel spectrum. By calculating the first-order difference and the second-order difference of the mel spectrum feature, and further supplementing the time-frequency information, the 3D mel-spectrum feature [34] was obtained, and its composition is shown in Figure 2. This article resized the 3D mel-spectrum feature to 256 × 256 × 3 as the input of the mean-teacher model, denoted by X C N N .

2.2.2. Acoustic Statistical Characteristics

Choosing the proper artificial statistical features is important for the effective learning of the model. Therefore, we employed the feature set [35] specified in the 2009 Sentiment Recognition Challenge in this paper. The feature set uses 16 low-level descriptors as well as 12 statistical functions, as shown in Table 1. The 16 low-level descriptors are, respectively, zero-crossing-rate (ZCR) from the time signal, root mean square (RMS) frame energy, pitch frequency (normalized to 500 Hz), harmonics-to-noise ratio (HNR) by autocorrelation function, and mel-frequency cepstral coefficients (MFCC) 1–12 in full accordance to HTK-based computation. To ensure the reproducibility of the experiments, we used opensmile software to extract these features from the feature set in speech. By calculating the low-level descriptors’ first-order difference and their 12 statistical functions, each speech could obtain 16 × 2 × 12 = 384 dimensional features as the input of the semi-supervised AE network, recorded as X A E .

2.3. Semi-Supervised Hybrid Network Model

In this paper, we collected some labeled data D L = { X L , Y L } N L , and a large number of unlabeled data D U = { X U } N U , where Y L was the label, i = 1 , 2 , , N , the total number of samples was N = N L + N U , N L , and N U , respectively, representing the total number of labeled data and unlabeled data. In this paper, the labeled data was recorded as X L = { x A E L , x C N N L } , and the unlabeled data was recorded as X U = { x A E U , x C N N U } . Among them, x A E L , x A E U were labeled input and unlabeled input of the semi-supervised AE network, with X A E = { x A E L , x A E U } ; x C N N L , x C N N U were labeled input and unlabeled input of mean-teacher model, respectively, with X C N N = { x C N N L , x C N N U } .

2.3.1. Mean-Teacher Model

As described in Refs. [36,37], the mean-teacher model has achieved excellent recognition performance in the case of insufficient numbers of labels. The specific network structure is shown in Figure 3. The mean-teacher model consists of a student network and a teacher network, which are composed of the same convolutional neural network. Their structure is shown in Table 2. To increase the amount of data that could be processed, the weight relationship between each feature factor and the corresponding category in the speech data was established at a deeper level. In this paper, the spectral features were horizontally flipped and randomly clipped (the processing was recorded as η ), and on that basis, unlabeled data was processed with noise enhancement (the processing was recorded as η ), and the processed features were used as the input of the mean-teacher model. The output of the network is shown as follows:
Y S t u L = f C N N ( x C N N L , Y L , η ; θ S t u ) Y S t u U = f C N N ( x C N N U , η , η ; θ S t u ) Y T e a U = f C N N ( x C N N U , η ; θ T e a )
where f C N N refers to the process of feature extraction of the convolutional network. Y S t u L refers to the output of labeled spectral data in the student network after processing η ; Y S t u U is the output of unlabeled spectral data on the student network after processing η and η . Y T e a U is the output of unlabeled spectral data on the teacher network after processing η .

2.3.2. Semi-Supervised AE Network

Refs. [38,39] show that the AE network can remove redundancy between features, fully dig the deep information between features, and obtain low-dimensional and high-level features. Therefore, a semi-supervised AE network was built in this paper and its parameter was θ A E . The network structure is shown in Figure 4, with its parameters can be seen in Table 3, and the extraction process is described as follows:
Y A E L = f A E ( x A E L , Y L ; θ A E ) Y A E N U = f A E ( x A E U , η ; θ A E ) Y A E U = f A E ( x A E U ; θ A E )
where f A E refers to the process of extracting features from the AE network, η refers to the noise enhancement processing of unlabeled features. Y A E L refers to the output of labeled acoustic statistics in the AE network, Y A E N U refers to the output of unlabeled acoustic statistics in the AE network after processing η , and Y A E U refers to the output of unlabeled acoustic statistics in the AE network.
Reconstruction loss is the key to unsupervised learning by AE networks. By reducing the reconstruction error, it helps the models to extract high-level, high-quality features. The reconstruction loss of the AE network is shown as follows:
L r e c o n = 1 M i = 1 M ( [ f R A E ( Y i A E ) · log ( Y i A E ) + ( 1 f R A E ( Y i A E ) ) log ( 1 Y i A E ) ] )
where M is the input number of each batch, f R A E refers to the process of reconstructing features from the AE network, Y i A E refers to the output of all features extracted by the AE network, and f R d a e ( Y i A E ) refers to the output obtained from manual statistical features after extraction and reconstruction by AE networks.

2.3.3. The AE–MT Model with Consistent Regularization

The semi-supervised AE–MT model constructed in this paper comprises two parts: the AE–Stu model and the AE–Tea model. The parameters of the AE–Stu model are expressed by θ = { θ A E , θ C N N } and the parameters of the AE–Tea model are expressed by θ = { θ A E , θ C N N } . The AE–MT model flowchart is shown in Figure 5.
Taking advantage of the characteristic that the feature spaces of artificial statistical features and depth features are different, the AE–MT model fuses the two types of features to obtain high-level features with a strong ability of representation, including unlabeled enhanced output Y A E _ S t u N U = { Y A E N U , Y S t u N U } , unlabeled output Y A E _ T e a U = { Y A E U , Y T e a U } , and labeled output Y A E _ S t u L = { Y A E L , Y S t u L } .
The supervised loss of the system model was represented by the cross-entropy loss function between the labeled output and the real label belonging to the AE–Stu model. Its equation is as follows:
L c e = 1 M i = 1 M [ Y i L · log ( Y i A E _ S t u L ) + ( 1 Y i L ) · log ( 1 Y i A E _ S t u L ) ]
The feature fusion method integrates different types of features to make up for the differences between features but also leads to the problem of dimension disaster, which will cause over-fitting and affect the classification ability of the model. Refs. [40,41,42] show that the consistency regularization method can utilize the potential information of unlabeled data and improve the generalization ability of the semi-supervised model. Meanwhile, to solve the problem that the perturbed fusion features are easily misclassified by the model, this algorithm introduces a consistency regularization method. As shown in Figure 6, although the original decision boundary is not enough to distinguish the perturbed features, the perturbed features are also correctly classified after optimization by the consistent regularization method.
In this paper, the loss of consistency regularization is defined as the expected distance between the unlabeled predicted output Y A E _ S t u N U of the AE–Stu model and the unlabeled predicted output Y A E _ T e a U of the AE–Tea model.
L c o n s i s = 1 M i = 1 M [ | | Y i A E _ S t u N U Y i A E _ T e a U | | 2 ]

2.3.4. Multiloss Model Optimization Model

In this paper, we optimized the AE–MT model using multiclass loss to improve its classification performance. The weighted sum of three types of losses was taken as the total loss of the AE–MT model, as shown below.
L = L c e + ω L c o n s i s + a L r e c o n
where a is the weighting factor of the reconstruction loss function, and ω is the dynamically adjustable weighting factor of the sum of the consistent regularization loss and the reconstruction loss, which is generally constant.
The total loss was back propagated, the parameters θ of the AE–Stu model were optimized and updated, and the parameters θ of the AE–Stu model were updated to those of the AE–Tea model through the exponential moving average method. The processing is shown below:
θ = α e m a θ + 1 α e m a θ
where α e m a represents the smoothing coefficient in the exponential moving average method, and θ represents the updated parameters of the teacher network.

3. Experiment and Result Analysis

3.1. Dataset

In order to complement the existing corpus of deception detection and verify the effectiveness of the AE–MT model proposed in this paper, we built the H-Wolf corpus for speech deception detection experiments by referring to the construction process of the Idiap Wolf database [43], and the Killer database [29]. We collected about 70 h of the “Werewolves of Miller’s Hollow” competition video, which can be found on the internet, and screened video clips containing truth and lies according to the ID card of players and competition rules in each werewolf killing competition. Firstly, we used free audio and video editing tools, as shown in Figure 7, to intercept video clips related to truth or lie. We then used Adobe Audition, an audio processing software, to separate the audio from the video and display the audio waveform and spectrogram, which can be seen in Figure 8. Finally, we imported the audio clip using Adobe Audition software and modified its sample rate to 48,000 Hz, changing its bit depth to 16 bits. The number of players in each “Werewolves of Miller’s Hollow” competition is 12, and one player can participate in multiple competitions. After screening and statistics, the detailed number of participants is shown in Table 4. After multiple people tests, we retained clear and recognizable speech extracts, obtaining 1103 speech extracts (521 deception speech extracts). Then we divide them by 9:1 to obtain 992 training data and 111 test data.

3.2. Experimental Configuration

In order to fully learn the feature information of the unlabeled data and reduce the impact of noise on the recognition results. we set η to 0.3 for enhancing unlabeled data. We used the small batch random gradient descent algorithm (SGD) for model training and set the learning rate to 0.0003. In this paper, the weight factor in Equation (4) is 0.5, and the cosine annealing attenuation method was used to adjust the learning rate so that the learning rate changed with the cycle.
We used accuracy and f1_score as classification evaluation criteria, accuracy as the main evaluation criterion for ablation experiments, and f1_score was used to further evaluate the performance of each module when the number of labels was 600. Their calculation method is shown in Formulas (8) and (9).
a c c u r a c y = n c o r r e c t N t o t a l
f 1 _ s c o r e = 2 × T P 2 × T P + F P + F N
n c o r r e c t represents the number of correctly predicted samples, N t o t a l represents the total number of samples. T P represents the positive samples with correct prediction, T N represents the negative samples with correct prediction, F P represents positive samples with prediction errors, F N represents negative samples with prediction errors.
To prevent data from over-fitting during training, we added dropout to the AE–MT model and set it to 0.8. All the experiments were carried out in the RTX 3080 and the 3.8 version of the python environment.

3.3. Results

3.3.1. Ablation Study

To verify the classification performance of the fusion features of the proposed semi-supervised model compared with the single feature, we removed the mean-teacher network (base), the AE network, as well as the consistency regularization algorithm(CR), respectively, and then conducted speech deception detection experiments when the number of labeled data was set to 200, 400, 600, with other parameters unchanged.
After a maximum of 100 epochs of iterative training, the experimental results are shown in Table 5 and Table 6.
As shown in Table 5, the accuracy of the mean-teacher model reached 61.35, 63.14, and 67.47% when the number of labeled data was 200, 400, and 600. From the results, the mean-teacher model made use of the potential information of unlabeled data and improved the accuracy of classification. The accuracy of the semi-supervised AE network in this paper achieved 63.0, 63.9, and 66.71% when the number of labeled data was 200, 400, and 600. It is noteworthy that the semi-supervised AE network attained the highest accuracy when the number of labeled data was 200. Because the AE network [44] is good at processing unsupervised data, even though the number of labeled data is fewer, the AE network can give better performance. However, when the number of labeled data is increased, the improvement in the semi-supervised AE networks’ performance is less than in the other model. According to the results of the AE + MT model in Table 5 and Table 6, it can be seen that simply combining the two models leads to dimensional disaster, causing an overfitting of the complex hybrid model, and reducing its classification performance. However, it should be noted that the AE + MT model combined with the consistent regularization method had better recognition performance than the other models. It was proved that the consistent regularization method could effectively solve the problem of overfitting of the model and improve the classification performance of the hybrid model. At the same time, the f1_score of the proposed model was also higher than the other models, confirming the effectiveness of the proposed method.

3.3.2. Comparison to Other Algorithms

In addition, we compared the proposed algorithm with other semi-supervised methods. The comparison algorithm includes the semi-supervised AE model used in reference and the semi-supervised LSTM model using a pseudo-labeling algorithm. The differences between our algorithm and other algorithms are shown in Table 7 and Table 8.
As shown in Table 7, when the number of labels is 200, compared with SS-AE, and SS-LSTM, the values of the proposed method are improved by 8.87 and 7.92%. When the number of labels was 400, compared with SS-AE, and SS-LSTM, the accuracy of the proposed method increased by 10.51 and 6.77%. Compared with SS-AE and SS-LSTM, the accuracy of the proposed method increased by 13.24 and 10.7%, when the number of labels was 600. When the number of labels was different, the proposed algorithm always performed better than the other algorithms. As shown in Table 8, the f1_score of the proposed algorithm was much higher than the other algorithms, which further proves that the classification performance of the proposed model is much better than other models.

3.3.3. Confusion Matrix

To further study the recognition accuracy, we introduced the confusion matrix to analyze the model. On the H-wolf dataset, we made the confusion matrix as shown in Figure 9. When the number of labeled data was 200, the recognition rate of truth and lies was 74 and 55%, respectively; when the number of labeled data was 400, the recognition rate of truth and deception was 64 and 61%, respectively; when the number of tags was 600, the recognition rate of truth and lies was 65 and 79%, respectively. Of these, the recognition rate of the truth was higher than 64%, and the accuracy rate of the lies was higher than 55%. With the increase in the number of labeled data, the recognition ability of the model to identify the lies was significantly improved.

4. Discussion

When people lie, they tend to use more complex language and take longer to respond to questions. This process is accompanied by changes in ERPs on the amygdala, insula, and prefrontal regions of the brain as well as changes in acoustic signature parameters associated with lying, with some studies demonstrating that these two changes are correlated [23]. Drawing on the work of Low et al. [47] and Pastoriza-Domínguez et al. [48] who used machine learning algorithms based on acoustic feature analysis for detecting major mental disorders, we focused, in this paper, on choosing the acoustic feature parameters associated with the act of lying and used the trained neural network model to detect subtle changes in the acoustic feature parameters under different speech patterns to discriminate between lies and truth. This can help us better understand how speech is processed in the brain and enable researchers to further investigate the brain’s cognitive neural mechanisms during the lying process. Our models can also be modified and applied to the assessment and diagnosis of speech prosody in mental disorders, in terms of the automatic classification of prosodic events for detection.
Due to the specific nature of the act of lying, it is difficult to insulate subjects from the effects of the equipment when collecting EEG signals and facial information related to lies, which can lead to biases between the data collected and the actual data. Moreover, in many cases, it is only after the act of lying has occurred that people’s brains become aware of the lie. As mentioned in the literature [49], the choice may have taken place before it was actually made. However, using speech signals alone for deception detection is not comprehensive; in some cases, EEG signals and facial information are more directly indicative of the true situation. Therefore, conducting multimodal lie detection research is meaningful [50], as it can comprehensively explore the neural mechanisms of the lying process from multiple perspectives.

5. Conclusions

In this work, we proposed a research framework for semi-supervised speech spoofing detection based on acoustic statistical features and time-frequency two-dimensional features. Unlike previous studies of semi-supervised speech spoofing detection algorithms based on a single feature and a single model, our proposed AE–MT model consists of two parallel components, namely, an AE network and an average teacher model, which deal with acoustic statistical features and time-frequency two-dimensional features, respectively. It is worth noting that applying feature fusion methods to the features extracted from the two networks can lead to high-level features with better representation. However, the feature fusion approach also increases the dimensionality, thus triggering a dimensionality catastrophe and exacerbating the overfitting of the model. Therefore, consistent regularization and dropout were also introduced in this paper to effectively improve the generalization ability of the model. Experiments showed that the AE–MT model could effectively mine feature information with good performance.

Author Contributions

Conceptualization, H.F. and H.Y.; data curation, H.F. and H.Y.; formal analysis, H.F., X.W. and C.Z.; funding acquisition, H.F.; investigation, H.Y., X.W. and X.L.; methodology, H.F. and H.Y.; project administration, H.F.; resources, H.F.; software, H.Y.; validation, H.Y.; supervision, H.F.; visualization, H.Y.; writing—original draft preparation, H.F. and H.Y.; writing—review and editing, H.F. and H.Y. All authors have read and agreed to the published version of the manuscript.

Funding

The research was supported by Henan Province Key Scientific Research Projects Plan of Colleges and Universities (Grant No.22A510001). The research was supported by Henan Province Key Scientific Research Projects Plan of Colleges and Universities (Grant No.22A520004).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Cole, T. Lying to the one you love: The use of deception in romantic relationships. J. Soc. Pers. Relatsh. 2001, 18, 107–129. [Google Scholar] [CrossRef]
  2. Christ, S.E.; Van Essen, D.C.; Watson, J.M.; Brubaker, L.E.; McDermott, K.B. The contributions of prefrontal cortex and executive control to deception: Evidence from activation likelihood estimate meta-analyses. Cereb. Cortex 2009, 19, 1557–1566. [Google Scholar] [CrossRef] [PubMed]
  3. Vrij, A.; Fisher, R.P. Which lie detection tools are ready for use in the criminal justice system? J. Appl. Res. Mem. Cogn. 2016, 5, 302–307. [Google Scholar] [CrossRef]
  4. Lykken, D.T. Psychology and the lie detector industry. Am. Psychol. 1974, 29, 725. [Google Scholar] [CrossRef] [PubMed]
  5. Levine, T.R. Truth-default theory and the psychology of lying and deception detection. Curr. Opin. Psychol. 2022, 47, 101380. [Google Scholar] [CrossRef] [PubMed]
  6. Gongola, J.; Scurich, N.; Quas, J.A. Detecting deception in children: A meta-analysis. Law Hum. Behav. 2017, 41, 44. [Google Scholar] [CrossRef] [PubMed]
  7. Rogers, R.; Boals, A.; Drogin, E.Y. Applying cognitive models of deception to national security investigations: Considerations of psychological research, law, and ethical practice. J. Psychiatry Law 2011, 39, 339–364. [Google Scholar] [CrossRef]
  8. Meng Luning, Z.Z. Multi-parameter psychological testing polygraph and application. People’s Procur. Semimon. 2000, 7, 56–58. [Google Scholar]
  9. Vrij, A.; Granhag, P.A.; Ashkenazi, T.; Ganis, G.; Leal, S.; Fisher, R.P. Verbal Lie Detection: Its Past, Present and Future. Brain Sci. 2022, 12, 1644. [Google Scholar] [CrossRef]
  10. Zhao, L.; Liang, R.; Xie, Y.; Zhuang, D. Progress and Outlook of Lie Detection Technique in Speech. J. Data Acquis. Process. 2017, 2, 246–257. [Google Scholar]
  11. Kirchhuebel, C. The Acoustic and Temporal Characteristics of Deceptive Speech; University of York: York, UK, 2013. [Google Scholar]
  12. Ekman, P.; O′Sullivan, M.; Friesen, W.V.; Scherer, K.R. Invited article: Face, voice, and body in detecting deceit. J. Nonverbal Behav. 1991, 15, 125–135. [Google Scholar] [CrossRef]
  13. Hansen, J.H.; Womack, B.D. Feature analysis and neural network-based classification of speech under stress. IEEE Trans. Speech Audio Process. 1996, 4, 307–313. [Google Scholar] [CrossRef]
  14. Kirchhübel, C.; Howard, D.M.; Stedmon, A.W. Acoustic correlates of speech when under stress: Research, methods and future directions. Int. J. Speech Lang. Law 2011, 18, 75–98. [Google Scholar] [CrossRef]
  15. Muhlenbruck, L.; Lindsay, J.G.; Malone, I.B.; Depaulo, B.M.; Charlton, K.; Cooper, C. Cues to deception. Psychol. Bull. 2002, 129, 74–118. [Google Scholar]
  16. Gopalan, K.; Wenndt, S. Speech analysis using modulation-based features for detecting deception. In Proceedings of the 2007 15th International Conference on Digital Signal Processing, Cardiff, UK, 1–4 July 2007; pp. 619–622. [Google Scholar]
  17. Enos, F.; Shriberg, E.; Graciarena, M.; Hirschberg, J.B.; Stolcke, A. Detecting deception using critical segments. In Proceedings of the 8th Annual Conference of the International Speech Communication Association, Antwerp, Belgium, 27–31 August 2007. [Google Scholar]
  18. Khawaja, M.A.; Chen, F.; Marcus, N. Measuring cognitive load using linguistic features: Implications for usability evaluation and adaptive interaction design. Int. J. Hum.-Comput. Interact. 2014, 30, 343–368. [Google Scholar] [CrossRef]
  19. Liu, Z.-T.; Xu, J.-P.; Wu, M.; Cao, W.-H.; Chen, L.-F.; Ding, X.-W.; Hao, M.; Xie, Q. Review of Emotional Feature Extraction and Dimension Reduction Method for Speech Emotion Recognition. Jisuanji Xuebao Chin. J. Comput. 2018, 41, 2833–2851. [Google Scholar] [CrossRef]
  20. Ding, H.; Zhang, Y. Speech prosody in mental disorders. Annu. Rev. Linguist. 2023, 9, 335–355. [Google Scholar] [CrossRef]
  21. Benesty, J.; Sondhi, M.M.; Huang, Y. Springer Handbook of Speech Processing; Springer: Berlin, Germany, 2008; Volume 1. [Google Scholar]
  22. Logan, B. Mel Frequency Cepstral Coefficients for Music Modeling. In Proceedings of the International Society for Music Information Retrieval Conference, Plymouth, MA, USA, 23–25 October 2000. [Google Scholar]
  23. Jiang, X.; Pell, M.D. On how the brain decodes vocal cues about speaker confidence. Cortex 2015, 66, 9–34. [Google Scholar] [CrossRef]
  24. Xie, Y.; Liang, R.; Bao, Y. Deception detection with spectral features based on deep belief network. Acta Acust. 2019, 2, 214–220. [Google Scholar]
  25. Xie, Y.; Liang, R.; Tao, H.; Zhu, Y.; Zhao, L. Convolutional bidirectional long short-term memory for deception detection with acoustic features. IEEE Access 2018, 6, 76527–76534. [Google Scholar] [CrossRef]
  26. Hirschberg, J.B.; Benus, S.; Brenier, J.M.; Enos, F.; Friedman, S.; Gilman, S.; Girand, C.; Graciarena, M.; Kathol, A.; Michaelis, L. Distinguishing deceptive from non-deceptive speech. In Proceedings of the 9th European Conference on Speech Communication and Technology, Lisbon, Portugal, 4–8 September 2005. [Google Scholar]
  27. Tarvainen, A.; Valpola, H. Mean teachers are better role models: Weight-averaged consistency targets improve semi-supervised deep learning results. Adv. Neural Inf. Process. Syst. 2017, 30, 1195–1204. [Google Scholar]
  28. Liu, J.-M.; Sun, H.; Lin, L. Pseudo-label Based Defensible Stable Network. Comput. Technol. Dev. 2022, 6, 34–38. [Google Scholar]
  29. Fu, H.; Lei, P.; Tao, H.; Zhao, L.; Yang, J. Improved semi-supervised autoencoder for deception detection. PLoS ONE 2019, 14, e0223361. [Google Scholar] [CrossRef] [PubMed]
  30. Su, B.-H.; Yeh, S.-L.; Ko, M.-Y.; Chen, H.-Y.; Zhong, S.-C.; Li, J.-L.; Lee, C.-C. Self-Assessed Affect Recognition Using Fusion of Attentional BLSTM and Static Acoustic Features. In Proceedings of the Interspeech, Hyderabad, India, 2–6 September 2018; pp. 536–540. [Google Scholar]
  31. Fang, Y.; Fu, H.; Tao, H.; Liang, R.; Zhao, L. A novel hybrid network model based on attentional multi-feature fusion for deception detection. IEICE Trans. Fundam. Electron. Commun. Comput. Sci. 2021, 104, 622–626. [Google Scholar] [CrossRef]
  32. Mao, Q.; Dong, M.; Huang, Z.; Zhan, Y. Learning salient features for speech emotion recognition using convolutional neural networks. IEEE Trans. Multimed. 2014, 16, 2203–2213. [Google Scholar] [CrossRef]
  33. Sugahara, R.; Osawa, M.; Sato, R. Ensemble of simple resnets with various mel-spectrum time-frequency resolutions for acoustic scene classifications. In Detection and Classification of Acoustic Scenes and Events Challenge; Rion Co., Ltd.: Tokyo, Japan, 2021. [Google Scholar]
  34. Zhu, M.; Jiang, P.; Li, Z. Speech emotion recognition based on full convolution recurrent neural network. Tech. Acoust. 2021, 5, 645–651. [Google Scholar]
  35. Schuller, B.; Steidl, S.; Batliner, A. The Interspeech 2009 Emotion Challenge; ISCA: Brighton, UK, 2009. [Google Scholar]
  36. Deng, J.; Li, W.; Chen, Y.; Duan, L. Unbiased mean teacher for cross-domain object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 4091–4101. [Google Scholar]
  37. Kim, J.-H.; Shim, H.-J.; Jung, J.-W.; Yu, H.-J. Learning metrics from mean teacher: A supervised learning method for improving the generalization of speaker verification system. arXiv 2021, arXiv:2104.06604. [Google Scholar] [CrossRef]
  38. Wang, W.; Huang, Y.; Wang, Y.; Wang, L. Generalized autoencoder: A neural network framework for dimensionality reduction. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Columbus, OH, USA, 23–28 June 2014; pp. 490–497. [Google Scholar]
  39. Yin, W.; Li, L.; Wu, F.-X. A semi-supervised autoencoder for autism disease diagnosis. Neurocomputing 2022, 483, 140–147. [Google Scholar] [CrossRef]
  40. Abuduweili, A.; Li, X.; Shi, H.; Xu, C.-Z.; Dou, D. Adaptive consistency regularization for semi-supervised transfer learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 6923–6932. [Google Scholar]
  41. Sohn, K.; Berthelot, D.; Li, C.L.; Zhang, Z.; Carlini, N.; Cubuk, E.D.; Kurakin, A.; Zhang, H.; Raffel, C. FixMatch: Simplifying Semi-Supervised Learning with Consistency and Confidence. Adv. Neural Inf. Process. Syst. 2020, 33, 596–608. [Google Scholar]
  42. Berthelot, D.; Carlini, N.; Goodfellow, I.; Papernot, N.; Oliver, A.; Raffel, C. MixMatch: A Holistic Approach to Semi-Supervised Learning. Adv. Neural Inf. Process. Syst. 2019, 32, 5049–5059. [Google Scholar]
  43. Hung, H.; Chittaranjan, G. The idiap wolf corpus: Exploring group behaviour in a competitive role-playing game. In Proceedings of the 18th ACM International Conference on Multimedia, Firenze, Italy, 25–29 October 2010; pp. 879–882. [Google Scholar]
  44. Baldi, P. Autoencoders, unsupervised learning, and deep architectures. In Proceedings of the ICML Workshop on Unsupervised and Transfer Learning, Bellevue, WA, USA, 2 July 2012; pp. 37–49. [Google Scholar]
  45. Deng, J.; Xu, X.; Zhang, Z.; Frühholz, S.; Schuller, B. Semisupervised autoencoders for speech emotion recognition. IEEE ACM Trans. Audio Speech Lang. Process. 2017, 26, 31–43. [Google Scholar] [CrossRef]
  46. Lee, D.-H. Pseudo-label: The simple and efficient semi-supervised learning method for deep neural networks. In Proceedings of the Workshop on Challenges in Representation Learning, ICML, Atlanta, GA, USA, 16 June 2013; p. 896. [Google Scholar]
  47. Low, D.M.; Bentley, K.H.; Ghosh, S.S. Automated assessment of psychiatric disorders using speech: A systematic review. Laryngosc. Investig. Otolaryngol. 2020, 5, 96–116. [Google Scholar] [CrossRef] [PubMed]
  48. Pastoriza-Domínguez, P.; Torre, I.G.; Diéguez-Vide, F.; Gomez-Ruiz, I.; Gelado, S.; Bello-López, J.; Ávila-Rivera, A.; Matias-Guiu, J.A.; Pytel, V.; Hernández-Fernández, A. Speech pause distribution as an early marker for Alzheimer’s disease. Speech Commun. 2022, 136, 107–117. [Google Scholar] [CrossRef]
  49. Bear, A.; Bloom, P. A simple task uncovers a postdictive illusion of choice. Psychol. Sci. 2016, 27, 914–922. [Google Scholar] [CrossRef] [PubMed]
  50. Nasri, H.; Ouarda, W.; Alimi, A.M. ReLiDSS: Novel lie detection system from speech signal. In Proceedings of the 2016 IEEE/ACS 13th International Conference of Computer Systems and Applications (AICCSA), Agadir, Morocco, 29 November–2 December 2016; pp. 1–8. [Google Scholar]
Figure 1. System model framework.
Figure 1. System model framework.
Brainsci 13 00725 g001
Figure 2. Components of 3D mel-spectrum. (a) shows the mel-spectrum, (b) shows the first-order difference feature, and (c) shows the second-order difference feature.
Figure 2. Components of 3D mel-spectrum. (a) shows the mel-spectrum, (b) shows the first-order difference feature, and (c) shows the second-order difference feature.
Brainsci 13 00725 g002aBrainsci 13 00725 g002b
Figure 3. Structure of the mean-teacher model.
Figure 3. Structure of the mean-teacher model.
Brainsci 13 00725 g003
Figure 4. Structure of the AE model.
Figure 4. Structure of the AE model.
Brainsci 13 00725 g004
Figure 5. The flowchart of the AE–MT model with consistent regularization.
Figure 5. The flowchart of the AE–MT model with consistent regularization.
Brainsci 13 00725 g005
Figure 6. Consistency regularization optimization diagram. The circle (Ο) represents the truth feature, the triangle (Δ) represents the lie feature, yellow represents the undisturbed truth feature, orange represents the disturbed truth feature, light blue represents the undisturbed truth feature, and dark blue represents the disturbed lie feature.
Figure 6. Consistency regularization optimization diagram. The circle (Ο) represents the truth feature, the triangle (Δ) represents the lie feature, yellow represents the undisturbed truth feature, orange represents the disturbed truth feature, light blue represents the undisturbed truth feature, and dark blue represents the disturbed lie feature.
Brainsci 13 00725 g006
Figure 7. Interception of relevant video clips.
Figure 7. Interception of relevant video clips.
Brainsci 13 00725 g007
Figure 8. The extraction process of audio by Adobe Audition. (a) shows the separation of audio and video, and (b) shows the audio waveform and spectrogram.
Figure 8. The extraction process of audio by Adobe Audition. (a) shows the separation of audio and video, and (b) shows the audio waveform and spectrogram.
Brainsci 13 00725 g008
Figure 9. Confusion matrix of the different numbers of labels: (a) shows the confusion matrix of the model when the number of labels is 200, (b) shows the confusion matrix of the model when the number of labels is 400, and (c) shows the confusion matrix of the model when the number of labels is 600.
Figure 9. Confusion matrix of the different numbers of labels: (a) shows the confusion matrix of the model when the number of labels is 200, (b) shows the confusion matrix of the model when the number of labels is 400, and (c) shows the confusion matrix of the model when the number of labels is 600.
Brainsci 13 00725 g009aBrainsci 13 00725 g009b
Table 1. 2009 International Speech Emotion Recognition Challenge feature set.
Table 1. 2009 International Speech Emotion Recognition Challenge feature set.
NameSpecific Content
LLDs(16)RMS, F0, ZCR, HNR, MFCC(1–12)
Statistical functions(12)mean, stddev, kurtosis, skewness, max, min, maxposition, minposition, range, offset, slope, MSE
Table 2. The model parameters of student network and teacher network.
Table 2. The model parameters of student network and teacher network.
Network LayerOutput Shape
Input256 × 256 × 3
3 × (Conv1 + BN + Relu)256 × 256 × 32
Max-pooling128 × 128 × 32
3 × (Conv2 + BN + Relu)128 × 128 × 64
Max-pooling64 × 64 × 64
Conv3 + BN + Relu62 × 62 × 128
Conv4 + BN + Relu62 × 62 × 64
Conv5 + BN + Relu62 × 62 × 32
Sum and average32
Full connection8
Table 3. The model parameters of AE network.
Table 3. The model parameters of AE network.
NetworkNetwork LayerNumber of Nerve Units
Input384
Encode networkEncode layer1 + BN + Elu256
Encode layer2 + BN + Elu128
Encode layer3 + BN + Elu64
Encode layer4 + BN + Elu + dropout32
Decode networkDecode layer1 + BN + Elu128
Decode layer2 + BN + Elu180
Decode layer3 + BN + Elu256
Decode layer4 + BN + Elu + dropout384
Full connection(32, 8)
Table 4. Number of Contestants.
Table 4. Number of Contestants.
Game NameMaleFemaleAll
Werewolf281240
Table 5. The recognition accuracy of each model in the ablation experiment.
Table 5. The recognition accuracy of each model in the ablation experiment.
DatabaseModelLabeled Examples
200400600
H-wolfMT (base)61.35%63.14%67.47%
AE63.0%63.9%66.71%
AE + MT59.62%64.42%67.31%
AE + MT + CR (proposed)62.5%65.38%68.62%
Table 6. The f1_score (%) of each model in the ablation experiment when the number of labels is 600.
Table 6. The f1_score (%) of each model in the ablation experiment when the number of labels is 600.
IndicatorsMT (Base)AEAE + MTAE + MT + CR (Proposed)
f1_score66.4%65.3%65.8%69.6%
Table 7. Comparison of recognition accuracy (%) between the proposed algorithm and other algorithms.
Table 7. Comparison of recognition accuracy (%) between the proposed algorithm and other algorithms.
DatabaseModelLabeled Examples
200400600
H-wolfSS-AE [45]53.63%54.87%55.38%
SS-LSTM [46]54.58%58.61%57.92%
Proposed62.5%65.58%68.62%
Table 8. Comparison between f1_score (%) of the proposed algorithm and f1_score (%) of other algorithms when the number of labels is 600.
Table 8. Comparison between f1_score (%) of the proposed algorithm and f1_score (%) of other algorithms when the number of labels is 600.
IndicatorsSS-AESS-LSTMProposed
f1_score45.5%49.8%69.6%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Fu, H.; Yu, H.; Wang, X.; Lu, X.; Zhu, C. A Semi-Supervised Speech Deception Detection Algorithm Combining Acoustic Statistical Features and Time-Frequency Two-Dimensional Features. Brain Sci. 2023, 13, 725. https://doi.org/10.3390/brainsci13050725

AMA Style

Fu H, Yu H, Wang X, Lu X, Zhu C. A Semi-Supervised Speech Deception Detection Algorithm Combining Acoustic Statistical Features and Time-Frequency Two-Dimensional Features. Brain Sciences. 2023; 13(5):725. https://doi.org/10.3390/brainsci13050725

Chicago/Turabian Style

Fu, Hongliang, Hang Yu, Xuemei Wang, Xiangying Lu, and Chunhua Zhu. 2023. "A Semi-Supervised Speech Deception Detection Algorithm Combining Acoustic Statistical Features and Time-Frequency Two-Dimensional Features" Brain Sciences 13, no. 5: 725. https://doi.org/10.3390/brainsci13050725

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop