Next Article in Journal
Implementation of a Long Short-Term Memory Neural Network-Based Algorithm for Dynamic Obstacle Avoidance
Previous Article in Journal
Electric Susceptibility at Partial Coverage of a Circular One-Side Access Capacitive Sensor with Rigid Polyurethane Foams
Previous Article in Special Issue
Transformer Monitoring with Electromagnetic Energy Transmission and Wireless Sensing
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Flexible Self-Powered Low-Decibel Voice Recognition Mask

1
Department of Physics, College of Physical Science and Technology, Research Institution for Biomimetics and Soft Matter, Xiamen University, Xiamen 361005, China
2
Jiujiang Research Institute, Xiamen University, Jiujiang 332000, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(10), 3007; https://doi.org/10.3390/s24103007
Submission received: 26 March 2024 / Revised: 22 April 2024 / Accepted: 23 April 2024 / Published: 9 May 2024
(This article belongs to the Special Issue Advances in Flexible Self-Powered Electronics Sensors)

Abstract

:
In environments where silent communication is essential, such as libraries and conference rooms, the need for a discreet means of interaction is paramount. Here, we present a single-electrode, contact-separated triboelectric nanogenerator (CS-TENG) characterized by robust high-frequency sensing capabilities and long-term stability. Integrating this TENG onto the inner surface of a mask allows for the capture of conversational speech signals through airflow vibrations, generating a comprehensive dataset. Employing advanced signal processing techniques, including short-time Fourier transform (STFT), Mel-frequency cepstral coefficients (MFCC), and deep learning neural networks, facilitates the accurate identification of speaker content and verification of their identity. The accuracy rates for each category of vocabulary and identity recognition exceed 92% and 90%, respectively. This system represents a pivotal advancement in facilitating secure and efficient unobtrusive communication in quiet settings, with promising implications for smart home applications, virtual assistant technology, and potential deployment in security and confidentiality-sensitive contexts.

1. Introduction

Vibrations represent a pervasive phenomenon in a practical context. The triboelectric nanogenerator (TENG) provides a compelling method for generating electrical current through mechanical friction, presenting opportunities for charge transfer within materials without requiring additional power sources. Through the conversion of vibrations into electrical signals, the TENG has found diverse applications in large buildings, energy harvesting, mechanical equipment, health monitoring, and motion sensing [1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26]. As science and technology progress, the TENG’s role in signal sensing via vibration has expanded into speech recognition, prompting the need for improved material structure, response speed, sensitivity, noise immunity, and deployment contexts. Notably, its capacity to detect signals through vibration has significantly contributed to advancements in acoustic science [27,28,29,30,31,32,33,34,35,36,37]. For instance, Zheng’s team engineered a TENG capable of recognizing road and traffic sounds and deployed it on roads or buildings to differentiate between passing vehicles and structural flaws [38]. Similarly, Wang et al. coated silk protein onto the copper film of the device and combined it with a layer of FEP coated with silver, forming upper and lower layers. This setup allows for the detection of sounds within the frequency range of 20 to 2000 Hz and exhibits good noise resistance [39]. Furthermore, Xia et al. developed a TENG adept at detecting minute pressures for monitoring breath and glottal vibration articulation using graphene oxide as the raw material [40]. Wang’s team fabricated layered thermoplastic polyurethane (PU) and ionic electrode layers atop a polytetrafluoroethylene film, enabling the detection of minute vibrations from acoustic pressures and the identification of diverse human voice signals [41]. In recent years, with the development of artificial intelligence technology, there have been corresponding breakthroughs in TENG in the field of speech recognition. Table S1 lists the progress of TENG in speech recognition applications.
Most of these studies have made significant progress in voiced speech recognition while contributing less to silent speech recognition. Additionally, integrating voiceprint recognition into silent speech recognition designs is necessary for system security. The integration of voice communication with modern technologies such as smart homes and voice assistants is also a growing trend. Therefore, there is an urgent need to further explore the expansion of TENG in speech recognition. The system to facilitate communication in environments necessitates careful attention to materials, precise speech recognition technology, and compatibility across various life aspects. The characteristics of TENG, such as self-powering capabilities, high sensitivity, and efficiency, facilitate the applications of intelligent sensing technologies in our daily lives. For instance, integrating TENG into gesture recognition [42,43,44], intelligent IoT [45,46,47], wearable devices [48,49,50], automated monitoring [51,52,53], and speech recognition [39,54,55,56]. Despite significant advancements in speech recognition technology, challenges persist in certain specific application scenarios regarding accuracy and stability. For example, challenges include speech recognition in noisy environments and the collection and processing of long-duration speech signals. Friction-based nanogenerators may offer a novel approach to addressing these challenges, potentially providing new possibilities for self-powered systems, particularly in scenarios requiring long-term operation without easy battery replacement, such as remote monitoring devices or wireless sensor networks. Furthermore, integrating speech recognition technology with friction-based nanogenerators can enable more intelligent self-powered systems, such as utilizing ambient sound signals to activate or control devices, thereby achieving more intelligent and adaptive functionality. However, the combination of friction-based nanogenerators and speech recognition also faces several challenges. Firstly, the energy conversion efficiency of friction-based nanogenerators is relatively low and needs further improvement to meet practical application demands. Secondly, the application scenarios are limited, necessitating the expansion of the functionalities of friction-based nanogenerators and the design of more stable and diversified nanogenerator systems. Additionally, addressing integration and optimization issues between friction-based nanogenerators and speech recognition systems is necessary to achieve good coordination and collaboration between the two.
Overall, the combination of friction-based nanogenerators and speech recognition holds tremendous potential, offering new solutions and applications for self-powered systems, the Internet of Things (IoT), and beyond. However, it is currently in the research exploration stage, requiring further engineering and scientific research to address technical and application challenges and achieve widespread application in real-world scenarios.
Here, we developed a triboelectric nanogenerator (TENG) tailored for silent speech recognition, as depicted in Figure 1a. Our device design entailed engineering a single-electrode, contact-separated TENG (CS-TENG) by selecting the contact-separated operation mode from the four TENG modes [57]. The contact-separated structure allows for free movement of a charged electrode, where periodic changes in the contact area between surfaces facilitate charge transfer and electrostatic induction. This in turn drives electron flow in external circuits and expands TENG’s applications across diverse contexts [58]. Practically, we affixed the CS-TENG to a mouthpiece to capture aeroacoustics signals and simulate a silent communication environment such as whispering and silent reading. Through the application of filtering, data processing, and deep learning algorithms, we have successfully achieved speech recognition. When integrated into scenarios such as smart homes and voice assistants, this approach simultaneously presents a novel solution for the silent operation of intelligent technology. Additionally, we bolster the system’s security based on a unique personal voiceprint feature mode to authenticate user’s identities.

2. Materials and Methods

The device structure consists of silk film, conductive fabric, and Polytetrafluoroethylene (PTFE), as shown in Figure 1b. PTFE and conductive fabric served as the friction electrodes. The two electrodes are then encapsulated with a treated soft silk fibroin (SF) film, which has good biocompatibility and mechanical qualities. The top and lower electrodes are separated and contacted when the CS-TENG vibrates. Owing to the materials’ different electronegativities, the mechanical energy from detected vibrations is transformed into electrical signals during contact separation. Subsequently, as shown in the flowchart of Figure 1c, TENG generates output signals for data processing by machine learning algorithms to establish a speech recognition system.
Figure 2a demonstrates that pure SF films exhibit poor tensile properties, breaking under a stress of 18 MPa at only 1.8% strain. In contrast, PU films have good tensile properties [59]. To improve the tensile properties of pure-SF films, the addition of PU leads to PU-modified SF films (treated-SF) exhibiting enhanced tensile strength over pure-SF films, achieving a strain of 120% and a stress of 23 MPa. The water loss rate of the treated SF film exceeded that of polyethylene (PE) film, approximately 5.2 mL h−1, demonstrating the silk film’s superior air permeability (Figure 2b). As depicted in Figure 2c,d, the biocompatibility was assessed using the CCK-8 cell counting method to evaluate cell viability and proliferation on the treated-SF films after one week of incubation. The result shows that the cell density on the samples increased significantly, and their surfaces became progressively covered with cells. The findings indicate that the cells demonstrated robust adhesion, spreading, and growth on the treated SF. This experimental result demonstrates the excellent biocompatibility of the modified silk membrane [55,60].

3. Results and Discussion

3.1. Performance of CS-TENG

As depicted in Figure 3a, after curve fitting, the sensitivity of CS-TENG reaches 7.43 × 10−2 nA kPa−1 under 44 kPa and decreases to 6.52 × 10−3 nA kPa−1 within 44 to 222 kPa. The rapid linear current increase at low pressure results from an accelerated expansion in the contact area between the upper and lower friction electrodes (conductive fabric and PTFE) upon external force application. This illustrates that the device’s sensitivity at low pressure significantly exceeds that at medium and high pressure, making it suitable for some applications like small vibration detection or whispered speech communication. Concurrently, the swift reduction in sensitivity at medium and high pressures broadens the sensor’s detection range, enabling sensitive and extensive dynamic pressure monitoring. To determine the optimal size and area for the TENG, we conducted tests across various dimensions to observe signal changes when the device was tapped by a motor at a consistent frequency. The comparison reveals an increase in device current with size (Figure S1a), likely due to the expanding device area increase in the contact area between the conductive fabric and PTFE. The current rises with the thickness increment from 1 mm to 4 mm; however, beyond 3 mm, the increase in current becomes less pronounced (Figure S1b). Here, a device measuring 2 cm × 2 cm with a 2 mm thickness was consistently selected for testing.
For CS-TENG to stably detect sound signals, it necessitates precise, stable, and repeatable vibration signal sensing across a broad frequency spectrum. CS-TENG accurately detects every high-frequency vibration (40 to 1020 Hz), as evidenced by fast Fourier transform (FFT) analysis (Figure 3b,c). The FFT spectrogram revealed that the response signal’s frequency shift compared to the original frequency resulted in a frequency deviation of no more than 0.63%. This demonstrates the device’s heightened accuracy in high-frequency vibration sensing with little signal distortion. By amplifying the shaker’s vibration intensity at a steady vibration frequency (100 Hz), the sensor’s current signal incrementally rose from 1 nA to 12 nA. This signifies that the current magnitude precisely mirrors the actual vibration intensity. Figure 3d depicts a fitting of the relationship between current output and vibration intensity, approximating a proportional curve. Figure 3e,f illustrates the CS-TENG vibration sensor’s current output under continuous scanning mode across frequencies ranging from 10 to 1020 Hz. The initial current amplitude is significantly higher than that at the end, demonstrating CS-TENG’s enhanced sensitivity to vibrations below 200 Hz. Human vocal frequencies normally occur within 50 to 200 Hz. Consequently, CS-TENG can effectively detect the vibrations associated with human vocalizations. To assess stability under low-frequency vibration, electrical signals were measured in 0.5 Hz intervals from 0.5 Hz to 3 Hz (Figure 3g,h). The test results indicated a modest increase in voltage signal magnitude with ascending frequency. The current amplitude experienced a more pronounced increase as the vibration frequency rose. Consequently, the current signal demonstrated greater sensitivity to vibration perception in low-frequency tests, meaning its output was more significant at identical vibration frequencies. In later voice vibration signal acquisitions, the analysis mainly relies on the current signal. To evaluate its repeatability and durability, voltage and current signals were measured after 30 days. We found that signals are closely matched to those recorded on the first day (Figure 3i), ensuring the TENG works soundly for long-term sound signal acquisition.

3.2. Speech Recognition Function of CS-TENG

Public spaces necessitate a degree of quiet, with studies indicating that noise levels between 20 and 40 dB are conducive to human activity and rest. In the current state of intelligent technology, silent/low-decibel speech recognition still requires further improvement and dissemination. Table S2 presents a comparison and analysis of current speech technologies. In this study, the CS-TENG we developed generates an electrical signal by facilitating contact separation between its upper and lower stages through aeroacoustics transfer airflow induced by whispering. To verify the speech perception capability of CS-TENG, we established a testing mechanism where CS-TENG is attached to a mouthpiece in a laboratory setting (52 dB ambient noise). By speaking at a soft air sound level under ambient noise, the TENG is induced to vibrate, allowing for the detection of the output current signal and the observation of waveforms. The results are then compared and analyzed, and the system is applied in various speech-control scenarios. Figure 4a illustrates the application schematic for this system. On the left are our physical figures. After signal acquisition, the processed signals are input into the model for recognition, and the recognition results are applied to applications such as smart homes and voice assistants. This chapter will elaborate on the processes of signal processing and speech recognition.
Initially, we uttered the same word at a consistent decibel level across varying distances directly in front of the TENG, observing that the output becomes more distinct and the signal change magnitude increases as the distance to the mouthpiece decreases. At 8 cm, the signal is weak. At 6 cm, it is the clearest. And beginning at 5 cm, although the signal strengthens, breathing starts to interfere with the speech signal (Figure S2). Consequently, we selected 6 cm as the optimal testing distance. Subsequently, we further investigated the characteristics of speech signals using a tester. Words of varied pronunciations were classified into three categories: commands, functions, and home. Furthermore, signals from seven distinct sentences were gathered. Figure 4b displays the real-time current signals detected by CS-TENG when a volunteer articulates words and sentences across the three categories (commands, functions, and home). Figure 4c depicts the process of filtering the original signal, followed by the application of the short-time Fourier transform (STFT), revealing three signals in the spectrogram. These signals correspond to three segments of the time-domain waveform, averaging a frequency range between 50 and 75 Hz.
Following the conversion of the signal from the time domain to the frequency domain by STFT, we utilized Mel-frequency cepstral coefficients (MFCC) to extract features from pertinent sound signals in the frequency domain and then trained these features using a neural network model. Following parameter optimization, the model can swiftly learn from the signal features and categorize them into specific groups. In this study, the training model is a neural network that combines convolutional neural networks (CNN) and long short-term memory (LSTM). Initially, the CNN is utilized for advanced feature extraction, isolating valid sound signals, and diminishing the original data volume. Subsequently, the LSTM neural network is selected to optimally process the output of the feature by the CNN, assimilate prolonged effective information, and facilitate its training and classification. Analyzing speech waveforms across three-word categories—commands, functions, and home—and seven common sentences, the recognition accuracies were 92.4%, 95.7%, 96.2%, and 88.2%, respectively. The corresponding confusion matrices for the word recognition accuracy are shown in Figures S4–S6. Figure 4d shows the confusion matrix for sentence recognition accuracy. The combined classification of instructions, functions, and home encompasses a total of 29 terms, achieving an accuracy of 89% (Figure S7). Here, an analysis of the recognition results has been conducted. The accuracy of recognition may be influenced by factors such as the testing environment and friction charges on the device. Therefore, to further increase accuracy, it is important to ensure consistency in signal collection and result verification environments. The accuracy of sentences may be slightly lower compared to individual words, possibly due to the fact that sentences are composed of different words, and there may be slight variations in pauses and pronunciation during each signal collection. When consolidating the three categories of words, the accuracy decreases. This could be because more diverse classifications require larger datasets for support. Therefore, to improve the accuracy of mixed vocabulary, extensive data collection is necessary, followed by multiple rounds of training.
Furthermore, traditional machine learning algorithms, including support vector machine (SVM), random forest (RF), decision tree (DT), and logistic regression (LR), were compared, with their accuracies detailed in Figure 4e. Overall, LR performed best and the DT algorithm the worst. However, all traditional models underperformed compared to the LSTM neural network mode. Our recognition outcomes are presented in a QT-developed GUI interface, which displays the speech’s electrical signal waveform and the recognition results; the interface is depicted in Figure 4f. The signal waveforms illustrating the pronunciation of common words in daily life are shown in Figure S9.

3.3. Identification and Security Defenses

In daily life, security system applications encompass a broad range, including mobile phone unlocking, private robot control, smart homes, and smart vehicles. Consequently, it is crucial to design a security system based on voice recognition. One advantageous method to enhance the security of speech systems is to utilize biometric features for encryption. Biometric features refer to certain physiological or behavioral characteristics unique to each individual that can be effectively measured, identified, and verified through technology, such as facial features, fingerprints, and voiceprints. Biometric recognition technology utilizes various sensing and data collection techniques, as well as artificial intelligence techniques like deep learning, to calculate individual physiological and behavioral features, establish mathematical models, and perform recognition and differentiation. In recent years, there has been a growing demand for biometric recognition technology in various fields. Among various biometric recognition technologies, intelligent speech-based technology is one of the most widely applied.
Figure 5a displays the speech waveforms of four volunteers. For the word “computer”, both the positive and negative amplitudes, along with the mean and extreme values of signal strength for each volunteer, were calculated. It was observed that the signal characteristics varied significantly among individuals (Figure 5c). Subsequently, the speech correlation among different users was further analyzed. Pearson’s coefficients for the four volunteers who articulated the word “computer” were computed based on feature parameters to assess the speech-signal correlation among different users. Coefficients closer to 1 indicated a stronger correlation, while those closer to −1 indicated a weaker correlation. Figure S8 reveals that the correlation between speakers and their signals approaches 1, indicating strong self-consistency, whereas correlations with signals from other testers are significantly lower, highlighting substantial individual variability in articulatory signals. This indicates that, despite volunteers having distinct voices, each possesses a unique speech pattern. Leveraging this, a personalized speech recognition system can be tailored to everyone, establishing a one-to-one match between their vibration signals and spoken words. Consequently, personalized model training, tailored to the unique spectral properties of each user’s voice signals, can accurately identify the four users, serving as a robust authentication mechanism for security. In this study, 10 distinct vocabulary sets comprising a total of 100 words were compiled into a dataset, achieving a recognition accuracy of 90.6%, with the confusion matrix depicted in Figure 5d. By training the developed classifier, the system can identify specific users. Upon a user inputting a unique voice password, the system determines if they are a registered user, thereby enhancing the smart device’s security. The judgment result is additionally presented in the GUI interface (Figure 5e).

4. Conclusions

In this study, we developed a novel device named CS-TENG and applied it in the field of acoustics. CS-TENG exhibits excellent stability and accuracy in both low- and high-frequency ranges, rendering it highly promising for applications in acoustic signal processing. Specifically, our research team utilized CS-TENG to detect vibrations produced by whispered speech and successfully recognized spoken content by analyzing the waveforms and spectrograms of these signals. This technology not only demonstrates good stability and accuracy in both low and high-frequency ranges but also enables the possibility of controlling smart technology through voice commands in quiet environments without causing disturbance, thus offering great flexibility for practical applications in various environments and scenarios.
Here, the feasibility and effectiveness of this technology were validated through experiments. We identified common words and sentences in daily life and applied them to voice assistants and smart home systems. In the experiments, we improved recognition accuracy through neural network algorithms, further enhancing system performance. These experimental results indicate that this technology not only performs well in laboratory environments but also holds potential for practical applications. In addition to its application in speech recognition, we also explored the potential application of this technology in user identity recognition. By analyzing the speaker’s voice characteristics, we can identify different individuals, thereby achieving personalized voice interaction experiences. This application of user identity recognition can enhance security defense and protect personal privacy information within speech systems, opening up more scenarios for the application of speech systems.
In summary, this study provides a new approach and method for the development and application of speech recognition technology. By utilizing CS-TENG’s vibration signals for speech recognition, we not only improve recognition accuracy and stability but also address some issues existing in traditional speech recognition systems. This technology has broad application prospects and is expected to play a significant role in human-computer interaction and communication in quiet places such as libraries and meeting rooms, as well as in information protection and security systems. In the future, with further research and improvements in this technology, it will bring new development opportunities for smart technology and human-computer interaction, providing more convenience and security for people’s lives.

5. Experimental Methods

Experimental Design for the treated-SF: Following the established protocol for silk solution preparation [61,62,63], the silk solution underwent degumming, washing, drying, dissolving, and centrifugation. Subsequently, the PU solution was diluted to a ratio of 1:13 (PU: water, v/v). A mixture of silk solution (1 mL) with PU solution (9 mL) was homogeneously blended and gradually transferred into a petri dish, which was then incubated at 26 °C for 48 h to yield a film.
Experimental Design of CS-TENG: PTFE and conductive fabrics were selected as the friction electrodes. The conductive fabrics, crafted from high-strength polyvinyl acetate (PVAC) fiber cloth, are plated with copper and nickel metals to enhance conductivity and coated with a conductive acrylic adhesive. Wires are attached to the conductive fabric. Following this, both the upper and lower electrodes are encapsulated within treated SF. A gap between the upper and lower electrodes is maintained and sealed around its perimeter with nanoglue to enable the contact separation of the electrodes.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/s24103007/s1, Figure S1:Effect of device area and thickness on signal; Figure S2:Testing the effect of distance on the signal; Figure S3:Test scenarios for voice signals; Figure S4:Confusion matrix for command-recognition results; Figure S5:Confusion matrix for function-recognition results; Figure S6:Confusion matrix for home-recognition results; Figure S7:Confusion matrix for mixed recognition results; Figure S8:Correlation coefficients between four users of the word "computer" are plotted; Figure S9:Waveforms of other speech signals; Table S1: Progress of TENG in Speech Recognition Applications; Table S2:The comparison of current speech recognition technologies; Table S3:Experimental instruments and equipment. The references [39,40,41,55,64,65,66,67,68,69,70,71,72,73,74,75,76,77] in the main text are cited in the Supplementary Information.

Author Contributions

J.L., as the first author, conceived and designed experiments, completed material testing and speech recognition, and wrote a paper; Y.S. provided assistance in the material preparation section; J.C. and M.Y. provided assistance in the drawing section of the paper. W.G. contributed to the data analysis and drafted the manuscript. Q.H. and W.G. edited the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Key Research and Development Project of Jiangxi Province (20212BBG73007, 20212BAB201023), the National Nature Science Foundation of China (22175146, 52001265), the Science and Technology Project of Fujian Province (2022H0001), the Science and Technology Project of Fujian Province (2021J01005), the “111′′ Project (B16029). The volunteer (Jianfeng Chen) agreed to all tests in the manuscript with informed consent, including the skin affinity experiment.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article and Supplementary Materials.

Conflicts of Interest

There are no conflicts to declare.

References

  1. Chen, J.; Wang, Z.L. Reviving Vibration Energy Harvesting and Self-Powered Sensing by a Triboelectric Nanogenerator. Joule 2017, 1, 480–521. [Google Scholar] [CrossRef]
  2. Khandelwal, G.; Maria Joseph Raj, N.P.; Kim, S.J. Materials Beyond Conventional Triboelectric Series for Fabrication and Applications of Triboelectric Nanogenerators. Adv. Energy Mater. 2021, 11, 2101170. [Google Scholar] [CrossRef]
  3. Kim, W.G.; Kim, D.W.; Tcho, I.W.; Kim, J.K.; Kim, M.S.; Choi, Y.K. Triboelectric Nanogenerator: Structure, Mechanism, and Applications. ACS Nano 2021, 15, 258–287. [Google Scholar] [CrossRef]
  4. Liu, Z.; Li, H.; Shi, B.; Fan, Y.; Wang, Z.L.; Li, Z. Wearable and Implantable Triboelectric Nanogenerators. Adv. Funct. Mater. 2019, 29, 1808820. [Google Scholar] [CrossRef]
  5. Cheng, J.; Zhang, X.; Jia, T.; Wu, Q.; Dong, Y.; Wang, D. Triboelectric Nanogenerator with a Seesaw Structure for Harvesting Ocean Energy. Nano Energy 2022, 102, 107622. [Google Scholar] [CrossRef]
  6. Shi, Y.; Wang, F.; Tian, J.; Li, S.; Fu, E.; Nie, J.; Lei, R.; Ding, Y.; Chen, X.; Wang, Z.L. Self-Powered Electro-Tactile System for Virtual Tactile Experiences. Sci. Adv. 2021, 7, eabe2943. [Google Scholar] [CrossRef] [PubMed]
  7. Xin, Y.; Liu, T.; Sun, H.; Xu, Y.; Zhu, J.; Qian, C.; Lin, T. Recent Progress on the Wearable Devices Based on Piezoelectric Sensors. Ferroelectrics 2018, 531, 102–113. [Google Scholar] [CrossRef]
  8. Choi, D.; Lee, Y.; Lin, Z.H.; Cho, S.; Kim, M.; Ao, C.K.; Soh, S.; Sohn, C.; Jeong, C.K.; Lee, J.; et al. Recent Advances in Triboelectric Nanogenerators: From Technological Progress to Commercial Applications. ACS Nano 2023, 17, 11087–11219. [Google Scholar] [CrossRef]
  9. Li, H.; Zhao, L.; Li, Z.; Ni, F.; Liu, Z.; Liu, R.; Fan, Y.; Li, Z. Progress in Self-Powered Health Monitoring and Physiological Function Regulation Devices. Chin. J. Biotechnol. 2019, 35, 2367–2385. [Google Scholar]
  10. Zhao, H.; Shu, M.; Ai, Z.; Lou, Z.; Sou, K.W.; Lu, C.; Jin, Y.; Wang, Z.; Wang, J.; Wu, C. A Highly Sensitive Triboelectric Vibration Sensor for Machinery Condition Monitoring. Adv. Energy Mater. 2022, 12, 2201132. [Google Scholar] [CrossRef]
  11. Wu, C.; Huang, H.; Yang, S.; Wen, G. Pagoda-Shaped Triboelectric Nanogenerator with High Reliability for Harvesting Vibration Energy and Measuring Vibration Frequency in Downhole. IEEE Sensors J. 2020, 20, 13999–14006. [Google Scholar] [CrossRef]
  12. Jin, L.; Zhang, B.; Zhang, L.; Yang, W. Nanogenerator as New Energy Technology for Self-Powered Intelligent Transportation System. Nano Energy 2019, 66, 104086. [Google Scholar] [CrossRef]
  13. Sun, J.; Schütz, U.; Tu, K.; Koch, S.M.; Roman, G.; Stucki, S.; Chen, F.; Ding, Y.; Yan, W.; Wu, C. Scalable and sustainable wood for efficient mechanical energy conversion in buildings via triboelectric effects. Nano Energy 2022, 102, 107670. [Google Scholar] [CrossRef]
  14. Zhang, C.; Yuan, W.; Zhang, B.; Yang, J.; Hu, Y.; He, L.; Zhao, X.; Li, X.; Wang, Z.L.; Wang, J. A Rotating Triboelectric Nanogenerator Driven by Bidirectional Swing for Water Wave Energy Harvesting. Small 2023, 19, 2304412. [Google Scholar] [CrossRef]
  15. Yang, D.; Ni, Y.; Kong, X.; Li, S.; Chen, X.; Zhang, L.; Wang, Z.L. Self-healing and elastic triboelectric nanogenerators for muscle motion monitoring and photothermal treatment. ACS Nano 2021, 15, 14653–14661. [Google Scholar] [CrossRef]
  16. Wen, Z.; Yang, Y.; Sun, N.; Li, G.; Liu, Y.; Chen, C.; Shi, J.; Xie, L.; Jiang, H.; Bao, D. A wrinkled PEDOT: PSS film based stretchable and transparent triboelectric nanogenerator for wearable energy harvesters and active motion sensors. Adv. Funct. Mater. 2018, 28, 1803684. [Google Scholar] [CrossRef]
  17. Gao, C.; Liu, T.; Luo, B.; Cai, C.; Zhang, W.; Zhao, J.; Yuan, J.; Fatehi, P.; Qin, C.; Nie, S. Cellulosic triboelectric materials for stable energy harvesting from hot and humid conditions. Nano Energy 2023, 111, 108426. [Google Scholar] [CrossRef]
  18. Li, Y.; Guo, Z.; Zhao, Z.; Gao, Y.; Yang, P.; Qiao, W.; Zhou, L.; Wang, J.; Wang, Z.L. Multi-layered triboelectric nanogenerator incorporated with self-charge excitation for efficient water wave energy harvesting. Appl. Energy 2023, 336, 120792. [Google Scholar] [CrossRef]
  19. Wang, Y.; Guo, H.; Liao, J.; Qin, Y.; Ali, A.; Li, C. Solid-Liquid triboelectric nanogenerator based on curvature effect for harvesting mechanical and wave energy. Chem. Eng. J. 2023, 476, 146571. [Google Scholar] [CrossRef]
  20. Bu, X.; Zhou, B.; Li, J.; Gao, C.; Guo, J. Orange peel-like triboelectric nanogenerators with multiscale micro-nano structure for energy harvesting and touch sensing applications. Nano Energy 2024, 122, 109280. [Google Scholar] [CrossRef]
  21. Xu, G.; Huang, X.; Shi, R.; Yang, Y.; Wu, P.; Zhou, J.; He, X.; Li, J.; Zen, Y.; Jiao, Y. Triboelectric Nanogenerator Enabled Sweat Extraction and Power Activation for Sweat Monitoring. Adv. Funct. Mater. 2024, 34, 2310777. [Google Scholar] [CrossRef]
  22. Zhang, H.; Zhang, D.-Z.; Wang, D.-Y.; Xu, Z.-Y.; Yang, Y.; Zhang, B. Flexible single-electrode triboelectric nanogenerator with MWCNT/PDMS composite film for environmental energy harvesting and human motion monitoring. Rare Met. 2022, 41, 3117–3128. [Google Scholar] [CrossRef]
  23. Guo, W.; Long, Y.; Bai, Z.; Wang, X.; Liu, H.; Guo, Z.; Tan, S.; Guo, H.; Wang, Y.; Miao, Y. Variable stiffness triboelectric nano-generator to harvest high-speed railway bridge’s vibration energy. Energy Convers. Manag. 2022, 268, 115969. [Google Scholar] [CrossRef]
  24. Xu, J.; Wei, X.; Li, R.; Kong, S.; Wu, Z.; Wang, Z.L. A capsule-shaped triboelectric nanogenerator for self-powered health monitoring of traffic facilities. ACS Mater. Lett. 2022, 4, 1630–1637. [Google Scholar] [CrossRef]
  25. Yuan, H.; Xiao, Z.; Wan, J.; Xiang, Y.; Dai, G.; Li, H.; Yang, J. A rolling-mode Al/CsPbBr3 Schottky junction direct-current triboelectric nanogenerator for harvesting mechanical and solar energy. Adv. Energy Mater. 2022, 12, 2200550. [Google Scholar] [CrossRef]
  26. Liu, C.; Wang, Y.; Zhang, N.; Yang, X.; Wang, Z.; Zhao, L.; Yang, W.; Dong, L.; Che, L.; Wang, G. A self-powered and high sensitivity acceleration sensor with V-Q-a model based on triboelectric nanogenerators (TENGs). Nano Energy 2020, 67, 104228. [Google Scholar] [CrossRef]
  27. Tong, K.; Zhang, Q.; Chen, J.; Wang, H.; Wang, T. Research on Throat Speech Signal Detection Based on a Flexible Graphene Piezoresistive Sensor. ACS Appl. Electron. Mater. 2022, 4, 3549–3559. [Google Scholar] [CrossRef]
  28. Yi, L.; Jiao, W.; Wu, K.; Qian, L.; Yu, X.; Xia, Q.; Mao, K.; Yuan, S.; Wang, S.; Jiang, Y. Nanoparticle Monolayer-Based Flexible Strain Gauge with Ultrafast Dynamic Response for Acoustic Vibration Detection. Nano Res. 2015, 8, 2978–2987. [Google Scholar] [CrossRef]
  29. Liu, Y.; Li, H.; Liang, X.; Deng, H.; Zhang, X.; Heidari, H.; Ghannam, R.; Zhang, X. Speech Recognition Using Intelligent Piezoresistive Sensor Based on Polystyrene Sphere Microstructures. Adv. Intell. Syst. 2023, 5, 2200427. [Google Scholar] [CrossRef]
  30. Chen, F.; Wu, Y.; Ding, Z.; Xia, X.; Li, S.; Zheng, H.; Diao, C.; Yue, G.; Zi, Y. A Novel Triboelectric Nanogenerator Based on Electrospun Polyvinylidene Fluoride Nanofibers for Effective Acoustic Energy Harvesting and Self-Powered Multifunctional Sensing. Nano Energy 2019, 56, 241–251. [Google Scholar] [CrossRef]
  31. Yuan, M.; Li, C.; Liu, H.; Xu, Q.; Xie, Y. A 3d-Printed Acoustic Triboelectric Nanogenerator for Quarter-Wavelength Acoustic Energy Harvesting and Self-Powered Edge Sensing. Nano Energy 2021, 85, 105962. [Google Scholar] [CrossRef]
  32. Zhao, H.; Xiao, X.; Xu, P.; Zhao, T.; Song, L.; Pan, X.; Mi, J.; Xu, M.; Wang, Z.L. Dual-Tube Helmholtz Resonator-Based Triboelectric Nanogenerator for Highly Efficient Harvesting of Acoustic Energy. Adv. Energy Mater. 2019, 9, 1902824. [Google Scholar] [CrossRef]
  33. Hu, S.; Yuan, Z.; Li, R.; Cao, Z.; Zhou, H.; Wu, Z.; Wang, Z.L. Vibration-driven triboelectric nanogenerator for vibration attenuation and condition monitoring for transmission lines. Nano Lett. 2022, 22, 5584–5591. [Google Scholar] [CrossRef]
  34. Zhang, D.; Guo, Q.; Wu, C.; Zhou, Q.; Yang, P. Development of a high temperature resistant downhole vibration sensor based on the triboelectric nanogenerator. Drill. Eng. 2022, 49, 120–127. [Google Scholar]
  35. Tan, D.; Zhou, J.; Wang, K.; Zhao, X.; Wang, Q.; Xu, D. Bow-type bistable triboelectric nanogenerator for harvesting energy from low-frequency vibration. Nano Energy 2022, 92, 106746. [Google Scholar] [CrossRef]
  36. He, C.; Yang, T.; Fang, J.; Pu, X.; Shang, K.; Tian, G.; Lu, X.; Wu, J.; Yang, W.; Qian, L. Tensegrity-inspired triboelectric nanogenerator for broadband and impact-resistive vibration sensing. Nano Energy 2023, 109, 108279. [Google Scholar] [CrossRef]
  37. Du, T.; Shen, D.; Xi, Z.; Yu, H.; Dong, F.; Zhao, C.; Zhu, M.; Zou, Y.; Sun, P.; Xu, M. Highly adaptive and broadband triboelectric energy harvester with stretching silicone rubber strip for variable harmonic frequency vibration. Nano Res. 2023, 17, 4089–4099. [Google Scholar] [CrossRef]
  38. Yao, H.; Wang, Z.; Wu, Y.; Zhang, Y.; Miao, K.; Cui, M.; Ao, T.; Zhang, J.; Ban, D.; Zheng, H. Intelligent Sound Monitoring and Identification System Combining Triboelectric Nanogenerator-Based Self-Powered Sensor with Deep Learning Technique. Adv. Funct. Mater. 2022, 32, 2112155. [Google Scholar] [CrossRef]
  39. Luo, H.; Du, J.; Yang, P.; Shi, Y.; Liu, Z.; Yang, D.; Zheng, L.; Chen, X.; Wang, Z.L. Human-Machine Interaction via Dual Modes of Voice and Gesture Enabled by Triboelectric Nanogenerator and Machine Learning. ACS Appl. Mater. Interfaces 2023, 15, 17009–17018. [Google Scholar] [CrossRef] [PubMed]
  40. Xia, S.-Y.; Long, Y.; Huang, Z.; Zi, Y.; Tao, L.-Q.; Li, C.-H.; Sun, H.; Li, J. Laser-Induced Graphene (Lig)-Based Pressure Sensor and Triboelectric Nanogenerator Towards High-Performance Self-Powered Measurement-Control Combined System. Nano Energy 2022, 96, 107099. [Google Scholar] [CrossRef]
  41. Wang, H.L.; Guo, Z.H.; Pu, X.; Wang, Z.L. Ultralight Iontronic Triboelectric Mechanoreceptor with High Specific Outputs for Epidermal Electronics. Nano-Micro Lett. 2022, 14, 86. [Google Scholar] [CrossRef] [PubMed]
  42. Zhou, B.; Liu, J.; Huang, X.; Qiu, X.; Yang, X.; Shao, H.; Tang, C.; Zhang, X. Mechanoluminescent-triboelectric bimodal sensors for self-powered sensing and intelligent control. Nano-Micro Lett. 2023, 15, 72. [Google Scholar] [CrossRef] [PubMed]
  43. Zhao, J.; Xiao, Y.; Yang, W.; Zhang, S.; Wang, H.; Wang, Q.; Sun, Z.; Li, W.; Gao, M.; Wang, Z. Ultrastretchable Triboelectric Nanogenerators Based on Ecoflex/Porous Carbon for Self-Powered Gesture Recognition. Adv. Mater. Technol. 2023, 8, 2201769. [Google Scholar] [CrossRef]
  44. Ge, X.; Gao, Z.; Zhang, L.; Ji, H.; Yi, J.; Jiang, P.; Li, Z.; Shen, L.; Sun, X.; Wen, Z. Flexible microfluidic triboelectric sensor for gesture recognition and information encoding. Nano Energy 2023, 113, 108541. [Google Scholar] [CrossRef]
  45. Zhang, D.-S.; Song, W.-Z.; Wu, L.-X.; Li, C.-L.; Chen, T.; Sun, D.-J.; Zhang, M.; Zhang, T.-T.; Zhang, J.; Ramakrishna, S. The influence of in-plane electrodes on TENG’s output and its application in the field of IoT intelligent sensing. Nano Energy 2023, 110, 108313. [Google Scholar] [CrossRef]
  46. Li, J.; Wu, C.; Dharmasena, I.; Ni, X.; Wang, Z.; Shen, H.; Huang, S.-L.; Ding, W. Triboelectric nanogenerators enabled internet of things: A survey. Intell. Converg. Netw. 2020, 1, 115–141. [Google Scholar] [CrossRef]
  47. Yang, Y.; Guo, X.; Zhu, M.; Sun, Z.; Zhang, Z.; He, T.; Lee, C. Triboelectric nanogenerator enabled wearable sensors and electronics for sustainable internet of things integrated green earth. Adv. Energy Mater. 2023, 13, 2203040. [Google Scholar] [CrossRef]
  48. Cui, X.; Wu, H.; Wang, R. Fibrous triboelectric nanogenerators: Fabrication, integration, and application. J. Mater. Chem. A 2022, 10, 15881–15905. [Google Scholar] [CrossRef]
  49. Liu, T.; Liang, R.; He, H.; Zeng, Y.; Hou, Z.; Liu, Y.; Yuan, J.; Luo, B.; Zhang, S.; Cai, C. Nanocellulosic triboelectric materials with micro-mountain arrays for moisture-resisting wearable sensors. Nano Energy 2023, 112, 108480. [Google Scholar] [CrossRef]
  50. Khanh, T.D.; Meena, J.S.; Choi, S.B.; Kim, J.-W. Breathable, self-healable, washable and durable all-fibrous triboelectric nanogenerator for wearable electronics. Mater. Today Adv. 2023, 20, 100427. [Google Scholar]
  51. Li, Y.; Tian, Z.; Gao, X.Z.; Zhao, H.Y.; Li, X.; Wang, Z.L.; Yu, Z.Z.; Yang, D. All-Weather Self-Powered Intelligent Traffic Monitoring System Based on a Conjunction of Self-Healable Piezoresistive Sensors and Triboelectric Nanogenerators. Adv. Funct. Mater. 2023, 33, 2308845. [Google Scholar] [CrossRef]
  52. Pang, Y.; Zhu, X.; Yu, Y.; Liu, S.; Chen, Y.; Feng, Y. Waterbomb-origami inspired triboelectric nanogenerator for smart pavement-integrated traffic monitoring. Nano Res. 2022, 15, 5450–5460. [Google Scholar] [CrossRef]
  53. Zhao, J.; Wang, D.; Zhang, F.; Pan, J.; Claesson, P.; Larsson, R.; Shi, Y. Self-powered, long-durable, and highly selective oil–solid triboelectric nanogenerator for energy harvesting and intelligent monitoring. Nano-Micro Lett. 2022, 14, 160. [Google Scholar] [CrossRef] [PubMed]
  54. Cui, M.; Guo, H.; Zhai, W.; Liu, C.; Shen, C.; Dai, K. Template-Assisted Electrospun Ordered Hierarchical Microhump Arrays-Based Multifunctional Triboelectric Nanogenerator for Tactile Sensing and Animal Voice-Emotion Identification. Adv. Funct. Mater. 2023, 33, 2301589. [Google Scholar] [CrossRef]
  55. Zhao, J.; Chen, D.; Zhao, L.; Shi, Y.; Guo, S.; Zhu, Z.; Liu, J.; Li, W.; Lei, W.; Chen, H. Self-powered speech recognition system for deaf users. Cell Rep. Phys. Sci. 2022, 3, 101168. [Google Scholar] [CrossRef]
  56. Du, M.; Cao, Y.; Qu, X.; Xue, J.; Zhang, W.; Pu, X.; Shi, B.; Li, Z. Hybrid nanogenerator for biomechanical energy harvesting, motion state detection, and pulse sensing. Adv. Mater. Technol. 2022, 7, 2101332. [Google Scholar] [CrossRef]
  57. Niu, S.; Wang, Z.L. Theoretical Systems of Triboelectric Nanogenerators. Nano Energy 2015, 14, 161–192. [Google Scholar] [CrossRef]
  58. Cao, X.; Jie, Y.; Wang, N.; Wang, Z.L. Triboelectric Nanogenerators Driven Self-Powered Electrochemical Processes for Energy and Environmental Science. Adv. Energy Mater. 2016, 6, 1600665. [Google Scholar] [CrossRef]
  59. Shi, C.Y.; Zhang, Q.; Yu, C.Y.; Rao, S.J.; Yang, S.; Tian, H.; Qu, D.H. An Ultrastrong and Highly Stretchable Polyurethane Elastomer Enabled by a Zipper-Like Ring-Sliding Effect. Adv. Mater. 2020, 32, e2000345. [Google Scholar] [CrossRef]
  60. Li, S.; Liu, A.; Qiu, W.; Wang, Y.; Liu, G.; Liu, J.; Shi, Y.; Li, Y.; Li, J.; Cai, W. An All-Protein Multisensory Highly Bionic Skin. ACS Nano 2024, 18, 4579–4589. [Google Scholar] [CrossRef]
  61. Hou, C.; Xu, Z.; Qiu, W.; Wu, R.; Wang, Y.; Xu, Q.; Liu, X.Y.; Guo, W. A Biodegradable and Stretchable Protein-Based Sensor as Artificial Electronic Skin for Human Motion Detection. Small 2019, 15, e1805084. [Google Scholar] [CrossRef]
  62. Huang, J.; Xu, Z.; Qiu, W.; Chen, F.; Meng, Z.; Hou, C.; Guo, W.; Liu, X.Y. Stretchable and Heat-Resistant Protein-Based Electronic Skin for Human Thermoregulation. Adv. Funct. Mater. 2020, 30, 1910547. [Google Scholar] [CrossRef]
  63. Xu, Z.; Qiu, W.; Fan, X.; Shi, Y.; Gong, H.; Huang, J.; Patil, A.; Li, X.; Wang, S.; Lin, H.; et al. Stretchable, Stable, and Degradable Silk Fibroin Enabled by Mesoscopic Doping for Finger Motion Triggered Color/Transmittance Adjustment. ACS Nano 2021, 15, 12429–12437. [Google Scholar] [CrossRef] [PubMed]
  64. Sun, H.; Gao, X.; Guo, L.Y.; Tao, L.Q.; Guo, Z.H.; Shao, Y.; Cui, T.; Yang, Y.; Pu, X.; Ren, T.L. Graphene-based dual-function acoustic transducers for machine learning-assisted human-robot interfaces. InfoMat 2023, 5, 66–79. [Google Scholar] [CrossRef]
  65. Lu, Y.; Tian, H.; Cheng, J.; Zhu, F.; Liu, B.; Wei, S.; Ji, L.; Wang, Z.L. Decoding lip language using triboelectric sensors with deep learning. Nat. Commun. 2022, 13, 1401. [Google Scholar] [CrossRef] [PubMed]
  66. Lee, J.P.; Jang, H.; Jang, Y.; Song, H.; Lee, S.; Lee, P.S.; Kim, J. Encoding of multi-modal emotional information via personalized skin-integrated wireless facial interface. Nat. Commun. 2024, 15, 530. [Google Scholar] [CrossRef] [PubMed]
  67. Yang, J.; Chen, J.; Su, Y.; Jing, Q.; Li, Z.; Yi, F.; Wen, X.; Wang, Z.; Wang, Z.L. Eardrum-inspired active sensors for self-powered cardiovascular system characterization and throat-attached anti-interference voice recognition. Adv. Mater. 2015, 27, 1316–1326. [Google Scholar] [CrossRef]
  68. Yang, H.; Li, Q.; Zhang, X.; Li, X.; Yang, Q.; Hu, Y.; Xi, Y.; Wang, Z.L. High-sensitive and ultra-wide spectrum multifunctional triboelectric acoustic sensor for broad scenario applications. Nano Energy 2022, 104, 107932. [Google Scholar] [CrossRef]
  69. Huang, X.; Acero, A.; Hon, H.-W.; Reddy, R. Spoken Language Processing: A Guide to Theory, Algorithm, and System Development; Prentice Hall: Upper Saddle River, NJ, USA, 2001. [Google Scholar]
  70. Bilmes, J.A. Graphical models and automatic speech recognition. In Mathematical Foundations of Speech and Language Processing; Springer: New York, NY, USA, 2004; pp. 191–245. [Google Scholar]
  71. Denby, B.; Schultz, T.; Honda, K.; Hueber, T.; Gilbert, J.M.; Brumberg, J.S. Silent speech interfaces. Speech Commun. 2010, 52, 270–287. [Google Scholar] [CrossRef]
  72. Herff, C.; Schultz, T. Automatic speech recognition from neural signals: A focused review. Front. Neurosci. 2016, 10, 204946. [Google Scholar] [CrossRef]
  73. Maier-Hein, L.; Metze, F.; Schultz, T.; Waibel, A. Session independent non-audible speech recognition using surface electromyography. In Proceedings of the IEEE Workshop on Automatic Speech Recognition and Understanding, Cancun, Mexico, 27 November–1 December 2005; pp. 331–336. [Google Scholar]
  74. Meltzner, G.S.; Sroka, J.J.; Heaton, J.T.; Gilmore, L.D.; Colby, G.; Roy, S.H.; Chen, N.; De Luca, C.J. Speech recognition for vocalized and subvocal modes of production using surface EMG signals from the neck and face. In Proceedings of the INTERSPEECH, Brisbane, Australia, 22–26 September 2008; pp. 2667–2670. [Google Scholar]
  75. Sun, H.; Zhang, X.; Zhao, Y.; Zhang, Y.; Zhong, X.; Fan, Z. A novel feature optimization for wearable human-computer interfaces using surface electromyography sensors. Sensors 2018, 18, 869. [Google Scholar] [CrossRef] [PubMed]
  76. Jorgensen, C.; Binsted, K. Web Browser Control Using EMG Based Sub Vocal Speech Recognition. In Proceedings of the Proceedings of the 38th Annual Hawaii International Conference on System Sciences, Big Island, HI, USA, 6 January 2005; p. 294c. [Google Scholar]
  77. Xiong, X. Fundamentals of Speech Recognition. Available online: https://slpcourse.github.io/materials/lecture_notes/Xiao_Guest_lecture_ASR.pdf (accessed on 21 April 2024).
Figure 1. Overall schematic and flowchart: (a) Schematic diagram of the structure and application scenarios of CS-TENG; (b) Working mechanism of CS-TENG; (c) Flowchart of the speech recognition system.
Figure 1. Overall schematic and flowchart: (a) Schematic diagram of the structure and application scenarios of CS-TENG; (b) Working mechanism of CS-TENG; (c) Flowchart of the speech recognition system.
Sensors 24 03007 g001
Figure 2. Characterizations of the treated SF: (a) Comparison of tensile properties between treated SF and pure SF; (b) Permeability test; (c) Biocompatible cell counting; (d) Cell fluorescence map (scale bar: 250 μm).
Figure 2. Characterizations of the treated SF: (a) Comparison of tensile properties between treated SF and pure SF; (b) Permeability test; (c) Biocompatible cell counting; (d) Cell fluorescence map (scale bar: 250 μm).
Sensors 24 03007 g002
Figure 3. Device performance tests: (a) Current sensitivity measurements; (b,c) High-frequency vibration test performance and its corresponding FFT; (d) Variation in current with vibration intensity; (e,f) Sweep signal and its frequency domain FFT; (g,h) Low-frequency vibration signals in voltage or current; (i) Working repeatability and durability after 30 days.
Figure 3. Device performance tests: (a) Current sensitivity measurements; (b,c) High-frequency vibration test performance and its corresponding FFT; (d) Variation in current with vibration intensity; (e,f) Sweep signal and its frequency domain FFT; (g,h) Low-frequency vibration signals in voltage or current; (i) Working repeatability and durability after 30 days.
Sensors 24 03007 g003
Figure 4. Acquisition of speech signals: (a) Schematic of the flow of speech signal acquisition and recognition; (b) Current waveform acquisition of the current of speech signals for different kinds of words (command, function, home) and sentences; (c) STFT after noise filtering of raw speech signals; (d) Confusion matrix of sentence recognition results; (e) Comparison of the accuracy of different algorithms for recognizing three categories of words (instruction, function, home); (f) Display interface of recognition results.
Figure 4. Acquisition of speech signals: (a) Schematic of the flow of speech signal acquisition and recognition; (b) Current waveform acquisition of the current of speech signals for different kinds of words (command, function, home) and sentences; (c) STFT after noise filtering of raw speech signals; (d) Confusion matrix of sentence recognition results; (e) Comparison of the accuracy of different algorithms for recognizing three categories of words (instruction, function, home); (f) Display interface of recognition results.
Sensors 24 03007 g004
Figure 5. Recognition and analysis of voice signals: (a) Speech signal waveform of four testers speaking the same word; (b) Application scenario of voiceprint recognition; (c) Comparison of signal eigenvalues for different testers; (d) Confusion matrix for identity recognition accuracy; (e) Display interface for identity recognition.
Figure 5. Recognition and analysis of voice signals: (a) Speech signal waveform of four testers speaking the same word; (b) Application scenario of voiceprint recognition; (c) Comparison of signal eigenvalues for different testers; (d) Confusion matrix for identity recognition accuracy; (e) Display interface for identity recognition.
Sensors 24 03007 g005
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, J.; Shi, Y.; Chen, J.; Huang, Q.; Ye, M.; Guo, W. Flexible Self-Powered Low-Decibel Voice Recognition Mask. Sensors 2024, 24, 3007. https://doi.org/10.3390/s24103007

AMA Style

Li J, Shi Y, Chen J, Huang Q, Ye M, Guo W. Flexible Self-Powered Low-Decibel Voice Recognition Mask. Sensors. 2024; 24(10):3007. https://doi.org/10.3390/s24103007

Chicago/Turabian Style

Li, Jianing, Yating Shi, Jianfeng Chen, Qiaoling Huang, Meidan Ye, and Wenxi Guo. 2024. "Flexible Self-Powered Low-Decibel Voice Recognition Mask" Sensors 24, no. 10: 3007. https://doi.org/10.3390/s24103007

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop