Next Article in Journal
Estimation of Respiratory Rate during Biking with a Single Sensor Functional Near-Infrared Spectroscopy (fNIRS) System
Previous Article in Journal
Incorporating Multivariate Auxiliary Information for Traffic Prediction on Highways
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Intelligent Microsystem for Sound Event Recognition in Edge Computing Using End-to-End Mesh Networking

1
School of Instrument Science and Opto-Electronics Engineering, Beijing Information Science and Technology University, Beijing 100192, China
2
Department of Precision Instrument, Tsinghua University, Beijing 100084, China
3
Key Laboratory of Smart Microsystem (Tsinghua University), Ministry of Education, Beijing 100084, China
4
State Key Laboratory of Precision Measurement Technology and Instruments, Beijing 100084, China
5
Beijing Laboratory of Biomedical Detection Technology and Instrument, Beijing 100084, China
6
Beijing Advanced Innovation Center for Integrated Circuits, Beijing 100084, China
7
China Mobile Research Institute, Beijing 100053, China
8
The IoT Intelligent Microsystem Center, Tsinghua University-China Mobile Joint Research Institute, Beijing 100084, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(7), 3630; https://doi.org/10.3390/s23073630
Submission received: 22 February 2023 / Revised: 22 March 2023 / Accepted: 29 March 2023 / Published: 31 March 2023
(This article belongs to the Section Internet of Things)

Abstract

:
Wireless acoustic sensor networks (WASNs) and intelligent microsystems are crucial components of the Internet of Things (IoT) ecosystem. In various IoT applications, small, lightweight, and low-power microsystems are essential to enable autonomous edge computing and networked cooperative work. This study presents an innovative intelligent microsystem with wireless networking capabilities, sound sensing, and sound event recognition. The microsystem is designed with optimized sensing, energy supply, processing, and transceiver modules to achieve small size and low power consumption. Additionally, a low-computational sound event recognition algorithm based on a Convolutional Neural Network has been designed and integrated into the microsystem. Multiple microsystems are connected using low-power Bluetooth Mesh wireless networking technology to form a meshed WASN, which is easily accessible, flexible to expand, and straightforward to manage with smartphones. The microsystem is 7.36 cm3 in size and weighs 8 g without housing. The microsystem can accurately recognize sound events in both trained and untrained data tests, achieving an average accuracy of over 92.50% for alarm sounds above 70 dB and water flow sounds above 55 dB. The microsystems can communicate wirelessly with a direct range of 5 m. It can be applied in the field of home IoT and border security.

1. Introduction

A Wireless Sensor Network (WSN) is a platform designed for efficient information acquisition that boasts quick deployment, wide coverage, and robust destruction resistance [1,2,3]. Sound signals are a rich source of information, and Wireless Acoustic Sensor Networks (WASNs) with sound event recognition capabilities are widely used in the field of Ambient Assisted Living (AAL), such as in smart homes [4]. Intelligent microsystems possess independent functionalities, including information acquisition, data processing, storage, communication, and energy harvesting [5]. Due to their advantageous features of intelligence, miniaturization, and affordability, intelligent microsystems are quintessential representatives of IoT terminals, garnering increasing attention [6,7].
Intelligent microsystems face severe limitations and interrelated constraints in practical applications and designs, including energy supply, communication, processing, sensing, and size. The utilization of high-performance processors and high-precision sensors increases the size and power consumption of microsystems. In contrast, the use of low-power devices, such as low-performing processors and low-precision sensors, impedes their intelligent sensing and processing capabilities [8]. Sending the data collected by microsystems to high-performance equipment for further processing results in an increase in communication costs and response time. Moreover, the sensing and communication range of a single microsystem is highly limited. As the number of IoT terminals grows, autonomous operation of microsystems and networking cooperation among microsystems become increasingly vital [9]. Thus, building a WASN with intelligent microsystems as nodes requires a balance between small size, low power consumption, sound sensing and recognition, and networked communication, making it a system optimization problem that warrants further research and practice.
Previous research has reported on some WASNs for sound event recognition [10]. However, a WASN proposed by [11] for noise measurement and sound event recognition in urban environments had poorly integrated sensing nodes and did not implement recognition locally. Other WASNs proposed for sound event recognition [12,13] utilized sensing nodes capable of local recognition assembled from microphones, Raspberry Pi, and power supplies. However, they had poor integration, large size, high power consumption, and high cost. Several reports have also focused on intelligent microsystems, including a 15 cm3 and 35 g microsystem for vibration sensing and target recognition [14]. Although the microsystem was well-integrated and could realize vibration wake-up and self-energy supply, its recognition accuracy was not optimal, and it did not implement wireless networking. Current research related to sound perception in intelligent microsystems has primarily focused on developing highly integrated, ultra-low-power sound sensor components, lacking effective system-level applications [15,16]. Zheng You presented an intelligent microsystem of 1.68 cm3 for sound source localization that consisted of a MEMS microphone and MCU, which had an excellent integration [17]. The acoustic source localization network consisted of four microsystem nodes with good localization accuracy, but the node access capability and scalability of the network had not been specified. To date, no intelligent microsystems with a small size, low power consumption, sound event recognition, and wireless networking capability have been reported.
This paper presents the design and implementation of an intelligent microsystem capable of sound sensing, sound event recognition, and wireless networking. The hardware and algorithms were optimized to reduce size, weight, and power consumption while improving sound event recognition performance. Figure 1 illustrates the overall working principle of the microsystem, which integrates a MEMS microphone and MCU and is fully packaged. It is powered using a lithium battery and supports plug-and-play functionality. A CNN-based sound event recognition algorithm is designed and ported to the microsystem for edge computing, enabling high accuracy and responsive sound event recognition with MCU-level computing resources. Multiple microsystems form a mesh WASN using low-power Bluetooth (BLE) Mesh wireless networking technology, which can be managed by connecting to smartphones. The microsystem’s performance, including sound event recognition and networking communication, was verified with tests and experiments. This intelligent microsystem is suitable for various applications, such as smart homes, environmental monitoring, and unattended operations, as an effective IoT terminal. This is the first reported implementation of an intelligent microsystem with a small size, low power consumption, sound event recognition, and wireless networking capabilities.

2. Hardware Design

The hardware of the intelligent microsystem consists of four basic units: the sensing module, the processing module, the wireless transceiver module, and the power supply module. The composition of the microsystem is shown in Figure 2a. The selected main MCU is a low-power processor containing a Bluetooth chip for data processing and transmission. A MEMS microphone is used to capture sound signals and send sound data to the MCU using the I2S interface. An antenna for RF communication is integrated. The power supply module consists of a lithium battery and a supporting energy management circuit.
As shown in Figure 2b, the MCU, MEMS microphone, antenna, and energy management circuit are compactly integrated into a single PCB, called the main board, to reduce the size of the microsystem. To improve sound sensing quality and minimize noise interference, the microphone is positioned on the back of the main board while the sound hole on the front of the board faces the sound source. The MCU, a model STM32WB55, has two cores and supports multiple operating modes, including the sleeping mode and the active mode. The application processor CPU1 is an ARM Cortex-M4 core for overall control, and the network processor CPU2 is an ARM Cortex-M0 with the Bluetooth protocol stack. In addition, the MCU has a floating-point unit (FPU) and embedded memory (1 MB Flash, 256 KB SRAM) to meet the computing and processing requirements of this microsystem [14]. The ICS-43434 MEMS omnidirectional microphone has an impressive ultra-high 120 dB SPL AOP, as well as a high SNR of 64 dB, and a broadband frequency response of 60 Hz to 20 kHz. It supports high performance, low power, and sleep modes of operation and comes in a compact surface mount package measuring only 3.5 mm × 2.65 mm × 0.98 mm with a bottom sound hole. The MEMS microphone can directly communicate with the MCU using the I2S interface without an audio codec, allowing it to collect environmental sound and send the data to the MCU for processing based on the sampling rate set by the software program. The energy supply module uses a lithium battery for energy storage and power supply. The energy management circuit uses the ADP5090 chip to achieve energy management for charging and discharging, which also has over-discharge protection and over-charge protection to make the system safer and more reliable.
The microsystem hardware integration is shown in Figure 2c. The main board and battery are the core of the microsystem and can function independently. For experimental and debugging purposes, an extension board with plug-and-play and program debugging interfaces has been added. The main board and extension board are secured with four screws to form a skeleton, inside of which the lithium battery is placed. The housing is made of semi-transparent photosensitive resin material created using 3D printing, with holes for electrical connection and sound acquisition. Packaged as a cylinder, the microsystem is highly practical and capable of being powered either with wire for extended periods or with a lithium battery for portable applications. The size of the microsystem is presented in Table 1. In sleep mode, the microsystem consumes 0.18 mW, and this increases to 36 mW in active mode [14].

3. Software Design

3.1. Sound Event Recognition

Alarm sounds indicate danger, and long-time water flow sounds represent a waste of water. In this paper, these two abnormal sound events are selected as recognition targets. To reduce interference and improve recognition accuracy, six types of sounds—door knocking, mouse-clicking, keyboard tapping, door creaking, washing machine running, and indoor background sound—are added to the recognition model as noise. Due to the limited sensing and computing resources of the intelligent microsystem, a simple and efficient sound event recognition algorithm and system workflow are designed to achieve accurate and responsive recognition of sound events, as shown in Figure 3.

3.1.1. Data Acquisition and Preprocessing

Figure 3a depicts the collection of data from two sources. The first was a public dataset named ESC-50 [18], which comprises 2000 environmental recordings of 50 types of sounds, each with a duration of 5 s, a sampling rate of 44.1 kHz, and a single channel. The sounds of interest, including alarm, door knocking, mouse-clicking, keyboard tapping, door creaking, and washing machine running, were extracted from ESC-50. The other source was an indoor experimental dataset, containing recordings of water flow sounds and indoor background sounds, captured with smartphones. The recordings in this dataset vary in duration, have a sampling rate of 48 kHz, and consist of two channels.
In order to ensure sample consistency, responsive recognition, and compliance with the microsystem processing capability, the data were pre-processed, as shown in Figure 3b. This involved segmenting the data into new samples with a duration of 1 s, a single channel, and resampling it to a rate of 16 kHz. It is worth noting that some of the new samples may not contain valid sounds and will need to be filtered out.
To ensure balance in the different types of sound samples, the authors used 300 samples for each of the water flow and alarm sounds as target samples, and 50 samples for each of the other six types of sounds, totaling 300 samples, as noise samples. Both target and noise samples were used for model training and testing. As there were fewer valid samples of water flow and alarm sounds, data augmentation was necessary, which not only expanded the dataset but also helped the classification model avoid overfitting [19,20]. For this study, data augmentation was performed by altering the timing and pitch of the sound samples.

3.1.2. Feature Extraction

The formula for converting ordinary frequency to mel-frequency is mel(f) = 2595 × log10(1 + f/700), where f is the frequency of the sound signal in Hertz (Hz).
The flow of MFSC feature extraction is shown in Figure 3c. First, the sound signal is framed and windowed. Then, the spectrum is obtained using fast Fourier transform (FFT) for each short-time window. Next, the Mel filter bank is used to convert the sound from the spectral domain to the Mel spectral domain. For non-vocal sounds, a Mel filter bank with the same bank height is selected [21], and the number of Mel filters is set to 30. Finally, the MFSC feature is obtained using natural logarithm processing of the Mel spectrum.
Using this process, sound data of 1 s duration and 16 kHz sampling rate can be transformed into a (30,32,1) MFSC feature vector, significantly reducing the number of operations and complexity of the recognition model input.

3.1.3. CNN Model Training

Artificial neural networks are commonly used for pattern recognition, and one of the most well-known networks is the Convolutional Neural Network (CNN) [22]. Compared to traditional machine learning algorithms, a CNN has good self-learning, fitting, and generalization abilities, as well as a simple structure and few parameters, making it easier to optimize the model. Additionally, a CNN has parallel processing and fast computing capabilities, making it suitable for edge computing devices [23,24]. Therefore, in this paper, a CNN is chosen as the classification model for sound event recognition, and the model is designed, debugged, and evaluated using the artificial neural network library Keras.
Figure 3d shows that the dataset is divided into training, validation, and test sets with a ratio of 7:1:2. The model training process is illustrated in Figure 3e. The model takes the MFSC feature vector as input and outputs the probability of each sound event type. To suit the model’s intended use in the microsystem, a CNN with a simple structure is designed with parameters as indicated in Table 2. The processing complexity of the model is represented by the MACC [25], which is a unit of multi-ply-and-accumulate complexity.
The model training process is illustrated in Figure 4a. After 100 epochs of training, the accuracy of both the training set and the validation set approaches 99%. Figure 4b shows the confusion matrix for the model test. The average accuracy of the test set is 95%, with the alarm sound and water flow sound achieving 100% accuracy, while the accuracy for other sounds is 85%.

3.1.4. Model Quantification and Porting

Due to limited computational resources in the microsystem, deploying the trained 32-bit floating-point model with large memory usage directly into the microsystem is not feasible. Therefore, the model needs to be converted into an 8-bit fixed-point model using quantization. Quantization involves establishing a mapping relationship between floating-point and fixed-point data with post-training quantization (PTQ) and quantization awareness training (QAT) being the commonly used quantization methods [26]. PTQ is relatively simple and allows the quantization of a pre-trained model with a limited representative dataset. On the other hand, QAT is more complex but yields better model accuracy by performing quantization during the training process.
As illustrated in Figure 3g, this paper performs 8-bit PTQ and ports the trained CNN model based on STM32 X_CUBE_AI. As shown in Table 3, the quantized model has a 0.25% decrease in accuracy, a 7.99% decrease in complexity, a 71.94% decrease in peak RAM usage, and a 73.98% decrease in peak flash usage. By trading off a small loss in accuracy, quantization achieves the goal of compressing model parameters, reducing model memory usage, lowering model complexity, and improving model running speed. Finally, the quantized model is converted into C files containing the model structure, weight parameters, and calling interfaces, which are then deployed to the microsystem to recognize sound events.

3.1.5. System Workflow

To optimize the software workflow of the microsystem, given the limited hardware resources and sound event recognition performance, several steps have been taken. These include system initialization, event initial detection, system work mode switching, preprocessing, feature extraction, and event recognition, as depicted in Figure 3h.
During system initialization, the microsystem determines the signal sampling period, microphone sampling rate, and ambient noise threshold before entering the low-power sleep mode. In this mode, the microphone remains active for sound detection, while the rest of the microsystem operates at a low-power state.
Event initial detection relies on sound amplitude to determine the presence of an event. If the sum of sound amplitudes in a sampling period is greater than five times the ambient noise threshold or twice the sum value of the previous sampling period, the microsystem switches to the high-power active mode. During this detection process, sound signals are continuously collected at a sampling rate of 16 kHz. A frame is made up of 1024 sampling points, with a frameshift of 512 sampling points. The Hanning window is used, and the window length is the same as the frame length.
Mel filtering of one frame of data outputs 30 Mel spectral coefficients, and a 1 s sound contains 32 frames. Therefore, the microsystem performs 32 cycles of processing to obtain the MFSC feature vector of a 1 s sound. This feature vector is used as input for a CNN classifier, which is run to obtain the sound event type.

3.2. Wireless Mesh Networking

The microsystem can recognize sound events independently, but its communication and sensing range is limited, so it needs to be networked with other microsystems. Topologies of WSNs include the star, tree, mesh, hierarchical topologies, etc. Meshed networks are also known as “multi-hop” networks, in which nodes communicate in a many-to-many manner, with the advantages of redundancy and reliability [27,28]. The common mesh networking technologies include BLE mesh [29], WIFI mesh [13], and ZigBee mesh [27]. BLE mesh has the advantages of self-organization, high security, simple routing, low power consumption, and a high device support [29]. Therefore, in this paper, BLE mesh is used as a networking method for the microsystems.

3.2.1. BLE Mesh Theory

BLE mesh is a Bluetooth Low Energy-based networking specification published by Bluetooth SIG [29]. It uses the “flooding” method [30] and the publish/subscribe mechanism [31] to transmit messages. There are different types of nodes: relay, proxy, friend, and low-power. Relay nodes retransmit received messages, and they enable multiple “hops” in the network. Proxy nodes expose the interface for a smartphone or other devices to interact with the mesh network. A BLE mesh network has a theoretical maximum of 32,767 nodes and up to 126 hops [31]. BLE mesh defines four address types, three of which are used for message delivery: unicast addresses, virtual addresses, and group addresses [32]. The architecture of BLE mesh is shown in Figure 5, where the model layer defines the functions and behaviors of the nodes [31]. From the control perspective, the models can be divided into Server, Client, and Control models. Server models expose their state to other models, Client models can access the state of Server models, and Control models are a combination of them. From the functional perspective, there are Generic models, Lighting models, Sensor models, and Vendor models.

3.2.2. Design and Implementation

The STM32WB55 MCU has been selected for its ability to support BLE mesh technology. Figure 6 illustrates how multiple microsystems can be networked together, enabling control and monitoring of sound event recognition results using a mobile device. The network is easily accessible, flexible to expand, and straightforward to manage. Each microsystem includes relay and proxy features, as well as Sensor Server models for sensing and Vendor models for control.
When connected to the network, the microsystem nearest to the smartphone within the direct Bluetooth communication range acts as a proxy node. The smartphone can publish messages, such as recognition commands, which the microsystems can receive if subscribed. The microsystems receive the recognition command and perform sound event recognition locally, then publish the sound event recognition results. Different microsystems can subscribe to specific message groups, such as those organized by floor. Messages can be relayed between any two microsystems within the direct Bluetooth communication range, allowing for flexible management of microsystems within the network and enabling monitoring of sound events in any location within the network coverage.
As depicted in Figure 6, a mesh network is formed by several microsystems and mobile phones, which are divided into two groups. In this configuration, Group 2 receives sound event recognition instructions from the phone and subscribes to the sound event recognition results. The sound recognition command is transmitted to the proxy microsystem that is directly connected to the phone. The “command” message is then relayed to the microsystem of Group 2 using several microsystems. The microsystem of Group 2 subscribes to the recognition command, performs sound event recognition locally, and publishes the sound event recognition results. The “recognition results” message is then relayed to the phone via several microsystem relays, enabling sound event monitoring.

4. Results and Discussion

In the sound event recognition experiment, the performance of microsystem sound event recognition in edge computing was tested using alarm and water flow sounds as targets.
In the networking communication experiment, the communication range between a smartphone and a microsystem, as well as the relay range between two microsystems, were tested.
The networking recognition experiment involved deploying multiple microsystems indoors to verify their networking communication and generalization recognition capabilities.

4.1. Sound Event Recognition Experiment

In the sound event recognition experiment, 25 alarm audios and 25 water flow audios from the dataset described in Section 3.1.1, each 5 s long, were used as experimental samples. The microsystem’s recognition accuracy was tested indoors at different distances from the sound source and with different sound intensities, as shown in Figure 7a.
Three sound samples, namely an alarm sound (shown in Figure 8a), a water flow sound (shown in Figure 8b), and noise samples (shown in Figure 8c), were randomly selected from the experimental dataset. The time domain distribution and MFSC features of these sound signals are presented in Figure 8.
While the amplitude of the alarm sound is greater than that of the water flow sound, this feature alone cannot distinguish between the two types of sound events since the distance between the microsystem and the sound source is flexible. However, the MFSC features of the alarm sound and water flow sound show clear differentiation. The frequency of the alarm sound is mainly concentrated above 500 Hz, while the frequency of the water flow sound is mostly concentrated between 0 and 100 Hz.
Figure 9 shows that the microsystem’s recognition accuracy improves as the distance between the microsystem and the sound source decreases and as the intensity of the sound source increases. As shown in Table 4, when tested indoors with low noise interference using the trained data, the microsystem can effectively identify sound events within 6 m with 100% accuracy and an average recognition time of no more than 3 s when the alarm sound is greater than 70 dB and the water flow sound is greater than 50 dB.
This suggests that the microsystems are capable of responsive and accurate sound event recognition. It also confirms the detection range of a single microsystem, which is important information for the networking communication experiment.

4.2. Networking Communication Experiment

In the networking communication experiment, the aim was to network the smartphone and the microsystems. The smartphone would send control commands to the microsystems while the microsystems would send sound event recognition results to the smartphone. As shown in Figure 7b, the range of direct communication between the smartphone and a single microsystem is around 10 m, while the direct communication distance between two microsystems is around 5 m.

4.3. Networking Recognition Experiment

Generalizability refers to a model’s ability to accurately predict on new, previously unseen data after being trained on a particular dataset. To evaluate the effectiveness and generalization of the sensing network in practical applications, a new test dataset is constructed, comprising audio samples of alarm sounds and water flow that were not used in model training. The sample parameters for the test dataset are shown in Table 5.
Seven microsystems were deployed in a house according to confirmed relay distances, as shown in Figure 7c, which illustrates the layout of the house and the location of the microsystems. Tests were conducted to recognize water flow sounds in Floor1-Kitchen and alarm sounds on Floor2-Room3, and the smartphone communication to the sensing network was verified once again. The sound source distance for this experiment was set to 0.1 m, 1 m, 2 m, and 3 m, considering that the communication distance between the two microsystems is approximately 5 m.
Figure 10 illustrates the interaction process between the smartphone and the microsystem sensing network, which proceeds as follows: scanning the unprovisioned microsystems, configuring them as network nodes, controlling them by group, and receiving the sound event recognition results from the microsystems. From anywhere in the house, the smartphone could monitor the sound event recognition results of all microsystem nodes in real time. The smartphone could also control this meshed network in a flexible manner, such as controlling a microsystem individually, controlling multiple microsystems in groups (e.g., by floor), or controlling all microsystems in the house. This demonstrates that the end-to-end meshed network based on BLE mesh and the microsystems has good topological flexibility and network scalability while also effectively extending the detection range.
The effective recognition standard was set to achieve a recognition accuracy rate greater than 90% within a range of 3 m. As illustrated in Figure 11, the sound intensity of the target sound event increases as the distance between the microsystem and the sound source decreases, resulting in higher recognition accuracy. According to Table 6, sound events, such as alarm sounds with a sound intensity greater than 70 dB, and water flow sounds with a sound intensity greater than 55 dB can be effectively recognized with an average recognition accuracy rate of 96.30% and 92.50%, respectively. This demonstrates that the proposed sound event recognition approach is effective and has general applicability.

5. Conclusions

This paper presents the design and implementation of an intelligent microsystem with sound perception, sound event recognition, and wireless networking capabilities. The hardware modules are optimized for small size, lightweight, and low power consumption. The microsystem, which can be powered with a lithium battery and supports plug-and-play, has a volume of 7.36 cm3 and a weight of 8 g without housing and 14.14 cm3 and a weight of 13 g with housing. The software includes a low-computation sound event recognition algorithm based on a CNN and an end-to-end wireless mesh network based on BLE mesh.
In sound event recognition testing, the microsystem demonstrated a high level of accuracy and responsiveness. When tested with trained samples and the sound source within 6 m, the microsystem achieved an average recognition accuracy of 100% for alarm sounds greater than 70 dB and water flow sounds greater than 50 dB, with an average recognition time of no more than 3 s. When tested with untrained samples within 3 m of the sound source, the microsystem achieved an average recognition accuracy of 96.30% for alarm sounds greater than 70 dB and 92.50% for water flow sounds greater than 55 dB. In wireless mesh networking, the direct communication range between microsystems is 5 m, and this meshed network can be monitored and controlled with devices such as smartphones.
This intelligent microsystem and its wireless meshed network have made progress in responsive and accurate recognition of sound events, size reduction, power consumption reduction, topology flexibility increase, and device adaptability improvement. This work can be used as the sensing layer of IoT and applied to various intelligent environment applications, such as smart homes and unattended surveillance networks. The presented intelligent microsystem is a promising solution for sound event recognition and wireless networking in intelligent environments.

Author Contributions

Conceptualization, L.H., W.D. and J.Z.; Methodology, L.H., W.D. and J.Z.; Software, L.H.; Validation, L.H. and G.X.; Formal analysis, L.H. and W.D.; Investigation, L.H.; Resources, G.X., S.X., Y.L. (Yuan Li), Y.L. (Yizheng Li) and J.Z.; Data curation, L.H.; Writing—original draft, L.H.; Writing—review & editing, L.H. and W.D.; Visualization, L.H. and W.D.; Supervision, W.D. and J.Z.; Project administration, S.X., Y.L. (Yuan Li), Y.L. (Yizheng Li) and J.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are openly available at [https://doi.org/10.1145/2733373.2806390], reference number [18]. This data can be found here: [https://github.com/karolpiczak/ESC-50].

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Sisinni, E.; Saifullah, A.; Han, S.; Jennehag, U.; Gidlund, M. Industrial Internet of Things: Challenges, Opportunities, and Directions. IEEE Trans. Ind. Inform. 2018, 14, 4724–4734. [Google Scholar] [CrossRef]
  2. Farooqi, A.H.; Khan, F.A. Intrusion Detection Systems for Wireless Sensor Networks: A Survey. In Proceedings of the International Conference on Future Generation Communication and Networking/International Conference on Advanced Communication and Networking (FGCN/ACN 2009), Jeju Island, Republic of Korea, 10–12 December 2009; Springer: Berlin, Germany, 2009; pp. 234–241. [Google Scholar]
  3. Kumar, D.P.; Amgoth, T.; Annavarapu, C.S.R. Machine learning algorithms for wireless sensor networks: A survey. Inf. Fusion 2019, 49, 1–25. [Google Scholar]
  4. Quintana-Suarez, M.A.; Sanchez-Rodriguez, D.; Alonso-Gonzalez, I.; Alonso-Hernandez, J.B. A Low Cost Wireless Acoustic Sensor for Ambient Assisted Living Systems. Appl. Sci. 2017, 7, 877. [Google Scholar] [CrossRef] [Green Version]
  5. Ha, N.; Xu, K.; Ren, G.H.; Mitchell, A.; Ou, J.Z. Machine Learning-Enabled Smart Sensor Systems. Adv. Intell. Syst. 2020, 2, 31. [Google Scholar] [CrossRef]
  6. Niccolai, L.; Bassetto, M.; Quarta, A.A.; Mengali, G. A review of Smart Dust architecture, dynamics, and mission applications. Prog. Aeosp. Sci. 2019, 106, 1–14. [Google Scholar] [CrossRef]
  7. Shafique, K.; Khawaja, B.A.; Sabir, F.; Qazi, S.; Mustaqim, M. Internet of Things (IoT) for Next-Generation Smart Systems: A Review of Current Challenges, Future Trends and Prospects for Emerging 5G-IoT Scenarios. IEEE Access 2020, 8, 23022–23040. [Google Scholar] [CrossRef]
  8. Yang, Y.C.; Nagarajaiah, S. Output-only modal identification by compressed sensing: Non-uniform low-rate random sampling. Mech. Syst. Signal Process. 2015, 56–57, 15–34. [Google Scholar] [CrossRef]
  9. Yu, W.; Liang, F.; He, X.F.; Hatcher, W.G.; Lu, C.; Lin, J.; Yang, X.Y. A Survey on the Edge Computing for the Internet of Things. IEEE Access 2018, 6, 6900–6919. [Google Scholar] [CrossRef]
  10. Alias, F.; Alsina-Pages, R.M. Review of Wireless Acoustic Sensor Networks for Environmental Noise Monitoring in Smart Cities. J. Sens. 2019, 2019, 13. [Google Scholar] [CrossRef] [Green Version]
  11. Luo, L.Y.; Qin, H.M.; Song, X.Y.; Wang, M.; Qiu, H.B.; Zhou, Z. Wireless Sensor Networks for Noise Measurement and Acoustic Event Recognitions in Urban Environments. Sensors 2020, 20, 2093. [Google Scholar] [CrossRef] [Green Version]
  12. Arce, P.; Salvo, D.; Pinero, G.; Gonzalez, A. FIWARE based low-cost wireless acoustic sensor network for monitoring and classification of urban soundscape. Comput. Netw. 2021, 196, 10. [Google Scholar] [CrossRef]
  13. Caro-Via, S.; Vidana-Vila, E.; Ginovart-Panisello, G.J.; Martinez-Suquia, C.; Freixes, M.; Alsina-Pages, R.M. Edge-Computing Meshed Wireless Acoustic Sensor Network for Indoor Sound Monitoring. Sensors 2022, 22, 7032. [Google Scholar] [CrossRef] [PubMed]
  14. Yang, D.; Duan, W.R.; Xuan, G.Z.; Hou, L.L.; Zhang, Z.; Song, M.X.; Zhao, J.H. Self-Powered Long-Life Microsystem for Vibration Sensing and Target Recognition. Sensors 2022, 22, 9594. [Google Scholar] [CrossRef] [PubMed]
  15. Jeong, S.; Chen, Y.; Jang, T.; Tsai, J.M.L.; Blaauw, D.; Kim, H.S.; Sylvester, D. Always-On 12-nW Acoustic Sensing and Object Recognition Microsystem for Unattended Ground Sensor Nodes. IEEE J. Solid State Circuit 2018, 53, 261–274. [Google Scholar] [CrossRef]
  16. Oh, S.; Cho, M.C.; Wu, X.; Kim, Y.; Chuo, L.X.; Lim, W.; Pannuto, P.; Bang, S.; Yang, K.Y.; Kim, H.S.; et al. IoT(2) the Internet of Tiny Things: Realizing mm-Scale Sensors through 3D Die Stacking. In Proceedings of the 2019 Design, Automation & Test in Europe Conference & Exhibition (DATE), Florence, Italy, 25–29 March 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 686–691. [Google Scholar]
  17. Xiong, C.S.; Lu, W.S.; Zhao, X.G.; You, Z. Miniaturized multi-topology acoustic source localization network based on intelligent microsystem. Sens. Actuator A Phys. 2022, 345, 11. [Google Scholar] [CrossRef]
  18. Piczak, K.J. ESC: Dataset for Environmental Sound Classification. In Proceedings of the ACM International Conference on Multimedia (ACM Multimedia), Brisbane, QLD, Australia, 26–30 October 2015; Association for Computing Machinery: New York, NY, USA, 2015; pp. 1015–1018. [Google Scholar]
  19. Nanni, L.; Maguolo, G.; Paci, M. Data augmentation approaches for improving animal audio classification. Ecol. Inform. 2020, 57, 8. [Google Scholar] [CrossRef] [Green Version]
  20. Lella, K.K.; Pja, A. Automatic diagnosis of COVID-19 disease using deep convolutional neural network with multi-feature channel from respiratory sound data: Cough, voice, and breath. Alex. Eng. J. 2022, 61, 1319–1334. [Google Scholar] [CrossRef]
  21. Raj, S.; Prakasam, P.; Gupta, S. Multilayered convolutional neural network-based auto-CODEC for audio signal denoising using mel-frequency cepstral coefficients. Neural Comput. Appl. 2021, 33, 10199–10209. [Google Scholar] [CrossRef]
  22. Kong, Q.Q.; Cao, Y.; Iqbal, T.; Wang, Y.X.; Wang, W.W.; Plumbley, M.D. PANNs: Large-Scale Pretrained Audio Neural Networks for Audio Pattern Recognition. IEEE ACM Trans. Audio Speech Lang. 2020, 28, 2880–2894. [Google Scholar] [CrossRef]
  23. Zhang, H.M.; McLoughlin, I.; Song, Y. Robust Sound Event Recognition using Convolutional Neural Networks. In Proceedings of the 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), South Brisbane, QLD, Australia, 19–24 April 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 559–563. [Google Scholar]
  24. Li, S.B.; Yao, Y.; Hu, J.; Liu, G.K.; Yao, X.M.; Hu, J.J. An Ensemble Stacked Convolutional Neural Network Model for Environmental Event Sound Recognition. Appl. Sci. 2018, 8, 1152. [Google Scholar] [CrossRef] [Green Version]
  25. Capra, M.; Bussolino, B.; Marchisio, A.; Masera, G.; Martina, M.; Shafique, M. Hardware and Software Optimizations for Accelerating Deep Neural Networks: Survey of Current Trends, Challenges, and the Road Ahead. IEEE Access 2020, 8, 225134–225180. [Google Scholar] [CrossRef]
  26. Wu, J.X.; Leng, C.; Wang, Y.H.; Hu, Q.H.; Cheng, J. Quantized Convolutional Neural Networks for Mobile Devices. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 27–30 June 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 4820–4828. [Google Scholar]
  27. Ompal; Mishra, V.M.; Kumar, A. Zigbee Internode Communication and FPGA Synthesis Using Mesh, Star and Cluster Tree Topological Chip. Wirel. Pers. Commun. 2021, 119, 1321–1339. [Google Scholar] [CrossRef]
  28. Garcia-Sanchez, A.J.; Losilla, F.; Rodenas-Herraiz, D.; Cruz-Martinez, F.; Garcia-Sanchez, F. On the Feasibility of Wireless Multimedia Sensor Networks over IEEE 802.15.5 Mesh Topologies. Sensors 2016, 16, 643. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  29. Darroudi, S.M.; Gomez, C.; Crowcroft, J. Bluetooth Low Energy Mesh Networks: A Standards Perspective. IEEE Commun. Mag. 2020, 58, 95–101. [Google Scholar] [CrossRef]
  30. Murillo, Y.; Reynders, B.; Chiumento, A.; Malik, S.; Crombez, P.; Pollin, S. Bluetooth Now or Low Energy: Should BLE Mesh Become a Flooding or Connection Oriented Network? In Proceedings of the 28th Annual IEEE International Symposium on Personal, Indoor, and Mobile Radio Communications (PIMRC), Montreal, QC, Canada, 8–13 October 2017; IEEE: Piscataway, NJ, USA, 2017. [Google Scholar]
  31. Kim, H.S.; Lee, J.; Jang, J.W. BLEmesh: A Wireless Mesh Network Protocol for Bluetooth Low Energy Devices. In Proceedings of the 2015 3rd International Conference on Future Internet of Things and Cloud, Rome, Italy, 24–26 August 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 558–563. [Google Scholar]
  32. Rondon, R.; Mahmood, A.; Grimaldi, S.; Gidlund, M. Understanding the Performance of Bluetooth Mesh: Reliability, Delay, and Scalability Analysis. IEEE Internet Things J. 2020, 7, 2089–2101. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Overall working principle of the intelligent microsystem.
Figure 1. Overall working principle of the intelligent microsystem.
Sensors 23 03630 g001
Figure 2. Hardware structure and components of the intelligent microsystem. (a) The composition of the microsystem. (b) The main board composition. (c) The microsystem hardware integration and packaging.
Figure 2. Hardware structure and components of the intelligent microsystem. (a) The composition of the microsystem. (b) The main board composition. (c) The microsystem hardware integration and packaging.
Sensors 23 03630 g002
Figure 3. Sound event recognition algorithm and system workflow.
Figure 3. Sound event recognition algorithm and system workflow.
Sensors 23 03630 g003
Figure 4. CNN model training and testing. (a) Model training and validation accuracy and loss. (b) The confusion matrix in model testing.
Figure 4. CNN model training and testing. (a) Model training and validation accuracy and loss. (b) The confusion matrix in model testing.
Sensors 23 03630 g004
Figure 5. The BLE mesh architecture.
Figure 5. The BLE mesh architecture.
Sensors 23 03630 g005
Figure 6. Design and implementation of the end-to-end mesh networking.
Figure 6. Design and implementation of the end-to-end mesh networking.
Sensors 23 03630 g006
Figure 7. Overall experimental design.
Figure 7. Overall experimental design.
Sensors 23 03630 g007
Figure 8. Distribution of sound signals in the time domain and MFSC features.
Figure 8. Distribution of sound signals in the time domain and MFSC features.
Sensors 23 03630 g008
Figure 9. (a) Recognition results for the alarm sound. (b) Recognition results for the water flow sound. The red dashed line in the figure represents a recognition accuracy of 90%.
Figure 9. (a) Recognition results for the alarm sound. (b) Recognition results for the water flow sound. The red dashed line in the figure represents a recognition accuracy of 90%.
Sensors 23 03630 g009aSensors 23 03630 g009b
Figure 10. The interaction process between the smartphone and the microsystem sensing network.
Figure 10. The interaction process between the smartphone and the microsystem sensing network.
Sensors 23 03630 g010
Figure 11. (a) Alarm sound recognition results in Floor2-Room3. (b) Water flow sound recognition results in Floor1-Kitchen. The red dashed line in the figure represents a recognition accuracy of 90%.
Figure 11. (a) Alarm sound recognition results in Floor2-Room3. (b) Water flow sound recognition results in Floor1-Kitchen. The red dashed line in the figure represents a recognition accuracy of 90%.
Sensors 23 03630 g011
Table 1. Microsystem size.
Table 1. Microsystem size.
SizeCoreWithout HousingWith Housing
Volume (cm3)0.887.3614.14
Weight (g)4.008.0013.00
Table 2. The CNN model structure.
Table 2. The CNN model structure.
IDLayer TypeOutput ShapeComplexity (MACC)
1Input(30,32,1)N/A
2Conv2D(28,30,16)120,976
Nonlinearity(28,30,16)13,440
3BatchNormalization(28,30,16)N/A
4MaxPooling2d(9,6,16)12,960
5Conv2D(9,4,8)13,832
Nonlinearity(9,4,8)288
6Conv2D(7,2,8)8072
Nonlinearity(7,2,8)112
7BatchNormalization(7,2,8)N/A
8Flatten(1,1,112)N/A
9Dropout(1,1,112)N/A
10Dense(1,1,32)3616
Nonlinearity(1,1,32)32
11Dropout(1,1,32)N/A
12Dense(1,1,3)99
Nonlinearity(1,1,3)45
N/A refers to the layer whose complexity is negligible.
Table 3. Comparison between the original model and the quantified model.
Table 3. Comparison between the original model and the quantified model.
Model TypeAccuracy
(%)
Complexity
(MACC)
Ram Usage
(KB)
Flash Usage
(KB)
Original model95.24173,47212.7618.95
Quantified model94.99159,6063.584.93
Table 4. Results of the sound event recognition experiment.
Table 4. Results of the sound event recognition experiment.
Sound Event TypeAverage Recognition Accuracy (%)
0.1 m1 m2 m3 m4 m5 m6 mTotal
Alarm
(51~78 dB)
97.0084.0083.3378.0075.3370.3368.6779.52
Alarm
(70~78 dB)
100.00100.00100.00100.00100.00100.00100.00100.00
Flow
(41~58 dB)
96.0090.6787.0084.6781.6780.0079.3385.62
Flow
(50~58 dB)
100.00100.00100.00100.00100.00100.00100.00100.00
Table 5. Sample parameters of the networking recognition experiment.
Table 5. Sample parameters of the networking recognition experiment.
Sound Event TypeSample SizeSample
Duration (s)
Average Sound
Pressure Level (dB)
Sampling Rate (kHz)Channels
Alarm50551~7444.12
Flow50549~64482
Table 6. Results of the networking recognition experiment.
Table 6. Results of the networking recognition experiment.
Sound Event TypeAverage Recognition Accuracy (%)
0.1 m1 m2 m3 mTotal
Alarm
(51~74 dB)
94.3370.8359.3357.3370.46
Alarm
(70~74 dB)
100.0098.4094.0092.8096.30
Flow
(49~64 dB)
88.6785.6781.5074.5082.59
Flow
(56~64 dB)
94.0093.0092.0091.0092.50
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Hou, L.; Duan, W.; Xuan, G.; Xiao, S.; Li, Y.; Li, Y.; Zhao, J. Intelligent Microsystem for Sound Event Recognition in Edge Computing Using End-to-End Mesh Networking. Sensors 2023, 23, 3630. https://doi.org/10.3390/s23073630

AMA Style

Hou L, Duan W, Xuan G, Xiao S, Li Y, Li Y, Zhao J. Intelligent Microsystem for Sound Event Recognition in Edge Computing Using End-to-End Mesh Networking. Sensors. 2023; 23(7):3630. https://doi.org/10.3390/s23073630

Chicago/Turabian Style

Hou, Lulu, Wenrui Duan, Guozhe Xuan, Shanpeng Xiao, Yuan Li, Yizheng Li, and Jiahao Zhao. 2023. "Intelligent Microsystem for Sound Event Recognition in Edge Computing Using End-to-End Mesh Networking" Sensors 23, no. 7: 3630. https://doi.org/10.3390/s23073630

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop