Next Article in Journal
Designing a Simple Fiducial Marker for Localization in Spatial Scenes Using Neural Networks
Previous Article in Journal
Landmark-Assisted Compensation of User’s Body Shadowing on RSSI for Improved Indoor Localisation with Chest-Mounted Wearable Device
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Environmental Noise Classification with Inception-Dense Blocks for Hearing Aids

by
Po-Jung Ting
1,
Shanq-Jang Ruan
1 and
Lieber Po-Hung Li
2,3,4,5,*
1
Department of Electronic and Computer Engineering, National Taiwan University of Science and Technology, Taipei 106, Taiwan
2
Department of Otolaryngology, Cheng Hsin General Hospital, Taipei 112, Taiwan
3
Faculty of Medicine, and Institute of Brain Science, National Yang-Ming Chiao-Tung University, Taipei 112, Taiwan
4
Department of Medical Research, China Medical University Hospital, China Medical University, Taichung 404, Taiwan
5
Department of Speech Language Pathology and Audiology, College of Health Technology, National Taipei University of Nursing and Health Sciences, Taipei 112, Taiwan
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(16), 5406; https://doi.org/10.3390/s21165406
Submission received: 30 June 2021 / Revised: 2 August 2021 / Accepted: 7 August 2021 / Published: 10 August 2021
(This article belongs to the Section Wearables)

Abstract

:
Hearing aids are increasingly essential for people with hearing loss. For this purpose, environmental noise estimation and classification are some of the required technologies. However, some noise classifiers utilize multiple audio features, which cause intense computation. In addition, such noise classifiers employ inputs of different time lengths, which may affect classification performance. Thus, this paper proposes a model architecture for noise classification, and performs experiments with three different audio segment time lengths. The proposed model attains fewer floating-point operations and parameters by utilizing the log-scaled mel-spectrogram as an input feature. The proposed models are evaluated with classification accuracy, computational complexity, trainable parameters, and inference time on the UrbanSound8k dataset and HANS dataset. The experimental results showed that the proposed model outperforms other models on two datasets. Furthermore, compared with other models, the proposed model reduces model complexity and inference time while maintaining classification accuracy. As a result, the proposed noise classification for hearing aids offers less computational complexity without compromising performance.

1. Introduction

The number of people who have hearing loss has been increasing. Nevertheless, the number hearing aid users is not obviously increasing at the same rate [1]. According to the latest statistics from the World Health Organization (WHO), around 466 million people worldwide suffer from hearing loss, and only 17% use hearing aids [2]. Wearing hearing aids compensates for the hearing loss of hearing-impaired listeners, providing many benefits for them [3]. However, one of the common complaints of hearing aid users is the hearing aid’s inability to reduce ambient noise completely; instead, they cause ambient noise to be amplified along with the human voice [4]. The hearing aid’s inability to reduce ambient noise completely is due to their operation in various, irregular, and random environmental sound scenarios [5]. Speech enhancement algorithms are an essential technique for hearing aids. At present, deep neural network-based speech-enhancement methods have been widely adopted and have shown significant performance advantages over conventional speech enhancement techniques in complex noise environments. Several studies [6,7,8] have noted that using machine-learning-based environmental noise classification techniques to classify environmental noise, before using speech enhancement, can improve speech enhancement algorithm performance. Thus, the environmental noise classification (ENC) algorithms for hearing aids deserve attention.
Figure 1 shows the typical processing flow of environmental sound classification (ESC), consisting of two essential components: an acoustic feature extractor and a classifier. In order to extract useful acoustic features, audio signals are first pre-processed in the time domain, including normalization and silence reduction. Then, the audio signals are converted to the audio frames with a cosine window function (Hanning or Hamming window). Finally, features are extracted from each audio frame, and the extracted acoustic features are fed into the classifier for training and testing. The result of classification is the prediction probability for which audio signals belong to which type of sound. In previous works [9,10], traditional acoustic feature techniques, such as mel frequency cepstral coefficients (MFCC), linear predictive coding (LPC), and perceptual linear predictive coefficients (PLP) are already used in the ESC field. However, using more discriminative representations, such as mel filter bank features [11], wavelet-based features [12], and Gammatone features [13], can achieve better performance. Many typical machine learning algorithms have been used to classify sounds, such as support vector machine (SVM) [14], Gaussian mixture model (GMM) [15], and k-nearest neighbors (KNN) [16] for ESC tasks. However, the performance of these typical machine learning algorithms is still unsatisfactory. One reason is that typical classifiers cannot capture time and frequency features when applied to spectrogram-like input.
In recent years, deep neural networks (DNNs) have succeeded in ESC tasks and have provided better performance than typical machine learning algorithms [17,18]. For audio signals, DNNs can extract features from raw data or simple hand-crafted features. However, the deep fully-connected architecture of DNNs is not robust for transformative features. Several studies have shown that convolution neural networks (CNNs) can capture relevant features from an image. Furthermore, training CNN models with spectrogram-like features from environmental sounds demonstrates significant performance benefits over other methods [16,19,20]. However, existing high-accuracy CNNs also have a higher model complexity, leading to more computational complexity, which means more floating-point operations (FLOPs). A higher number of FLOPs is directly proportional to CPU consumption [21].
Although much work has been done on environmental sound classification, environmental noise classification for hearing aids is still critically lacking. The computational resources of hearing aids pose the major limitation that they cannot compute a large set of high-level features, and such features cannot be fed into a sophisticated environmental sound classifier [7]. One way to overcome these limited computational resources is to train the model on the server; the trained model can then be downloaded to hearing aids, which can then perform the environmental sound classification in real-time. However, the CNN model used in the ESC tasks mentioned above have a high computational complexity. Hence, reducing the computational model complexity while maintaining the model accuracy is essential.
This study attempts to provide a new model architecture for environmental noise classification on hearing aids that minimizes complexity by utilizing Inception-Dense block and depthwise-separable convolution. In order to verify the effectiveness of the proposed model, two index data analyses are performed: 1. FLOP counts are taken, representing the model’s computational complexity; 2. inference time—the running time for the model on a smartphone—is measured. For this objective to be achieved, the paper is structured as follows. Section 2 is a review of the literature, including the current research on feature extraction, environmental sound classification, and environmental noise classification for hearing aid. Section 3 provides a detailed introduction of proposed methods. In Section 4, the setting of the experiments on the UrbanSound8K dataset and the hearing aids noisy sound (HANS) dataset are described. Section 5 compares the performance of the proposed approach with previous models used in environmental noise classification and environmental sound classification. The conclusions and perspectives of future work are presented in the last section.

2. Related Work

This section first introduces the feature which will be used as the input of the environmental noise classifier. Then, the current research on environmental sound classification and environmental noise classification will be introduced.

2.1. Time-Frequency Representations for Noise Signal

Several types of desirable information are contained in raw audio data, such as the short-time Fourier transform (STFT) spectrogram, log-scaled mel-spectrogram, and gammatone spectrogram. This paper uses the log-scaled mel-spectrogram as the input of our proposed model. The reasoning for choosing this feature is stated below.
With a growing amount of evidence, the log-scaled mel-spectrogram outperforms other features in environmental sound classification tasks. Huzaifah [22] compared five signal processing methods, such as STFT-spectrogram, log-scaled mel-spectrogram, constant-Q transform (CQT), continuous wavelet transform (CWT), mel frequency cepstral coefficients (MFCCs), using these features as the input of two different models. The results indicate that the model that consisted of three convolution layers and two fully-connected layers, and which used the log-scaled mel-spectrogram as a feature, performed the best. Su et al. [23] resulted in the finding that the performance of the log-scaled mel-spectrogram is better than the gammatone spectrogram and the mel-spectrogram. Furthermore, and more essential, the log-scaled mel-spectrogram is computationally more efficient for real-time implementation than the CQT spectrogram, CWT scalogram, and MFCC.
The log-scaled mel-spectrogram generation process includes:
1.
Signal pre-processing.
2.
A Fourier transformation to obtain the signal spectrogram.
3.
Mapping of the spectrogram into a mel-spectrogram through triangular overlapping windows whose center frequencies are distributed on the mel scale. The function B for computing the mth mel-frequency from frequency f in Hertz and its inverse B 1 are given by [24]:
B ( f ) = 2595 log 10 ( 1 + f 700 ) ,
B 1 ( m ) = 700 ( 10 m 2595 1 ) .
4.
Taking a log calculation (decibles) on the mel spectrogram,
L M ( n ) = log ( k = 0 K H n ( k ) * F ( k ) 2 ) , n = 1 N ,
where H n ( k ) denotes the amplitude of the nth filter at frequency bin k, | F ( k ) | 2 denotes the FFT power spectrum.

2.2. Conventional Noise Classification Algorithms

Many conventional classification algorithms are used in noise classification. The KNN classification algorithm calculates the distance of the new input data from the k nearest points which determines the class of a new input data point. It is suitable for simple classification problems with basic training features. As the number of training features increases, KNNs computational complexity and time increases. SVM and Neural Networks [25] are feasible when there is a clear margin of separation between classes and are more effective in high dimensional spaces. However, SVM performs poorly when the data set has more noise, or the data set is too large. HMM is a widely used statistical method for speech recognition. One major advantage of HMMs over the previously described classifiers is that they account for the temporal statistics of different states’ occurrence in the features.
Nordqvist and Leijon [26] used the HMM and vector quantizer to classify three kinds of auditory environments (traffic noise environment, pure voice environment, and babble noise environment). Büchler et al. [27] selected characteristic parameters from the perspective of acoustic scene analysis to distinguish four different acoustic scenes (voice, noisy voice, noise, and music). Those characteristics have been evaluated together with different pattern classifiers. Simple classifiers, such as rule-based and minimum-distance classifiers, have been compared with more complex approaches, such as Bayes classifier, neural network, and HMM. Abe et al. [28] selected eleven features and used Bayes classifier, SVM, and Logistic regression to classify four kinds of auditory environments (speech, speech in noise, noise, and classical music). The conventional noise classification algorithms for hearing aids focused on voices and music. There is a paucity of conventional noise classification algorithms for hearing aids on the sound field outside of voices and music.

2.3. Deep Convolutional Neural Network

A CNN is a deep learning technology based on supervised learning and is widely used for image processing while maintaining the spatial information of the image. Recently, several deep learning methods for environmental sound classification have been conceived. Piczak [16] created a two-channel feature by applying the log-scaled mel-spectrogram and its delta information as the inputs of his CNN model, and the model achieved 73% accuracy on the UrbanSound8K dataset. Salamon and Bello [19] compared different data augmentations that could influence the accuracy of each class. They used the log-scaled mel-spectrogram as the input of the CNN model, and the accuracy was 79%. Zhang et al. [29] applied mixup and data augmentation to ESC tasks. They used the log-scaled mel-spectrograms and their delta information a as two-channel feature, and a similar CNN architecture to VGG net; the accuracy achieved was 82%. Palanisamy et al. [20] computed three different window sizes and hop lengths as three-channel features for the input to Inception [30], ResNet [31], and DenseNet [32]. The result showed that DenseNet is the best of the three.
Although there is much discussion on environmental sound classification based on CNNs in the literature, there is a paucity of noise classification algorithms for hearing aids based on CNNs. Singh and Joshi [33] used log-scaled mel-spectrograms as the input of a similar VGG net to classify background sound in a speech audio segment. Park and Lee [34] processed the spectrogram image through a sharpening mask and median filter, which was then used as the input of the CNN. However, the dataset is not public, and the model structure is not clear. Roedily et al. [35] used MFCC as the input of a CNN-LSTM model and an inference model on a smartphone. In order to evaluate the performance of our proposed model, the model is evaluated against the environmental sound classifications of other CNNs on the UrbanSound8k dataset, and is then evaluated against the noise classification of the Roedily and Singh models on the HANS dataset.

3. Proposed Methodology

3.1. Inception Block with Dense Connectivity

Figure 2 shows three blocks; the dense connectivity [32], Inception [30], and Inception-Dense blocks. The advantage of dense connectivity is that it bypasses connections. It can reuse feature maps from the previous layers. Figure 2a illustrates the layout of the dense connectivity block. The t h layer receives the feature-maps of all preceding layers, and x 0 , , x 1 are used as input to H ( . ) :
x = H ( [ x 0 , x 1 , , x 1 ] ) ,
where [ x 0 , x 1 , , x 1 ] refers to the concatenation of the feature-maps produced in layers 0 , , 1 . H ( . ) is a composite function of three consecutive operations: batch normalization (BN), followed by a rectified linear unit (ReLU) and a 3 × 3 convolution (Conv). However, various frequency bands and time intervals are important characteristics of the individual sound types. The advantage of Inception is that a combination of different kernel sizes can take multi-level feature maps. Using multi-level feature maps from multiple filters improves the performance of the network. Moreover, all the architectures prior to Inception performed convolution on the spatial and channel-wise domains. Figure 2b illustrates the layout of the Inception block. By performing the 1 × 1 convolution, the Inception block performs cross-channel correlations, ignoring spatial dimensions, followed by cross-spatial and cross-channel correlations via the 3 × 3 and 5 × 5 filters.
Therefore, dense connectivity is utilized to connect all Inception blocks on the proposed structure. As shown in Figure 2c, Inception blocks are connected by dense connectivity, allowing each block to receive input directly from its previous block. Similar to DenseNet, a layer in the proposed model implements a non-linear transformation F ( . ) , where is the index of the layer. F ( . ) is a composite operation function such as BN, ReLU, Conv, or Pool. The output of t h layer is denoted as x , which can be defined as :
x = c o n c a t e ( x 1 , F ( [ x b 1 , x b 2 , x b 3 , x b 4 ] ) ) ,
where x 1 is the output of 1 layer, x b 1 , x b 2 , x b 3 , x b 4 are the outputs of each branch in Inception block. [ x b 1 , x b 2 , x b 3 , x b 4 ] is a filter concatenation. It concatenates feature maps along the channel dimension.

3.2. Depthwise-Separable Convolution

Inspired by [36], the standard convolution layer is replaced with a depthwise-separable convolution layer on the Inception-Dense block. The depthwise-separable convolution layer can reduce the total number of operations. It combines a depthwise convolution layer and a 1 × 1 convolution called a pointwise convolution. The computational cost of a standard convolutional layer, a depthwise convolutional layer, and a pointwise convolutional layer is defined as (6)–(8), respectively. The depthwise-separable convolution layer could be simplified as (9), which is the sum of (7) and (8).
W i n * H i n * C i n * K * K * C o u t ,
W i n * H i n * C i n * K * K ,
W i n * H i n * C i n * C o u t ,
W i n * H i n * C i n * ( C o u t + K * K ) ,
where the computational cost depends multiplicatively on the number of input channels C i n , the number of output channels C o u t , the kernel size K × K , and the size of input feature map W i n × H i n . The depthwise-separable convolution layer can save 80 to 90 percent of the computation on the 3 × 3 convolution layer; on the other hand, the accuracy of the model is likely to decrease. Consequently, two types of Inception block with dense connectivity are used on the proposed model, as shown in Figure 3.

3.3. Network Structure

This section describes the proposed model in detail and shows that the model is compact. The proposed model architecture is comprised of two convolutional layers, three Inception-Dense blocks, and one fully-connected layer. The compared state-of-the-art noise classification network structure is based on VGG Net; the classification accuracy of this network is better than that of others [33]. The network structure is presented in Figure 4a. It stacks two convolutional layers as a module and stacks three or four modules into their network structure. In our model, we also use two initial convolutional layers. Then, the module of the convolutional layer is replaced by the Inception-Dense block; the Inception-Dense block is stacked three times in our network structure. However, considering DenseNet, we use a transition layer after three Inception-Dense blocks. Finally, the fully-connected layer is applied. The proposed network structure is presented in Figure 4b.
The detailed framework of the proposed model is shown in Figure 5. Batch Normalization (BN) is used to normalize the input data at first. Next, two 3 × 3 convolutional layers are used as a basic feature extractor, each convolutional layer followed by a BN and a Rectified Linear Unit (ReLU) activation function. After the second convolutional layer, a max-pooling layer is used to retain the most prominent features of the feature maps. The pool size of the max-pooling layer is 4 × 1 . Then, three Inception-Dense blocks are used to extract multiple features. The first block uses the Inception-Dense block A. Because of the channel size of the feature maps, all of the 1 × 1 convolutional layers use 16 filters, and all of the 3 × 3 convolutional layers use 32 filters in the first block. The second block and the third block use the Inception-Dense block B and the same settings. A BN and a ReLU activation function are followed by a convolutional layer on all Inception-Dense blocks. Next, the transition layer of DenseNet is considered, using a 1 × 1 convolutional layer and an average pooling layer, followed by a BN and a ReLU activation function.
Inspired by [37], instead of adding a fully-connected layer, a global average pooling layer is added at last, and the vector is fed directly into the softmax layer. The advantage of global average pooling over the fully-connected layer is that it is more native to the convolution structure by enforcing correspondences between feature maps and categories. Thus, the feature maps can be easily interpreted as categories confidence maps. The softmax function is used at the final layer to obtain class probabilities, and the chosen loss function is the cross-entropy loss function. The cross-entropy loss and softmax are used together because they provide a smooth and straightforward gradient, making computations much easier.
The configuration of the proposed model is described in Table 1. As an illustration, the 128 × 128 × 1 features are used as the input of the model at the configuration. The model is designed with full consideration of computational efficiency and practicality. The trained model has tiny size, only 1.6 MB. It can be applied to individual devices, including even those with limited computational resources.

4. Experiments

4.1. Dataset

The proposed model is evaluated on the following datasets: the UrbanSound8K [38] and the Hearing Aids Noisy Sound (HANS) dataset.
UrbanSound8K. The UrbanSound8k dataset contains 8732 clips (each of length less than or equal to 4 s). The original sampling rate of each audio clip varies from 16,000 Hz to 44,100 Hz. All audio clips are resampled to a sampling rate of 16,000 Hz, which is the appropriate frequency for signal processing for hearing aids. The dataset is officially split into ten folds and is divided into ten classes: air conditioner (ac), car horn (cr), children playing (ch), dog bark (db), engine idling (ei), gunshot (gs), jackhammer (jh), siren (si), and street music (sm). Note that the dataset is not rearranged.
Hearing Aids Noisy Sound (HANS). Many datasets are available publicly for environmental sounds, but there is no dedicated public dataset with a specific focus on hearing aid applications. Inspired by [5], we built a dataset called the Hearing Aid Noisy Sound (HANS) dataset; it contains sounds considered common and difficult for hearing-aids users. The top five categories of annoying sounds for hearing aids users are verbal human sounds, vehicle sounds, machine tools sounds, natural sounds, and household appliance sounds. The dataset considers the categories of annoying sounds in its description. It selects 5 classes and 10 classes from the UrbanSound8K dataset and the ESC50 [16] dataset, respectively. The 15 classes are separated into five major categories on the HANS dataset as shown in Figure 6. The UrbanSound8K and ESC50 datasets involve nonoverlapping, short clips of environmental sounds. The ESC50 dataset is a weak label dataset, and the length of each audio clip is equal to 5 s. The majority of the audio clip is empty; only a minor part with one sound is included. The Urbansound8k dataset is a strong label dataset. Its sounds are continuous, seldom empty and may contain multiple sounds which are not labeled, such as wind and people speaking. Some sound files on the UrbanSound8K dataset were removed because they conflict with the classes of the HANS dataset. Note that the class of human speech is the only one that contains sound files with speech. If a file in the class of street music only had music, or a file in the class of children playing included the sound of laughter, then these files were removed from the HANS dataset.
Because the UrbanSound8K and ESC50 datasets are small datasets, the official literature uses k-fold cross-validation to evaluate model performance. For that reason, the HANS dataset also uses k-fold validation to evaluate model performance. However, the UrbanSound8K dataset uses 10-fold cross-validation, and the ESC50 dataset uses 5-fold cross-validation. In order to reasonably combine two datasets, the HANS dataset uses 5-fold cross-validation. For using 5-fold cross-validation, two folds of the UrbanSound8K dataset are combined into one fold, such as combining the first fold and the second fold into the first fold of the HANS dataset and combining the third fold and the fourth fold into the second fold of the HANS dataset. The ESC50 dataset directly transferred the original fold to the new fold of the HANS dataset, i.e., by putting the first fold of the ESC50 dataset into the first fold of the HANS dataset. To maintain the appropriate frequency for signal processing on hearing aids, all audio clips are subsequently down-sampled to 16,000 Hz in the HANS dataset.

4.2. Data Preprocessing

Log-scaled mel-spectrograms are more efficient for real-time implementation than CQT spectrograms, CWT scalograms, and MFCCs. Regarding model performance, some studies [35,38] show that using MFCCs as input to the model is better than the log-scaled mel-spectrograms, and some studies [22,23] show the opposite. Consequently, a simple comparison experiment was carried out with log-scaled mel-spectrograms and MFCCs. The result showed that using the log-scaled mel spectrograms as the input of the proposed model provides better performance. Thus, as mentioned above, the proposed model uses log-scaled mel-spectrograms as the input in the following experiment.
The parameters of data processing used were the same on the UrbanSound8K dataset and the HANS dataset. The audio signals consist of 16,000 samples per second. Specifically, log-scaled mel-spectrograms were extracted by the Librosa [39]. First, short-time Fourier transform spectrograms were generated with the hamming window size of 512, hop length of 256, and covering the audible frequency range (300–16,000 Hz). Second, the spectrograms were mapped to the 128 mel filters (bands). Then, the resulting spectrograms were converted to a logarithmic scale (decibels). Finally, the spectrograms were divided into multiple data by frame length. The data frame lengths of 60 (approximately 1 s), 90 (approximately 1.5 s), and 128 (approximately 2 s) represent Set A, Set B, and Set C, respectively. Many studies use different frame lengths, which could affect the classification accuracy. Thus, we used statistics from those studies and experiment with three frame lengths.

4.3. Data Augmentation

Raw Data. In order to avoid overfitting and to effectively utilize the limited data on UrbanSound8K, pitch-shifting [19] and time-stretching [19] deformation methods were used for each audio sample in order to generate new audio samples. These augmentations were applied using the MUDA [40] library. As shown in Table 2, the augmentation data of Set A, Set B, and Set C were increased to 13.2, 13.2, and 14.7 times, respectively, as compared with the corresponding original data.
Because of the data imbalance problem in the HANS dataset, the pitch of each audio sample was shifted by a factor r. Next, the pitch-shifted sample was stretched for time by a random factor, to generate a new audio sample. The factor ranged from 0.9 to 1.2. However, the audio samples from the ESC50 dataset are fewer than those of from the UrbanSound8K dataset. Therefore, data augmentation was carried out with four sets of the relevant parameter in this research, as shown in Table 3. The differences between the four sets are the pitch-shifting factors and the amount of new audio samples. Table 4 presents the augmentation rules that apply to the four sets to balance the fold and class data. The augmentation was applied using the Librosa [39] library. The result after augmentation is illustrated in Figure 7.
Spectrogram. Mixup is a simple but effective method to generate training data [41], which we utilized. It uses mini-batch data which is selected from the whole of the original training data. Differently from traditional approaches, mixup generates new samples through the linear interpolation of two samples and their labels. New samples are determined by
x ˜ = λ x i + ( 1 λ ) x j , y ˜ = λ y i + ( 1 λ ) y j ,
where ( x i , y i ) and ( x j , y j ) are two samples randomly selected from training data. x represents raw input samples, y represents a one-hot label. The mix factor λ is a hyper-parameter α and λ ∼ Beta( α , α ). Note that mixup was only used for the training phase.

4.4. Training Settings

For the training stage, the Adam optimizer was used to train all models. The learning rate was initialized as 0.01 and reset as 0.01 when the learning rate was lower than 0.00001. The learning rate was automatically decreased by a factor of 0.5 when the validation loss stopped improving; after seven epochs for the UrbanSound8K dataset and five epochs for the HANS dataset. Every batch consisted of 32 samples randomly selected from the training set without repetition. The models were trained for 100 epochs and 50 epochs for UrbanSound8K and HANS, respectively.
For the testing stage, feature extraction and audio cropping patterns were the same as those used in the training stage. The k-fold cross-validation was used to evaluate the classification performance of the methods. For the UrbanSound8K dataset, k was set as 10 to produce a fair comparison with the results reported by Salmon et al. [38]. For the HANS dataset, k was set as 5 to validate mean accuracy.
All models were trained using the Keras [42] library with the TensorFlow backend on Nvidia GeForce RTX2080 with 16GB RAM. The inference time of all models was measured on the android mobile device Pixel 3a, and the models used TensorFlow Lite (TFLite). This mobile device possesses a Qualcomm Snapdragon 670 64-bit ARM-based octa-core system on a chip (SoC). The clock speed of this CPU varies between 1.7 and 2.0 GHz depending on the core being used. The internal memory of this smartphone is 4GB LPDDR4x RAM. It also possesses an Adreno 614 GPU. Note that TensorFlow for Mobile does not utilize this GPU.

5. Results

In this section, the performance of the proposed model and its classification results are shown. First, the performance of the proposed model on four conditions of the UrbanSound8k dataset are introduced. We provide comprehensive comparisons with other existing models [16,19,20,29] including the accuracy, the number of parameters, and total FLOPs. Afterward, the performances of the proposed model and three existing models [29,33,35] on the HANS dataset for noise classification is examined. We also provide various other comparisons with those models, including the accuracy, the number of parameters, total FLOPs, and inference time. Note that the android mobile device is used to measure the inference time.

5.1. Classification Results on Urbansound8k

Figure 8 shows the classification accuracy of the proposed model. The values in the figure are the average classification accuracy (%), which represent the rate of correct predictions on the total set of input data. MO represents the proposed model on the original data without data augmentation; MM represents the proposed model using the mixup technique on the original dataset; MA represents the proposed model on the augmentation dataset; MMA represents the proposed model using the mixup technique on the augmentation dataset. Figure 8a displays the classification results of Set A under four conditions. Comparing the accuracy of MO and other conditions, MM, MA, and MMA improvements are up to 2.28%, 2.49%, and 4.24%, respectively. Figure 8b displays the classification results of Set B under four conditions. Comparing the accuracy between MO and other conditions, MM, MA, and MMA improvements are up to 2.96%, 3.12%, and 5.27%, respectively. Figure 8c displays the classification results of Set C under four conditions. Comparing the accuracy between MO to other conditions—MM, MA, and MMA—the improvements are up to 2.81%, 3.61%, and 6.82%, respectively. Thus, data augmentation is shown to be an important technique for increasing the performance on a limited dataset. Moreover, the mixup technique is a powerful way to improve the classification accuracy in the training phase, resulting in better performance than for the model without mixup. Furthermore, when the two techniques are used simultaneously, the performance of the proposed model is at its best.
Figure 8d shows the average classification accuracy of three sets under four conditions. Set C produces the highest classification accuracy in the time division compared to other sets.With the increasing frame length of the spectrogram, the classification accuracy increases. Although greater sound length increases the number of frames, it provides more information about the environmental sound.
The mean classification accuracy of the proposed model compared with other existing models [16,19,20,29] is shown in Table 5. It can be observed that our method achieves a mean accuracy of 83.03% on UrbanSound8K. The mean classification accuracy of our method outperforms PizcakCNN [16] (baseline) and SBCNN [19] by 10.33% and 4.06%, respectively. In comparison with the ZhangCNN [29], the mean accuracy of the proposed model only increases 0.43%. However, the parameters and FLOPs of the proposed model are much lower than for ZhangCNN, indicating that the proposed model has significant advantages in computational efficiency over other models. Moreover, the mean classification accuracy of our method outperformed ResNet, Inception, and DenseNet by 9.77%, 7.79%, and 6.73%, respectively. The results show that our model, which uses the Inception-Dense block, is effective.

5.2. Classification Results on the Hearing Aids Noisy Sound (HANS) Dataset

The mean classification accuracy of the proposed model and three existing models [29,33,35] is presented in Figure 9 for noise classification. It can be observed that the proposed model is superior to other models due to its higher average classification accuracy. Figure 9a shows that Set C produces the highest classification accuracy compared to the other sets. The mean classification accuracy of Set A, Set B, and Set C can achieve the best precision by using the proposed model on the HANS dataset. The accuracies of Set A, Set B, and Set C are 72.73%,73.22%, and 74.03%, respectively. As a result, the proposed model is more effective than other models. Figure 9b shows that all models using the mixup technique significantly improved their accuracy. Compared with other models, the proposed model achieves the best accuracy in every Set. The improvements on Set A, Set B, and Set C are up to 0.47%, 1.62%, and 1.24%, respectively. Thus, we can state that the mixup technique is helpful on the HANS dataset.
Table 6 presents the performance of different architectures on the HANS dataset. It is apparent that the proposed model has the best accuracy and the fewest parameters among all models. The FLOPs of the proposed model are the second least of all models. Although the Roedily model has fewer FLOPs than those of our model, the accuracy of our model is much higher. Therefore, the proposed model can use fewer parameters and FLOPs while achieving better accuracy and inference time.

6. Conclusions

In this article, we present a specially designed network for accurately recognizing urban and noise sounds. The proposed model aims to take full advantage of the low-level information in log-scaled mel-spectrograms to make its classification decisions. This architecture is shown to be competitive with other deep learning architectures evaluated on the UrbanSound8K dataset and our HANS dataset. Moreover, the input spectrograms are generated using three different frame lengths. We found that the classification accuracy of the proposed model gradually increased with an increased frame length of the input spectrogram. Thus, we can deduce that more information can be obtained due to the longer audio frame length.
Additionally, the mixup technique is used to increase data diversity in the training phase. The experimental results showed that the mixup technique could enhance classification accuracy. The proposed model performed better than other models on two datasets, and the inference time of the proposed model is short. Despite the proposed model’s performance being competitive, there are limitations to improving the accuracy of the proposed model on the HANS dataset. Thus, we plan to use different input-transform methods and other data-augmentation methods to further improve the proposed model in our future work. The proposed noise classifier will be used as part of a noise reduction app for hearing-improvement purposes.

Author Contributions

Conceptualization, P.-J.T.; methodology, P.-J.T.; software, P.-J.T.; validation, P.-J.T.; formal analysis, P.-J.T.; investigation, P.-J.T.; resources, L.P.-H.L.; data curation, P.-J.T.; writing—original draft preparation, P.-J.T.; writing—review and editing, S.-J.R.; visualization, P.-J.T.; supervision, S.-J.R.and L.P.-H.L.; project administration, P.-J.T.; funding acquisition, L.P.-H.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Ministry of Science and Technology of the Republic of China under the project “Deep Learning-Based Noise Reduction Model Design on Hearing Aids-Application of MobileNet and signal compression” (Grant No. MOST 108-2314-B-350-002-MY2).

Institutional Review Board Statement

This study conformed to the Declaration of Helsinki, and was approved by the Institutional Ethics and Research Committee of Cheng-Hsin General Hospital.

Informed Consent Statement

Informed consent was non-applicable in the study.

Data Availability Statement

The data presented in this study are available in this published article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Löhler, J.; Walther, L.E.; Hansen, F.; Kapp, P.; Meerpohl, J.; Wollenberg, B.; Schönweiler, R.; Schmucker, C. The prevalence of hearing loss and use of hearing aids among adults in Germany: A systematic review. Eur. Arch. Oto-Rhino 2019, 276, 945–956. [Google Scholar] [CrossRef] [Green Version]
  2. World Health Organization. Deafness and Hearing Loss. 2020. Available online: https://www.who.int/news-room/fact-sheets/detail/deafness-and-hearing-loss (accessed on 26 April 2021).
  3. Mulrow, C.D.; Tuley, M.R.; Aguilar, C. Sustained benefits of hearing aids. J. Speech Lang. Hear. Res. 1992, 35, 1402–1405. [Google Scholar] [CrossRef]
  4. Vestergaard Knudsen, L.; Öberg, M.; Nielsen, C.; Naylor, G.; Kramer, S.E. Factors influencing help seeking, hearing aid uptake, hearing aid use and satisfaction with hearing aids: A review of the literature. Trends Amplif. 2010, 14, 127–154. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Skagerstrand, A.; Stenfelt, S.; Arlinger, S.; Wikström, J. Sounds perceived as annoying by hearing-aid users in their daily soundscape. Int. J. Audiol. 2014, 53, 259–269. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Shi, W.; Zhang, X.; Zou, X.; Han, W. Deep neural network and noise classification-based speech enhancement. Mod. Phys. Lett. B 2017, 31, 1740096. [Google Scholar] [CrossRef]
  7. Park, G.; Cho, W.; Kim, K.S.; Lee, S. Speech Enhancement for Hearing Aids with Deep Learning on Environmental Noises. Appl. Sci. 2020, 10, 6077. [Google Scholar]
  8. Zhang, L.; Wang, M.; Zhang, Q.; Liu, M. Environmental Attention-Guided Branchy Neural Network for Speech Enhancement. Appl. Sci. 2020, 10, 1167. [Google Scholar] [CrossRef] [Green Version]
  9. Lee, K.; Ellis, D.P. Audio-based semantic concept classification for consumer video. IEEE Trans. Audio Speech Lang. Process. 2009, 18, 1406–1416. [Google Scholar] [CrossRef] [Green Version]
  10. Temko, A.; Monte, E.; Nadeu, C. Comparison of sequence discriminant support vector machines for acoustic event classification. In Proceedings of the 2006 IEEE International Conference on Acoustics Speech and Signal Processing Proceedings, Toulouse, France, 14–19 May 2006; Volume 5, p. V. [Google Scholar]
  11. Chu, S.; Narayanan, S.; Kuo, C.C.J. Environmental sound recognition with time frequency audio features. IEEE Trans. Audio Speech Lang. Process. 2009, 17, 1142–1158. [Google Scholar] [CrossRef]
  12. Geiger, J.T.; Helwani, K. Improving event detection for audio surveillance using gabor filterbank features. In Proceedings of the 2015 23rd European Signal Processing Conference (EUSIPCO), Nice, France, 31 August–4 September 2015; pp. 714–718. [Google Scholar]
  13. Valero, X.; Alias, F. Gammatone cepstral coefficients: Biologically inspired features for non-speech audio classification. IEEE Trans. Multimed. 2012, 14, 1684–1689. [Google Scholar] [CrossRef]
  14. Uzkent, B.; Barkana, B.D.; Cevikalp, H. Non-speech environmental sound classification using SVMs with a new set of features. Int. J. Innov. Comput. Inf. Control 2012, 8, 3511–3524. [Google Scholar]
  15. Stowell, D.; Giannoulis, D.; Benetos, E.; Lagrange, M.; Plumbley, M.D. Detection and classification of acoustic scenes and events. IEEE Trans. Multimed. 2015, 17, 1733–1746. [Google Scholar] [CrossRef]
  16. Piczak, K.J. ESC: Dataset for environmental sound classification. In Proceedings of the 23rd ACM International Conference on Multimedia, Brisbane, Australia, 26–30 October 2015; pp. 1015–1018. [Google Scholar]
  17. Kons, Z.; Toledo-Ronen, O.; Carmel, M. Audio Event Classification Using Deep Neural Networks. Interspeech. 2013, pp. 1482–1486. Available online: https://www.isca-speech.org/archive/archive_papers/interspeech_2013/i13_1482.pdf (accessed on 21 December 2020).
  18. McLoughlin, I.; Zhang, H.; Xie, Z.; Song, Y.; Xiao, W. Robust sound event classification using deep neural networks. IEEE/ACM Trans. Audio Speech Lang. Process. 2015, 23, 540–552. [Google Scholar] [CrossRef] [Green Version]
  19. Salamon, J.; Bello, J.P. Deep convolutional neural networks and data augmentation for environmental sound classification. IEEE Signal Process. Lett. 2017, 24, 279–283. [Google Scholar] [CrossRef]
  20. Palanisamy, K.; Singhania, D.; Yao, A. Rethinking cnn models for audio classification. arXiv 2020, arXiv:2007.11154. [Google Scholar]
  21. Sehgal, A.; Kehtarnavaz, N. Guidelines and benchmarks for deployment of deep learning models on smartphones as real-time apps. Mach. Learn. Knowl. Extr. 2019, 1, 450–465. [Google Scholar] [CrossRef] [Green Version]
  22. Huzaifah, M. Comparison of time-frequency representations for environmental sound classification using convolutional neural networks. arXiv 2017, arXiv:1706.07156. [Google Scholar]
  23. Su, Y.; Zhang, K.; Wang, J.; Zhou, D.; Madani, K. Performance analysis of multiple aggregated acoustic features for environment sound classification. Appl. Acoust. 2020, 158, 107050. [Google Scholar] [CrossRef]
  24. Stevens, S.S.; Volkmann, J.; Newman, E.B. A scale for the measurement of the psychological magnitude pitch. J. Acoust. Soc. Am. 1937, 8, 185–190. [Google Scholar] [CrossRef]
  25. Hagan, M.T.; Demuth, H.B.; Beale, M. Neural Network Design; PWS Publishing Co.: Boston, MA, USA, 1997. [Google Scholar]
  26. Nordqvist, P.; Leijon, A. An efficient robust sound classification algorithm for hearing aids. J. Acoust. Soc. Am. 2004, 115, 3033–3041. [Google Scholar] [CrossRef]
  27. Büchler, M.; Allegro, S.; Launer, S.; Dillier, N. Sound classification in hearing aids inspired by auditory scene analysis. EURASIP J. Adv. Signal Process. 2005, 2005, 387845. [Google Scholar] [CrossRef] [Green Version]
  28. Abe, K.; Sakaue, H.; Okuno, T.; Terada, K. Sound classification for hearing aids using time-frequency images. In Proceedings of the 2011 IEEE Pacific Rim Conference on Communications, Computers and Signal Processing, Victoria, BC, Canada, 23–26 August 2011; pp. 719–724. [Google Scholar]
  29. Zhang, Z.; Xu, S.; Cao, S.; Zhang, S. Deep convolutional neural network with mixup for environmental sound classification. In Chinese Conference on Pattern Recognition and Computer Vision (PRCV); Springer: Cham, Switzerland, 2018; pp. 356–367. [Google Scholar]
  30. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  31. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  32. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
  33. Singh, J.; Joshi, R. Background Sound Classification in Speech Audio Segments. In Proceedings of the 2019 International Conference on Speech Technology and Human-Computer Dialogue (SpeD), Timisoara, Romania, 10–12 October 2019; pp. 1–6. [Google Scholar]
  34. Park, G.; Lee, S. Environmental Noise Classification Using Convolutional Neural Networks with Input Transform for Hearing Aids. Int. J. Environ. Res. Public Health 2020, 17, 2270. [Google Scholar] [CrossRef] [Green Version]
  35. Roedily, W.; Ruan, S.J.; Li, L.P.H. Real-Time Noise Classifier on Smartphones. IEEE Consum. Electron. Mag. 2020, 10, 37–42. [Google Scholar] [CrossRef]
  36. Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv 2017, arXiv:1704.04861. [Google Scholar]
  37. Lin, M.; Chen, Q.; Yan, S. Network in network. arXiv 2013, arXiv:1312.4400. [Google Scholar]
  38. Salamon, J.; Jacoby, C.; Bello, J.P. A dataset and taxonomy for urban sound research. In Proceedings of the 22nd ACM International Conference on Multimedia, Orlando, FL, USA, 3–7 November 2014; pp. 1041–1044. [Google Scholar]
  39. McFee, B.; Raffel, C.; Liang, D.; Ellis, D.P.; McVicar, M.; Battenberg, E.; Nieto, O. librosa: Audio and music signal analysis in python. In Proceedings of the 14th Python in Science Conference, Austin, TX, USA, 6–12 July 2015; Volume 8, pp. 18–25. [Google Scholar]
  40. McFee, B.; Humphrey, E.; Bello, J. A software framework for musical data augmentation. In Proceedings of the 16th International Society for Music Information Retrieval Conference, Malaga, Spain, 26–30 October 2015. [Google Scholar]
  41. Zhang, H.; Cisse, M.; Dauphin, Y.N.; Lopez-Paz, D. mixup: Beyond empirical risk minimization. arXiv 2017, arXiv:1710.09412. [Google Scholar]
  42. Chollet, F. Keras. 2015. Available online: https://keras.io (accessed on 5 April 2021).
Figure 1. Typical processing flow of an environmental sound classification system.
Figure 1. Typical processing flow of an environmental sound classification system.
Sensors 21 05406 g001
Figure 2. (a) Dense connectivity. (b) Inception block. (c) Inception-Dense block. (a,b) have two common points: the first is that they all use 1x1 convolution as a BottleNeck layer; the second is that they all use filter concatenation connections. The difference between (a,b) is that the input tensor directly passes to the output tensor without any other middle operations in (a).
Figure 2. (a) Dense connectivity. (b) Inception block. (c) Inception-Dense block. (a,b) have two common points: the first is that they all use 1x1 convolution as a BottleNeck layer; the second is that they all use filter concatenation connections. The difference between (a,b) is that the input tensor directly passes to the output tensor without any other middle operations in (a).
Sensors 21 05406 g002
Figure 3. Two kinds of basic Inception-Dense blocks in our model. The differences between the blocks (a,b): block (a) is the traditional Inception block, while block (b) replaces the 3 × 3 convolution layer with the 3 × 3 depthwise-separable convolution layer.
Figure 3. Two kinds of basic Inception-Dense blocks in our model. The differences between the blocks (a,b): block (a) is the traditional Inception block, while block (b) replaces the 3 × 3 convolution layer with the 3 × 3 depthwise-separable convolution layer.
Sensors 21 05406 g003
Figure 4. The network structures. (a) The state-of-the-art network structure based on VGG Net [33]. (b) The proposed network structure.
Figure 4. The network structures. (a) The state-of-the-art network structure based on VGG Net [33]. (b) The proposed network structure.
Sensors 21 05406 g004
Figure 5. The overall framework of the proposed model. The purple block represents a 3 × 3 convolution, and the blue block represents a 1 × 1 convolution. The grey block represents a max-pooling layer, an average pooling layer, or a global average pooling layer. Here only one type of layer on a grey block is present.
Figure 5. The overall framework of the proposed model. The purple block represents a 3 × 3 convolution, and the blue block represents a 1 × 1 convolution. The grey block represents a max-pooling layer, an average pooling layer, or a global average pooling layer. Here only one type of layer on a grey block is present.
Sensors 21 05406 g005
Figure 6. For the hearing aids application, 15 subclasses are grouped into five major categories, and each category is described in detail. A yellow block indicates that the subclass is selected from the UrbanSound8k dataset. A blue block represents that the subclass is selected from the ESC50 dataset.
Figure 6. For the hearing aids application, 15 subclasses are grouped into five major categories, and each category is described in detail. A yellow block indicates that the subclass is selected from the UrbanSound8k dataset. A blue block represents that the subclass is selected from the ESC50 dataset.
Sensors 21 05406 g006
Figure 7. (a) The frame lengths of 60. (b) The frame lengths of 90. (c) The frame lengths of 128. (d) The statistic data of the three sets, shown distributed into five folds.
Figure 7. (a) The frame lengths of 60. (b) The frame lengths of 90. (c) The frame lengths of 128. (d) The statistic data of the three sets, shown distributed into five folds.
Sensors 21 05406 g007
Figure 8. The classification results under four conditions of MO, MM, MA, and MMA and three sets Set A, Set B, and Set C. (a) The classification results of Set A. (b) The classification results of Set B. (c) The classification results of Set C. (d) The classification results of three sets.
Figure 8. The classification results under four conditions of MO, MM, MA, and MMA and three sets Set A, Set B, and Set C. (a) The classification results of Set A. (b) The classification results of Set B. (c) The classification results of Set C. (d) The classification results of three sets.
Sensors 21 05406 g008
Figure 9. (a) The mean classification accuracy of models on the HANS dataset (b) The mean classification accuracy of models on the HANS dataset with the mixup technique. The frame length of Set A is 60, Set B is 90, and Set C is 128. The mean accuracy is the average score when performing 5-fold cross-validation.
Figure 9. (a) The mean classification accuracy of models on the HANS dataset (b) The mean classification accuracy of models on the HANS dataset with the mixup technique. The frame length of Set A is 60, Set B is 90, and Set C is 128. The mean accuracy is the average score when performing 5-fold cross-validation.
Sensors 21 05406 g009
Table 1. The configuration of the proposed model. Out Shape represents the dimension in ( f r a m e , m e l f e a t u r e s , c h a n n e l ) . The model only has 97 k parameters.
Table 1. The configuration of the proposed model. Out Shape represents the dimension in ( f r a m e , m e l f e a t u r e s , c h a n n e l ) . The model only has 97 k parameters.
LayerFilter Size/Stride (Number of Filters)Out ShapeParams
Input(128, 128, 1)
Conv1 3 × 3 /1 (16)(128, 128, 16)228
Conv2 3 × 3 /1 (16)(128, 128, 16)2,384
MaxPool1 3 × 3 /1 (None)(42, 42, 16)
Inception (a)[ 1 × 1 /1(16)conv]*1,
[ 1 × 1 /1(16)conv, 3 × 3 /1(32)conv]*1,
[ 1 × 1 /1(16)conv, 3 × 3 /1(32)conv, 3 × 3 /1(32)]*1,
[ 3 × 3 /1(None)avergepool, 1 × 1 /1(16)conv]
(42, 42, 112)20,128
Inception (b)[ 1 × 1 /1(32)conv]*1,
[ 1 × 1 /1(32)conv, 3 × 3 /1(32)conv]*1,
[ 1 × 1 /1(32)conv, 3 × 3 /1(32)conv, 3x3/1(32)conv]*1,
[ 3 × 3 /1(None)avergepool, 1 × 1 /1(32)conv]
(42, 42, 240)21,152
Inception (c)[ 1 × 1 /1(32)conv]*1,
[ 1 × 1 /1(32)conv, 3 × 3 /1(32)conv]*1,
[ 1 × 1 /1(32)conv, 3 × 1 /1(32)conv,
3 × 3 /1(32)conv]*1,
[ 3 × 3 /1(None)avergepool, 1 × 1 /1(32)conv]
(42, 42, 368)39,584
Conv3 1 × 1 /1(32)(42, 42, 32)13,408
AveragePool 2 × 2 / 2 × 2 (None)(42, 10, 32)
GlobalAvergePool1(32)
Densenumber of classesnumber of classes330
Total Params 97,214
Table 2. The UrbanSound8K dataset is divided into three sets by the original data and the augmentation data.
Table 2. The UrbanSound8K dataset is divided into three sets by the original data and the augmentation data.
FoldUrbanSound8KUrbanSound8k (aug)
SetASetBSetCSetASetBSetC
15238261987369,59334,79212,841
25328266488870,74035,37212,964
35550277592573,76636,88513,528
45940297099078,85339,42914,390
55616280893674,55637,28013,555
64938246982365,59932,80011,966
75028251483866,80433,40512,250
84836241880664,21132,11411,964
94896244881665,03532,51912,118
105022251183766,69833,35012,376
Table 3. The description of four groups.
Table 3. The description of four groups.
GroupSampler
1.5Even2
2Each2
5Each−2.5, −2, −1, 1, 2
6Each−2.5, −2, −1, 1, 2, 2.5
Table 4. The HANS dataset augmentation rule.
Table 4. The HANS dataset augmentation rule.
Major CategorieClassFold 1Fold 2Fold 3Fold 4Fold 5
Vehicle engine (ve)Helicopter66666
Airplane66666
Train66666
Machine tools (ma)DrillingNoneNoneNone21.5
Jackhammer2NoneNone1.5None
Chainsaw66666
Household appliance (ha)Air conditionNone1.5NoneNone1.5
Vacuum cleaner66666
Washing machine66666
Natural (na)Thunderstorm55555
Sea waves55555
Rain55555
Wind55555
Human speech (hu)Children playingNoneNoneNoneNoneNone
Street music1.5NoneNone22
Table 5. Comparison of classification accuracy with other models on UrbanSound8K dataset. The bold number is our result.
Table 5. Comparison of classification accuracy with other models on UrbanSound8K dataset. The bold number is our result.
ApproachMean AccParametersFLOPs
PiczakCNN [16]73.09109,134,090515,806,780
ResNet [20]73.2623,608,2025,044,643,516
Inception [20]75.2421,823,2743,196,935,548
DenseNet [20]76.3018,341,1945,367,216,252
SBCNN [19]79.00874,746170,694,732
ZhangCNN [29]82.601,186,322882,779,336
Proposed model83.0397,214394,483,170
Table 6. Comparison of the four models on HANS dataset. The bold is our result.
Table 6. Comparison of the four models on HANS dataset. The bold is our result.
ApproachParamSetAcc (%)FLOPsInference Time (s)
ModelFLM
Roediy [35]116,869A57.706,184,6060.0120.033
B57.489,234,5900.0230.045
C58.7213,191,7100.0360.057
Zhang [29]1,183,081A71.29377,229,4760.0640.084
B72.56582,501,7960.0930.112
C71.01860,721,3160.1300.146
Singh [33] 4,694,473A72.533,072,945,4440.2640.283
B73.784,637,731,6200.3910.408
C74.756,676,496,6760.5490.565
Proposed model97,049A73.20187,247,7480.0320.061
B74.84280,86,43880.0490.067
C75.27394,482,8200.0670.084
FLM: First Load tflite Model on mobile device.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Ting, P.-J.; Ruan, S.-J.; Li, L.P.-H. Environmental Noise Classification with Inception-Dense Blocks for Hearing Aids. Sensors 2021, 21, 5406. https://doi.org/10.3390/s21165406

AMA Style

Ting P-J, Ruan S-J, Li LP-H. Environmental Noise Classification with Inception-Dense Blocks for Hearing Aids. Sensors. 2021; 21(16):5406. https://doi.org/10.3390/s21165406

Chicago/Turabian Style

Ting, Po-Jung, Shanq-Jang Ruan, and Lieber Po-Hung Li. 2021. "Environmental Noise Classification with Inception-Dense Blocks for Hearing Aids" Sensors 21, no. 16: 5406. https://doi.org/10.3390/s21165406

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop