Next Article in Journal
Decoupling Capacitors Placement at Board Level Adopting a Nature-Inspired Algorithm
Previous Article in Journal
A Mechanism of Masking Identification Information regarding Moving Objects Recorded on Visual Surveillance Systems by Differentially Implementing Access Permission
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Machine Learning Approach to Fitting Prescription for Hearing Aids

Department of Electronic Engineering, Inha University, Incheon 22212, Korea
*
Author to whom correspondence should be addressed.
Electronics 2019, 8(7), 736; https://doi.org/10.3390/electronics8070736
Submission received: 8 May 2019 / Revised: 2 June 2019 / Accepted: 24 June 2019 / Published: 28 June 2019
(This article belongs to the Section Artificial Intelligence)

Abstract

:
A successful Hearing-Aid Fitting (HAF) is more than just selecting an appropriate Hearing Aid (HA) device for a patient with Hearing Loss (HL). The initial fitting is given by the prescription based on user’s hearing loss; however, it is often necessary for the audiologist to readjust some parameters to satisfy the user demands. Therefore, in this paper, we concentrated on a new application of Neural Network (NN) combined with a Transfer Learning (TL) strategy to develop a fitting algorithm with the prescription database for hearing loss and readjusted gain to minimize the gap between fitting satisfaction. As prior information, we generated the data set from two popular hearing-aid fitting software, then fed the training data to our proposed model, and verified the performance of the architecture. Pondering real life circumstances, where numerous fitting records may not always be accessible, we first investigated the number of minimum fitting records required for possible sufficient training. After that, we evaluated the performance of the proposed algorithm in two phases: (a) NN with refined hyper parameter showed enhanced performance in compare to state-of-the-art DNN approach, and (b) the TL approach boosted the performance of the NN algorithm in a broad way. Altogether, our model provides a pragmatic and promising tool for HAF.

1. Introduction

Hearing-Aid Fitting (HAF) formulas estimate the required electro-acoustic amplification for individuals with hearing loss [1]. The National Acoustic Laboratories (NAL) procedures (NAL-NL1 [2,3]; NAL-NL2 [4,5]) and the Desired Sensation Level (DSL) [6] procedures (DSL v4.1 [7] and DSL m[i/o] [8]) have been used extensively for prescribing hearing-aid gain in recent time.
NAL-NL1 [2,3], the first prescription fitting procedure for prescribing nonlinear gain was developed by National Acoustic Laboratories (NAL) in 1999. The purely theoretical fitting formula aims at maximizing speech intelligibility while ensuring the overall loudness does not exceed the loudness that perceived by a normal-hearing person. Like its progenitor NAL-NL1, the second generation prescription procedure of NAL-NL2 [4,5] also aims to make speech intelligible and overall loudness comfortable. However, the theoretical derivation of NAL-NL2 differs from that of NAL-NL1 in two major points: (a) Intelligibility model and (b) gain constraint. The result of this is that NAL-NL2 prescribes relatively more gain in high and low frequencies than at the mid frequencies. Moreover, NAL-NL2 takes the profile of the hearing-aid user’s age, gender, experience, language, and compressor speed into consideration [4]. On the other hand, the DSL fitting [6] aims at comfortable listening levels to maximize the speech recognition performance in each frequency region. The procedure uses the desired sensation levels to calculate its empirical real-ear aided gain.
Standard machine learning approaches to predict hearing aid fitting could be a potential solution for audiologists in clinical applications. Recently, deep learning is becoming a mainstream technology in solving major issues that have been stacked for ages in the artificial intelligence (AI) community. It has turned out to be very promising at discovering complicated high-dimensional data structures and is, thus, applicable to the domains of business, administration, and science at the industrial scale [9]. In addition to beating records in image recognition [10,11] and speech recognition [12,13,14], it has beaten other machine-learning-based approaches at predicting the activity of living organism, DNA sequencing to bioimaging, and neuroimaging and in different biological applications [15]. Researchers believe that, in the near future, deep machine learning will have many more successes because it requires very light engineering by hand, so it can easily take advantage of the increasing amount of available data and computing. New learning algorithms that are currently being developed for deep neural networks will only accelerate this progress to a large extent. However, in many real world applications, like the availability of hearing-aid fitting data is limited due to privacy concerns. In addition, the data can be in a different feature space or follow a different data distribution. In such cases, the knowledge transfer would significantly improve the learning performance by avoiding very costly data labeling efforts [16].
Many applications can be found where transfer learning can be truly beneficial. One example is cancer drug sensitivity prediction [17], where a subset of transferred data from the auxiliary data of the related task and the target training data is used to build a model. The resulting transfer learning model is then applied on the test set of the target task to make predictions. In Reference [18], transfer learning with deep network is used to predict salient area in natural video. The knowledge learned with the deep network on a large data set is transferred to train the network on a small data set to predict salient areas. In Reference [19], a Deep Neural Network based fitting prescription for hearing aids has been offered, but the information about detailed methodology, techniques, and how this is accomplished are proprietary and unpublished. In this paper, we propose a transfer learning approach for the application of hearing-aid fitting, where our proposed approach combined several techniques. Our contribution and the structure of the paper are as follows:
  • First, we extracted the fitting data set from two popular nonlinear hearing-aid fitting software (NAL-NL1 and NAL-NL2) and divided the randomly shuffled data set in two groups for training two different models.
  • Second, we investigated the number of minimum fitting records required for sufficient training and observed that more than 1500 records are essential.
  • Third, we trained the NN model with improved hyper parameters where we used random weight initialization and exponential weight decay concept and the approach shows enhanced performance.
  • Fourth, we applied the inductive parameter transfer learning approach in our second model in a smaller data set condition with the same hyper parameter. We transferred the final weights from the NN model as the initial weights for the TL model, and it performed surprisingly well over the traditional NN model.
  • Finally, we compared and analyzed the output of the two models for verification.
The model acquired in this research was selected by reviewing the top performing machine learning algorithm as applied to hearing-aid fitting prescription. As no other papers were found on the topic of transfer learning based fitting algorithm, all of our research on engineering features to estimate the expected gain for hearing loss patient is fancy.

2. Background

In this section, we introduced basic concepts of neural network and transfer learning, as the background of our proposed NN-TL based algorithm.

2.1. Fully Connected Neural Networks

Deep Neural Networks (DNN) are distinguishable from other types of Neural Networks by their depth, computational complexity, and performance. In a fully connected feed-forward DNN model [20], nodes are fully connected with all the nodes in the previous layer. The computational procedure of a fully connected DNN involved with matrix–vector arithmetics and transformation by the activation function can be expressed as follows:
Y = σ ( W T · X + b )
where Y and X are the outputs of the current layer and previous layer, respectively; W is the weight matrix; b is the bias vector; and σ is the activation function.
Each node in a neural network does a weighted sum of all of its inputs, adds a constant named bias, and then feeds the result through some nonlinear activation function (e.g., sigmoid, softmax, or relu). Nodes from the final layer produces the prediction, and a loss function measures the distance between the actual output and predicted output. The connections in a neural network can be shown in Figure 1.

2.2. Transfer Learning

In most applications in the real world, it is expensive or impossible to collect the necessary training data and to rebuild different models. Here comes the knowledge transfer or transfer learning approach that could reduce the effort of data collection. The main aim is to extract the knowledge from one domain and to apply the knowledge in another domain.
In transfer learning what to transfer, when to transfer, and how to transfer are three main research issues. Inductive transfer learning, transductive transfer learning, and unsupervised transfer learning are three popular sub-settings in this approach. Based on “what to transfer”, the approaches to transfer learning in the above sub-settings can be further classified into four classes [16]. The specific class used in this paper can be referred to as the parameter-transfer approach. The parameters (weights) sharing concepts have been widely used in different machine learning models. In general, very common hard weight sharing uses previously trained weights as the initial weights of a neural network model. A general parameter-based transfer learning method can be shown as in Figure 2.

3. Proposed Architecture

A simple block diagram of our proposed hearing-aid fitting model is shown in Figure 3.
The data set contains hearing loss information for 6 different frequencies for each of the individual subjects. Other than hearing loss and insertion gain information for the corresponding loss, the data set does not contain any other features. Therefore, we did a binary conversion of the hearing loss data as pre-processing and then positioned them based on frequency hierarchy to make more features (6 frequencies * 7 bit binary conversion * 6 positions) as shown in Figure 4.
Then we applied the NN and TL concept as two separate models. Finally, we compare and verify the results of these two models.

3.1. Neural Network with Refined Hyper-Parameter

In the first part of our proposed design, we considered a simple 4-layer (252 units of input layer, 85 units of first and 30 units of second hidden layers, and 6 units of output layer) Multilayer Perceptron (MLP) Neural Network as shown in Figure 5.
The architecture adopted here is similar to the DNN regression approach [21] that is based on a feed-forward neural network [20,22], having many levels of nonlinearities and allowing them to densely illustrate an immensely varying and distinctly nonlinear function. In Reference [21], the training involves an unsupervised pre-training and supervised fine-tuning. However, in our approach, we only considered the simple supervised training with a refined hyper-parameter.
We choose random weight initialization, ensuring the weights varied in the range of 0.5 to 0.5 ( ± 0.1 ) from input layer to first hidden layer, 1.0 to 1.0 ( ± 0.1 ) from first hidden layer to second hidden layer, and 1.5 to 1.5 ( ± 0.1 ) from second hidden layer to output layer. We also considered the exponential weight decay concept in our approach. After initializing all the parameters, in the training session, the feed forward propagation is calculated first, the back propagation is calculated second, then the parameter is updated, and loss function is calculated at the end.

3.1.1. Feed Forward Propagation

Feed forward neural network learned to map a fixed size input to a fixed size output. The results of weighted sums from hidden layers to the output layer passes through a nonlinear activation function. We use sigmoid ( σ = 1 / ( 1 + e x ) ), the most classical activation function in all the layers. In the feed forward propagation, the output of two hidden layers and output layer can be expressed as follows:
First hidden layer:
H j 1 = σ ( i = 1 252 w i j · I i )
Second hidden layer:
H k 2 = σ ( j = 1 85 w j k · H j 1 )
Output layer:
O l = σ ( k = 1 30 w k l · H k 2 )
where I i is the input; w i j , w j k , and w k l are the corresponding weight matrices; H j 1 , H k 2 , and O l are the outputs of the first hidden layer to the output layer respectively; and σ is the activation function.

3.1.2. Error Calculation

The error is calculated with the simple subtraction as follows:
e l = y l O l
where y l is the real sample output.

3.1.3. Back Propagation and Weight Update

To properly adjust the weight vector, we considered the mini-batch Stochastic Gradient Descent (SGD) [23] optimizer. To lower the computation cost and to get lower variance, we used 100 subset of our data set each time we trained the network. First, we get the gradients of each layer of the networks for the current mini-batch, and then, we use that gradients to update the weights of each layers of the networks. To make the update operation very easy, we just add the gradient of particular weight matrix to our existing weight matrix. In addition, to make the learning better, we scale the gradients with a suitable learning rate. This less-complicated procedure usually finds a good set of weights surprisingly well in comparison to far more elaborate optimization techniques [9].

3.2. Transfer Learning Approach

In the second part of our proposed design, we use the concept of simple inductive parameter transfer learning approach [16]. We use the same NN model that includes the same hyper-parameter here in the TL approach. However, instead of random weight initialization, we retained the final weights from NN model and employed them here as the initial weights of the Transfer Learning model as shown in Figure 6.

4. Experimental Results and Evaluation

The performance of our proposed fitting algorithm have been evaluated in different phases. In the first subsection, the complete information about the data set has provided. The the next section, we tried to find the minimum number of fitting records requires for sufficient training. After that, performance of the NN algorithm with a refined hyper-parameter and the performance of the TL approach have been evaluated. Evaluation with a Mean Squared Error (MSE) analysis and some statistical parameter has also been considered.

4.1. Data Sets

Hearing loss vs. insertion gain data have been extracted manually for 1100 subjects from the two renowned nonlinear hearing-aid fitting softwares; National Acoustic Laboratories’ nonlinear fitting procedure, version 1 (NAL-NL1) and version 2 (NAL-NL2). We considered flat hearing loss type for each subject, and the hearing loss varied maximally by 3 dB within 6 frequency bands (250 Hz, 500 Hz, 1 KHz, 2 KHz, 4 KHz, and 8 KHz). The maximum and minimum hearing loss we considered were in between 10 dB and 70 dB respectively. We manually put the hearing loss information in the software and noted the corresponding insertion gain information for the 50 dB, 65 dB, and 80 dB input levels. The 1000 training and 100 test sets were selected randomly in such a way that no data appears more than once in the combined training and test sets. There were no correlation between the two data sets; hence, we considered them separately. The full data sets can be found in the Supplementary Materials.

4.2. Number of Fitting Records for Sufficient Training

Number of training data always plays an important role in the performance of the machine learning based algorithm. In general, the more the data set, the better the training [9]. However, pondering real life situations, a big number of fitting records from audiologist may not always be practical. Therefore, we did an experiment to find the minimum number of fitting data set required for sufficient training by varying the amount of fitting data from 2800 to 1000. It was noticed that, the more training data used, the smaller the gain difference became. Furthermore, a monotonic difference in the gain with the increase in hearing loss was examined. Then, we were convinced that about 1500 data are necessary for sufficient training.

4.3. Efficiency of Predicting Fitting Records

The algorithm performance for 3 different input levels (50 dB, 65 dB, and 80 dB) for NAL-NL1 are shown in Figure 7, Figure 8 and Figure 9 and for NAL-NL2 are shown in Figure 10, Figure 11 and Figure 12 correspondingly. The solid green line illustrates the insertion gain extracted from the fitting software, and the solid red and blue line illustrate the gain predicted by the NN and TL algorithms. On the other hand, the dashed line indicates the corresponding gain differences. Many of the acronyms used in the performance comparison are defined in the legends of the figures.
Even though the proposed NN approach have followed the trend of the fitting formula pretty well compared to the state-of-the-art approach [19], it is clearly visible that the TL approach has outperformed it and that the predictions were very close to the original fitting records extracted by the formula. In most of the cases, the differences were in between −2 dB and +2 dB. However, some unexpected gain fluctuation can be noticed in some cases but can be neglected.

4.4. Mean Square Errors Comparison

Predictive mean square error results are shown in Table 1. In the table, we described the prediction error rate of our proposed TL method compared with NN method for three different input levels, and the results were taken after the 3rd epoch for each cases. From the table, it is clear that the TL algorithm for each of the input level significantly outperforms the results of the NN algorithm. In addition, our proposed transfer-learning-based fitting algorithm obtains better predictions and thus had smaller MSE for all the cases. The average MSE for TL and NN were 0.8662% and 1.1792% respectively.

4.5. Statistical Evaluation

Table 2 and Table 3 use three statistical indicator (average, minimum, and maximum) to analyze the performance of TL algorithm over the NAL-NL1 and NAL-NL2 data set, respectively. In both of the tables, it is clearly noticeable that the differences were quite negligible except some cases that were highlighted in bold.

5. Discussion

All the advanced and modern fitting methodologies are fortunate to provide at least some corrective gain to the hearing-impaired ear. Although not yet developed, the optimum fitting method would seek to restore all the dynamic acoustic properties lost through cochlear and conductive causes. However, each patient functions as a unique experiment with variable and unpredictable outcomes. Therefore, the insertion gain calculated by fitting formulas might not be the preferred gain for a hearing loss patient, and there is no potential benchmark for an acceptable prediction error. The preferred gain can vary largely from subject to subject, male to female, young to old, or one geographical region to another. In addition, we also do not know how each fitting software calculates the insertion gain for a corresponding hearing loss information, thus it remains a black box.
Therefore, in this paper, we extracted the patient hearing loss and insertion gain data from fitting software and applied the machine learning strategies to see if they can follow the trend of fitting formulas. By applying this approach with the clinical data from audiologist, we can mitigate the gap between fitting satisfaction of a hearing loss patient. The collection of large hearing aid fitting data from audiologist for further investigation will be a immense challenge due to the regulations and privacy concerns. That is why we considered parameter transfer learning approach, so that we can deal with the smaller data set problem. The TL prediction error for insertion gain is basically the calculated difference between predictive gain and the original gain from fitting formula. Therefore, very small or no difference will be expected in ideal case.

6. Conclusions and Future Work

MLP-NN model with refined hyper-parameters empowers computational prototypes consisting of multiple processing layers to learn data representation in multiple levels of abstraction and TL approach speeds up the training and enhances the performance of the model even in a smaller data set condition. For future work, we will be using the hearing-aid fitting data that would be collected from the health care professionals. We are expecting more features in the data set that includes the patients’ personal preference to the hearing-aid type. We will be considering more challenging and smaller data set conditions to get a more realistic impression about the performance of the proposed algorithm. More consideration will have to be provided to optimize and evaluate the proposed approach.

Supplementary Materials

The following are available online at https://www.mdpi.com/2079-9292/8/7/736/s1. A “Readme.txt” file has also been included mentioning all the details about the included files.

Author Contributions

Conceptualization, S.I.M.M.R.M. and S.L.; data curation, S.I.M.M.R.M. and S.L.; formal analysis, S.I.M.M.R.M. and S.L.; funding acquisition, S.L.; investigation, S.I.M.M.R.M. and S.L.; methodology, S.I.M.M.R.M.; project administration, S.L.; resources, S.L.; software, S.I.M.M.R.M.; supervision, S.L.; validation, S.I.M.M.R.M.; visualization, S.I.M.M.R.M.; writing—original draft, S.I.M.M.R.M.; writing—review and editing, S.I.M.M.R.M. and S.L.

Funding

This work was supported by INHA UNIVERSITY Research Grant and by Basic Science Research Program (NRF-2016R1A2B4015370) through the National Research Foundation of Korea (NRF).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Sandlin, R.E. Textbook of Hearing Aid Amplification; Singular Publishing Group, Thomson Learning: San Diego, CA, USA, 2000. [Google Scholar]
  2. Dillon, H. NAL-NL1: A new procedure for fitting non-linear hearing aids. Hear. J. 1999, 52, 10–16. [Google Scholar] [CrossRef]
  3. Byrne, D.; Dillon, H.; Ching, T.; Katsch, R.; Keidser, G. NAL-NL1 procedure for fitting nonlinear hearing aids: Characteristics and comparisons with other procedures. J. Am. Acad. Audiol. 2001, 12, 37–51. [Google Scholar] [PubMed]
  4. Keidser, G.; Dillon, H.; Flax, M.; Ching, T.; Brewer, S. The NAL-NL2 prescription procedure. Audiol. Res. 2011, 1, 88–90. [Google Scholar] [CrossRef] [PubMed]
  5. Keidser, G.; Dillon, H.; Carter, L.; O’Brien, A. NAL-NL2 Empirical Adjustments. Trends Amplif. 2012, 16, 211–223. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Dillon, H. Hearing Aids; Boomerang Press: Turramurra, Australia, 2012; Volume 30, p. 169. [Google Scholar] [CrossRef]
  7. Seewald, R.; Cornelisse, L.; Ramji, K.; Sinclair, S.; Moodie, K.; Jamieson, D. DSL 4.1 for Windows; Hearing Health Care Research Unit, University of Western Ontario: London, ON, Canada, 1997. [Google Scholar]
  8. Seewald, R.; Moodie, S.; Scollie, S.; Bagatto, M. The DSL method for pediatric hearing instrument fitting: Historical perspective and current issues. Trends Amplif. 2005, 9, 145–157. [Google Scholar] [CrossRef] [PubMed]
  9. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  10. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  11. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 1–9. [Google Scholar] [CrossRef] [Green Version]
  12. Hinton, G.; Deng, L.; Yu, D.; Dahl, G.; Mohamed, A.R.; Jaitly, N.; Senior, A.; Vanhoucke, V.; Nguyen, P.; Sainath, T.; et al. Deep Neural Networks for Acoustic Modeling in Speech Recognition: The Shared Views of Four Research Groups. IEEE Signal Process. Mag. 2012, 29, 82–97. [Google Scholar] [CrossRef]
  13. Graves, A.; Mohamed, A.R.; Hinton, G. Speech Recognition with Deep Recurrent Neural Networks. In Proceedings of the 2013 IEEE International Conference on Acoustics, Speech and Signal Processing, Vancouver, BC, Canada, 26–31 May 2013. [Google Scholar] [CrossRef]
  14. Zhang, Y.; Pezeshki, M.; Brakel, P.; Zhang, S.; Bengio, C.L.Y.; Courville, A. Towards End-to-End Speech Recognition with Deep Convolutional Neural Networks. arXiv 2017, arXiv:1701.02720. [Google Scholar]
  15. Mahmud, M.; Kaiser, M.S.; Hussain, A.; Vassanelli, S. Applications of Deep Learning and Reinforcement Learning to Biological Data. IEEE Trans. Neural Netw. Learn. Syst. 2018, 29, 2063–2079. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Pan, S.J.; Yang, Q. A Survey on Transfer Learning. IEEE Trans. Knowl. Data Eng. 2010, 22, 1345–1359. [Google Scholar] [CrossRef]
  17. Turki, T.; Wei, Z.; Wang, J.T.L. A transfer learning approach via procrustes analysis and mean shift for cancer drug sensitivity prediction. J. Bioinform. Comput. Biol. 2018, 16, 1840014. [Google Scholar] [CrossRef] [PubMed]
  18. Chaabouni, S.; Benois-Pineau, J.; Ben Amar, C. Transfer learning with deep networks for saliency prediction in natural video. In Proceedings of the 2016 IEEE International Conference on Image Processing (ICIP), Phoenix, AZ, USA, 25–28 September 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 1604–1608. [Google Scholar] [CrossRef]
  19. Haruta, C.; Kamimura, Y.; Sunohara, M.; Tateno, M.; Co, R. Deep Neural Network-based fitting prescription for hearing aids: The efficacy of the proposed system. In Proceedings of the 2016 International Hearing Aid Research Conference (IHCON2016), Tahoe City, CA, USA, 10–14 August 2016; p. 1. [Google Scholar]
  20. Khadivi, P.; Tandon, R.; Ramakrishnan, N. Flow of Information in Feed-Forward Deep Neural Networks. arXiv 2016, arXiv:1603.06220. [Google Scholar]
  21. Xu, Y.; Du, J.; Dai, L.R.; Lee, C.H. A Regression Approach to Speech Enhancement Based on Deep Neural Networks. IEEE Trans. Audio Speech Lang. Process. 2015, 23, 7–19. [Google Scholar] [CrossRef]
  22. Zemouri, R. An evolutionary building algorithm for Deep Neural Networks. In Proceedings of the 2017 12th International Workshop on Self-Organizing Maps and Learning Vector Quantization, Clustering and Data Visualization (WSOM), Nancy, France, 28–30 June 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 1–7. [Google Scholar] [CrossRef]
  23. Li, M.; Zhang, T.; Chen, Y.; Smola, A.J. Efficient mini-batch training for stochastic optimization. In Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining—KDD’14, New York, NY, USA, 24–27 August 2014; ACM Press: New York, NY, USA, 2014; pp. 661–670. [Google Scholar] [CrossRef]
Figure 1. Neural network connections.
Figure 1. Neural network connections.
Electronics 08 00736 g001
Figure 2. Parameter-based transfer learning method (weight transfer).
Figure 2. Parameter-based transfer learning method (weight transfer).
Electronics 08 00736 g002
Figure 3. Block diagram of the proposed hearing-aid fitting.
Figure 3. Block diagram of the proposed hearing-aid fitting.
Electronics 08 00736 g003
Figure 4. Binary conversions of input features.
Figure 4. Binary conversions of input features.
Electronics 08 00736 g004
Figure 5. Multilayer Perceptron Neural Network (MLP-NN) Model.
Figure 5. Multilayer Perceptron Neural Network (MLP-NN) Model.
Electronics 08 00736 g005
Figure 6. Parameter transfer learning model.
Figure 6. Parameter transfer learning model.
Electronics 08 00736 g006
Figure 7. NAL-NL1: Performance comparison for NN and TL vs. fitting formula for the 50 dB input level. Comparisons are made for 250 Hz, 500 Hz, 1 KHz, 2 KHz, 4 KHz, and 8 KHz frequencies from top left to right to bottom left to right.
Figure 7. NAL-NL1: Performance comparison for NN and TL vs. fitting formula for the 50 dB input level. Comparisons are made for 250 Hz, 500 Hz, 1 KHz, 2 KHz, 4 KHz, and 8 KHz frequencies from top left to right to bottom left to right.
Electronics 08 00736 g007
Figure 8. NAL-NL1: Performance comparison for NN and TL vs. fitting formula for the 65 dB input level. Comparisons are made for 250 Hz, 500 Hz, 1 KHz, 2 KHz, 4 KHz, and 8 KHz frequencies from top left to right to bottom left to right.
Figure 8. NAL-NL1: Performance comparison for NN and TL vs. fitting formula for the 65 dB input level. Comparisons are made for 250 Hz, 500 Hz, 1 KHz, 2 KHz, 4 KHz, and 8 KHz frequencies from top left to right to bottom left to right.
Electronics 08 00736 g008
Figure 9. NAL-NL1: Performance comparison for NN and TL vs. fitting formula for the 80 dB input level. Comparisons are made for 250 Hz, 500 Hz, 1 KHz, 2 KHz, 4 KHz, and 8 KHz frequencies from top left to right to bottom left to right.
Figure 9. NAL-NL1: Performance comparison for NN and TL vs. fitting formula for the 80 dB input level. Comparisons are made for 250 Hz, 500 Hz, 1 KHz, 2 KHz, 4 KHz, and 8 KHz frequencies from top left to right to bottom left to right.
Electronics 08 00736 g009
Figure 10. NAL-NL2: Performance comparison for NN and TL vs. fitting formula for the 50 dB input level. Comparisons are made for 250 Hz, 500 Hz, 1 KHz, 2 KHz, 4 KHz, and 8 KHz frequencies from top left to right to bottom left to right.
Figure 10. NAL-NL2: Performance comparison for NN and TL vs. fitting formula for the 50 dB input level. Comparisons are made for 250 Hz, 500 Hz, 1 KHz, 2 KHz, 4 KHz, and 8 KHz frequencies from top left to right to bottom left to right.
Electronics 08 00736 g010
Figure 11. NAL-NL2: Performance comparison for NN and TL vs. fitting formula for the 65 dB of nput level. Comparisons are made for 250 Hz, 500 Hz, 1 KHz, 2 KHz, 4 KHz, and 8 KHz frequencies from top left to right to bottom left to right.
Figure 11. NAL-NL2: Performance comparison for NN and TL vs. fitting formula for the 65 dB of nput level. Comparisons are made for 250 Hz, 500 Hz, 1 KHz, 2 KHz, 4 KHz, and 8 KHz frequencies from top left to right to bottom left to right.
Electronics 08 00736 g011
Figure 12. NAL-NL2: Performance comparison for NN and TL vs. fitting formula for the 80 dB input level. Comparisons are made for 250 Hz, 500 Hz, 1 KHz, 2 KHz, 4 KHz, and 8 KHz frequencies from top left to right to bottom left to right.
Figure 12. NAL-NL2: Performance comparison for NN and TL vs. fitting formula for the 80 dB input level. Comparisons are made for 250 Hz, 500 Hz, 1 KHz, 2 KHz, 4 KHz, and 8 KHz frequencies from top left to right to bottom left to right.
Electronics 08 00736 g012
Table 1. MSE Comparison for NN and TL.
Table 1. MSE Comparison for NN and TL.
Input LevelNN (in %)TL (in %)
50 dB1.09140.7101
65 dB1.08710.8330
80 dB1.35931.0555
Average1.17920.8662
Table 2. Statistical analysis of NAL-NAL1.
Table 2. Statistical analysis of NAL-NAL1.
GroupInput Level
50 dB65 dB80 dB
TLNL1DiffTLNL1DiffTLNL1Diff
250 HzAverage4.854.76−0.094.104.02−0.082.842.74−0.10
Min0.490.500.010.631.000.370.930.50−0.43
Max15.1818.503.3211.7215.003.287.5010.002.50
500 HzAverage13.0413.040.009.149.190.054.034.150.12
Min3.392.00−1.392.582.00−0.580.841.000.16
Max23.3226.002.6816.9319.502.5710.9613.502.54
1 KHzAverage23.4623.680.2119.6019.760.1610.7511.010.25
Min10.7210.00−0.728.387.00−1.383.102.50−0.60
Max34.6637.502.8429.7332.502.7720.1523.002.85
2 KHzAverage23.4824.030.5518.3718.490.1210.5810.690.11
Min12.2912.500.217.547.00−0.542.482.00−0.48
Max33.5637.003.4427.9730.502.5321.0423.502.46
4 KHzAverage19.8720.160.2915.2215.05−0.179.839.890.06
Min10.8810.00−0.886.054.00−2.052.021.50−0.52
Max27.8030.002.2022.7824.501.7218.7420.001.26
8 KHzAverage5.525.690.177.137.260.136.947.060.12
Min3.613.00−0.614.474.500.034.534.00−0.53
Max8.9814.005.0210.8015.004.2010.6615.004.34
Table 3. Statistical analysis of NAL-NAL2.
Table 3. Statistical analysis of NAL-NAL2.
GroupInput Level
50 dB65 dB80 dB
TLNL2DiffTLNL2DiffTLNL2Diff
250 HzAverage10.7710.20−0.565.404.96−0.440.720.50−0.22
Min1.780.00−1.780.460.00−0.460.050.00−0.05
Max21.1722.471.3015.4716.741.276.147.151.01
500 HzAverage12.0311.44−0.595.444.98−0.460.720.48−0.24
Min1.770.00−1.770.430.00−0.430.060.00−0.06
Max24.4826.361.8816.4617.941.486.067.161.10
1 KHzAverage16.3415.66−0.6810.5810.55−0.044.124.170.05
Min3.830.00−3.832.210.00−2.210.550.00−0.55
Max28.9732.063.0920.8423.382.5411.7213.151.43
2 KHzAverage18.4817.73−0.7412.0911.94−0.154.814.80-0.02
Min6.051.72−4.333.070.00−3.070.690.00−0.69
Max30.8833.312.4323.2825.392.1114.2415.501.26
4 KHzAverage18.9118.46−0.4513.2213.300.076.086.200.12
Min6.492.77−3.723.400.00−3.400.760.00−0.76
Max30.5433.633.0924.0327.103.0715.7417.752.01
8 KHzAverage19.0618.54−0.5114.9714.91−0.068.638.890.26
Min5.111.48−3.633.390.00−3.391.440.00−1.44
Max32.5135.783.2727.2730.313.0419.6722.002.33

Share and Cite

MDPI and ACS Style

Mondol, S.I.M.M.R.; Lee, S. A Machine Learning Approach to Fitting Prescription for Hearing Aids. Electronics 2019, 8, 736. https://doi.org/10.3390/electronics8070736

AMA Style

Mondol SIMMR, Lee S. A Machine Learning Approach to Fitting Prescription for Hearing Aids. Electronics. 2019; 8(7):736. https://doi.org/10.3390/electronics8070736

Chicago/Turabian Style

Mondol, S I M M Raton, and Sangmin Lee. 2019. "A Machine Learning Approach to Fitting Prescription for Hearing Aids" Electronics 8, no. 7: 736. https://doi.org/10.3390/electronics8070736

APA Style

Mondol, S. I. M. M. R., & Lee, S. (2019). A Machine Learning Approach to Fitting Prescription for Hearing Aids. Electronics, 8(7), 736. https://doi.org/10.3390/electronics8070736

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop