Next Article in Journal
Brain Strategy Algorithm for Multiple Object Tracking Based on Merging Semantic Attributes and Appearance Features
Previous Article in Journal
Modeling and Parallel Operation of Exchange-Biased Delta-E Effect Magnetometers for Sensor Arrays
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Analysis of Gastrointestinal Acoustic Activity Using Deep Neural Networks

1
Institute of Computer Science, Warsaw University of Technology, 00-665 Warsaw, Poland
2
Graduate School of Information, Production and Systems, Waseda University, Tokyo 169-8050, Japan
3
Department of Pediatric Gastroenterology and Metabolic Diseases, Poznan University of Medical Sciences, 60-572 Poznan, Poland
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(22), 7602; https://doi.org/10.3390/s21227602
Submission received: 21 October 2021 / Revised: 9 November 2021 / Accepted: 14 November 2021 / Published: 16 November 2021
(This article belongs to the Section Biomedical Sensors)

Abstract

:
Automated bowel sound (BS) analysis methods were already well developed by the early 2000s. Accuracy of ~90% had been achieved by several teams using various analytical approaches. Clinical research on BS had revealed their high potential in the non-invasive investigation of irritable bowel syndrome to study gastrointestinal motility and in a surgical setting. This article proposes a novel methodology for the analysis of BS using hybrid convolutional and recursive neural networks. It is one of the first methods of using deep learning to be widely explored. We have developed an experimental pipeline and evaluated our results with a new dataset collected using a device with a dedicated contact microphone. Data have been collected at night-time, which is the most interesting period from a neurogastroenterological point of view. Previous works had ignored this period and instead kept brief records only during the day. Our algorithm can detect bowel sounds with an accuracy >93%. Moreover, we have achieved a very high specificity (>97%), crucial in diagnosis. The results have been checked with a medical professional, and they successfully support clinical diagnosis. We have developed a client-server system allowing medical practitioners to upload the recordings from their patients and have them analyzed online. This system is available online. Although BS research is technologically mature, it still lacks a uniform methodology, an international forum for discussion, and an open platform for data exchange, and therefore it is not commonly used. Our server could provide a starting point for establishing a common framework in BS research.

1. Introduction

The gastroenterologist’s basic tools, including medical history, physical examination, laboratory tests, endoscopic examinations and standard imaging, complement several expensive and usually invasive specialist investigations. They include, among others, methods of examining bowel function (motility) such as multiple X-rays following oral tracer administration or rectal infusion of contrast media, as well as antroduodenal or colonic manometry, requiring general anesthesia before insertion of the probe. These methods have several disadvantages, such as a high degree of invasiveness (radiation, anesthesia, and catheter insertion), financial and time costs, and in many cases, very poor availability. Above all, they do not assess the motility of the small intestine, apart from at a few specialist centers. There is a need for a new, non-invasive method to assess the motor function of the intestine.
We propose analyzing sounds made by the digestive system, called bowel sounds (BS), by computer programs (software). BS analysis has greatly developed since 1958, when the first recordings were described [1]. First computerized analyses of BS were conducted in 1970s [2], however technical solutions have still not been made available to physicians, despite technological progress.
One of the key studies that best demonstrated the clinical potential of BS research was published by Craine et al. in 1999 and discussed differences in characteristics of BS patterns between patients with irritable bowel syndrome and controls [3]. Furthermore, at this time, stationary–nonstationary filters with wavelet transformation were introduced [4], later enriched with fractal dimension analysis. Wavelet transformations have also been used with Wiener filtering [5]. Principal component analysis of BS was done in early 2000s, along with considerable work on identification of anatomical location of BS sources [6]. Much research has also been done on signal denoising using various filters, mostly based on wavelet transformations or frequency thresholds [7]. Importantly, Kim et al. who reported BS detection with back-propagation neural networks [8] also correlated BS with colonic transit time. Real-time BS detection was achieved using a naive Bayesian classifier [9]. Support vector machines have also been used for BS identification [10], including a wearable device [11]. Other more recent approaches included spectral entropy analysis, Gaussian Hamming distance analysis [12]. In China, Liu et al. applied voice recognition neural networks to BS identification and obtained encouraging results [13]. Much progress was achieved by the Noisy Guts team from Australia. They have not only proposed their own neural network-based approach to BS identification, but also studied irritable bowel syndrome and demonstrated a link between BS and migrating motor complexes in the large bowel [14]. For more historical context we refer to our recent review [15].
The Noisy Guts team [14], who have also studied intestinal sounds, distinguished five types of sounds based on 20 h of recordings: short sounds (single bursts), multiple sounds (multiple bursts), long irregular sounds (random continued sounds), harmonic sounds (harmonic sounds) and a combination of the sounds previously mentioned (combination sounds). This large variety of detected sounds strongly suggests that there is still room for improvement in BS analysis.
The division mentioned above does not result from various mechanisms in the human body but only from an analysis of intestinal sound recordings. As a result, characteristics of some types may easily overlap. In the case of short sounds, some of them are very quiet, and it is sometimes debatable whether a sound should be considered an intestinal sound or just be ignored. The method described later in this paper does not focus on classifying sounds into types but only on their detection. This is because it is potentially most useful in medical research to quantify intestinal sounds.
Most current solutions to detect intestinal sounds are based on the extraction of features from recordings without the use of machine learning methods [16]. One of the simplest methods is to detect a sound if the average amplitude of a sound signal in a given time interval exceeds a specific value. The disadvantage of this method is that it does not solve the problem of recording noise. A simple improvement of the method is the determination of the average amplitude solely based on components with specific frequencies. More sophisticated solutions also include noise reduction using adaptive filters [17].
Machine learning methods are widely used in bowel sound analysis [15]. One solution using machine learning algorithms is the use of the parameters Jitter and Shimmer, i.e., parameters that determine the variation of the fundamental frequency in relation to the frequency of the fundamental tone and the variation of the amplitude in relation to the average amplitude. On this basis, detailed attributes are distinguished, which constitute the input to the logistic regression model. Based on a small sample of data, this method has achieved a sensitivity of 86.3 % and specificity of 91.0 % .
An example of using a neural network to recognize intestinal sounds is the AIMAS (Autonomous Intestinal Motility Analysis System) system [18]. The first step in its algorithm is to remove noise and amplify the signal using predefined filters. The time, frequency, and signal parameters derived from wavelet transform are then inserted as inputs into the dense neural network. This method allows the detection and classification of intestinal sounds with accuracy of 94.3 % . The usage of neural networks is also presented in [19].
Our approach, not explored in any depth elsewhere, is to analyze sounds made by the digestive system using machine learning models based on a particular type of deep neural network. There are solutions for recognizing intestinal sounds that use dense neural networks, but so far, there have been no methods based on convolutional or recursive networks. Due to the widespread application of these networks elsewhere, an attempt to apply them to detecting intestinal sounds is interesting and worth investigating.
We have collected data using a new device with a dedicated contact microphone. We then created a dataset with 321,000 records of length 10 ms that medical doctors labeled. Our research hypothesis is that developing a new model using a deep neural network trained on a high-quality dataset enables bowel sound detection with high accuracy, specificity and sensitivity.

2. Methods

2.1. Data Acquisition

An intestinal sound-dedicated contact microphone was designed and produced at Poznan University of Medical Sciences (Poland). It consisted of an electromagnetically (copper foil) and acoustically isolated (foam) low resonance frequency (1.3 kHz) piezoelectric transducer (Ningbo Kepo Electronics, Ningbo, Zhejiang, China) enclosed in a custom design low-profile (≈1 cm) 3D-printed head containing a standard stethoscope membrane made of epoxy resin and glass fiber. The enclosure was printed with a notch around its circumference, similar to a stethoscope head, enabling mounting of the stethoscope membrane rim. The microphone was attached to the skin using 2–3 cm of two-sided medical tape (hair system tape, low-strength). The microphones work with lightweight Tascam DR-10CH (TEAC, Tokyo, Japan) recorders that capture sound with a depth of 24 bits, making it possible to record both quiet tones and individual events that are very loud.
The following protocol was established, guaranteeing the highest sound quality: the patient was instructed not to eat large meals after 18:00 and snacks after 21:00. Before going to bed, the patient was required to stick the microphone on the lower right part of their abdomen and switch on the recorder. In the evening, the patient also had to complete a clinical and nutritional questionnaire. A morning stool sample was collected to correlate selected parameters with automatically obtained quantitative data assessing bowel function.
Apart from eliminating private conversations and other disturbances, the protocol adopted had a significant advantage from the point of view of physiology: due to frequent meals, in many patients the third phase of intestinal motor activity (migrating motor complexes) appears only in sleep. It is characterized by increased peristalsis, the purpose of which is to cleanse the small intestine. The first and the second phases occur both at night and during the day. A person is not conscious and movement is minimized. There is a high probability of a long period without food. In the morning, the activity of the autonomic nervous system changes. Our preliminary results suggest that the third phase is audible and generates a loud rhythmic signal, coupled with respiratory movements. The recordings obtained may also allow detection of sleep apnea and heart rate changes.
We currently have a collection of almost one hundred recordings created by a team from the Department of Pediatric Gastroenterology and Metabolic Diseases in Poznań, Poland. Ultimately, we intend to have the largest and most diverse collection in the world. Some of the material has been tagged by marking the start and end times of thousands of recording fragments that contain intestinal sounds of interest to us. The dataset marked in this way was used to train the above-described neural network to detect interesting sound events.
The recordings from 19 subjects were divided into 2-s fragments and then mixed. The data set consisted of 321,000 records of length 10 ms. 15 % of randomly selected records were extracted as test data, with the remaining 85 % used for training and fivefold cross-validation.

2.2. Algorithm Overview

The recording is input for the algorithm depicted in Figure 1. This algorithm reduces the detection and characterization of intestinal sounds to a binary classification of recording fragments called frames.
The conversion between the audio signal to spectrogram is performed using Discrete Fourier Transform. The spectrogram is created by calculating spectral components using the window that shifts alongside the time axis. The signal was converted from the function of time x ( t ) to the function of time and frequency X ( t , f ) . The process involves using Short-Time Fourier Transform, depicted in Equation (1), where L is the window size, w n multiplier dependent on the window function, M is the number of frames related to the total signal length and the window shift.
X k , m = n = 0 L 1 w n x m L + n e 2 π i k n L , k { 0 , , L 1 } m { 0 , . . . , M 1 }

2.3. Parameters Optimization

In all numerical experiments, we used binary cross-entropy as a function of cost. The algorithm ADAM [20] implemented in the Keras library with default parameters was used for optimization. The experiments examining the influence of a given parameter were carried out on one architecture in which the only factor that changed was the parameter under study.
The metrics presented later show the quality of frame classification by a given model. The data presented contains the results of the best epochs during the training. The figures with the learning curves are from a one-time training of the model lasting 100 epochs.
Due to the huge space of neural network hyperparameter searches and the high computational cost of training, we carried out only a small subset of possible experiments. For this reason, we assume that the parameters could be significantly better, and, as a result, it would be possible to obtain a better quality of the model. In particular, we did not test the influence of the optimizer and several types of augmentation used to improve the model’s quality.
For parameters optimization We used sub-set containing 114,000 records.
First, we store the predictions performed on a straightforward underlying algorithm to obtain a baseline. It consisted of determining the average value of the spectrogram elements from the 0–1500 Hz interval. The frame was then considered as a bowel sound or not by comparing its mean with a predetermined threshold value. The results are shown in Table 1. The highest accuracy achieved was 86.55 % , with a sensitivity of 13.83%. The base model shows that bowel sounds are mostly quiet compared to the noise in the recording. Algorithms based on straightforward attributes extracted from the recording are poorly suited to detecting these sounds.

2.4. Spectrogram Frequency Range

The maximum frequency determines the highest frequency of the spectrogram entering the neural network. The goal is to find a frequency above which sounds are sporadic because allowing scarcer sounds to be part of the main group makes it difficult for the model to do its job.
Based on the results of Table 2, it can be seen that the best value for the maximum frequency is 1500 Hz. Further increasing the range does not improve the results and even worsens them slightly, although the differences are not great, as can be seen in Figure 2. This is consistent with the fact that the range of short sounds is 60–1500 Hz. The biggest drop in model quality is seen for max freq equal to 500 Hz. This proves that the frequency range in the 500–1000 Hz range is important for prediction and intestinal sounds are often visible in this range.

2.5. Frame Width and Frame Smoothing Influence

The smaller the width of the fft window, the greater the resolution of the time domain spectrogram. However, its increase comes at the cost of reducing the resolution of the frequency domain. Intuitively, there should be a window width for which the resulting spectrograms perform best as inputs to the neural network. For each width change, the hop length window step was changed as follows f f t 4 . Such a step value serves to smooth and increase the frame resolution.
Based on the results from Table 3 and the Figure 3, it can be concluded that by far, the best value of the fft window width is 441 samples, which coincides with the observation that for this value, intestinal sounds are best seen on the spectrogram.
Two kinds of window functions were tested: Hann (2) and Hamming (3) showed on Figure 4.
w H a n n ( n ) = 0 , 5 1 c o s 2 π n N
w H a m m i n g ( n ) = 0 , 54 0 , 46 c o s 2 π n N
The results are presented in Table 4. It shows that using the Hann window clearly gives better results. In this case, the value chosen on the basis of the best visibility in the spectrogram also turned out to be the best.

2.6. Number of Frames for Analysis

The study of the effect of the number of frames was to find the optimal number of contiguous frames used for prediction in the linking model. Intuitively, one 10 ms frame is not enough to classify. The average duration of the briefest short intestinal sound is approximately 20–30 ms. Analysis of the intestinal sound environment on a spectrogram, which is not visible from one frame, is also useful in prediction. On the other hand, too many frames lead to an increase in the number of model parameters and make the training process more difficult.
The results are shown in Table 5 and Figure 5. It can be seen that for a small number of frames, the quality of the model is clearly worse. Above 7 frames, the difference becomes less and less noticeable. With 13 frames, the accuracy of the model reaches the highest value. Despite the best result obtained for 13 frames, the number of 9 contiguous frames was selected as the best model due to the smaller number of training parameters of the model and only slightly worse results, when compared to 13 frames.

2.7. Spectrogram Creation

Based on the research conducted, the best model hyperparameters were determined. The following parameters were selected to create the spectrogram:
  • window width: 441 samples,
  • window step: 110 samples (25%),
  • frequency range: 0–1500 Hz,
  • window smoothing function: Hann window.
The classified frame width is 10 ms (441 samples). This value was adopted because the frame width should be as small as possible. Otherwise, the problem is caused by having several intestinal sounds in one frame or adjacent frames with two different intestinal sounds wrongly identified as one sound.
A frame is classified as an intestinal sound if either of the two conditions is met. The first is when the duration of the intestinal sound in the frame is longer than half the frame width. The second is when more than half the duration of the entire intestinal sound is in a frame. Due to the short frame width, the second condition is met very rarely but allows the potential use of the algorithm for a larger frame size in the future.
The algorithm uses decibel spectrograms. Due to the low frequencies of intestinal sounds, we used a linear scale—there was no point in using the Mel frequency scale. The spectrogram was standardized before entering the network, based on the mean and variance of decibel amplitudes determined from all available recordings.

2.8. Classification

Machine learning methods detected sound events in the solution presented. The artificial neural network detected signals with specific acoustic characteristics by analyzing the spectrogram obtained from recordings. Our model involved a convolutional network (CNN) that played the role of extracting features from the images (spectrograms). We chose two neural network architectures. Both used the ability of convolutional networks to generate well generalizing activation maps. The first model was a combination of the CNN and a recursive network (RNN), and the second model was a combination of the CNN with the dense neural network. The recursive part and dense part operated on a sequence of input data rather than on individual samples. We presnet the schema of the classifier in Figure 6 and Figure 7. In both our models, the first step was to obtain attributes representing the classified frame based on the convolution layers. Then, based on the sequence of vectors of these attributes, classification of each frame was performed. After prediction by the net, adjacent frames considered to be a single bowel sound were combined and treated as the same longer sound. The exact model architectures (layers description, number of neurons, etc.) along with learning hyperparameters are described on the project web-page.
Note that we examined and experimented with a wide variety of classifiers, ranging from more traditional approaches like decision trees and SVM, naive Bayesian to state-of-the-art deep learning classifiers. We fine-tuned the parameters and meta parameters of our algorithms during the experiments. The details and results are available on the project webpage.

2.8.1. Convolutional and Recursive Models

The first model described is a convolutional model with a recursive model called the CRNN [21] model. It works by extracting graphical information from the sequence of vectors obtained, thanks to the convolutional network information about the class using a recursive network. The network diagram is shown in Figure 6. The solution uses convolution layers with two-dimensional filters and max pooling layers to extract the attributes from the frame. For sequence analysis, the bidirectional gated recurrent units layer (GRUs) is used to return the sequence. The dropout layers are used to regularize the network. We used the ReLU activation function at the output of each convolutional and recursive layer. The output was the probability value of being a bowel sound, calculated for a given frame. The output was obtained using a sigmoid.
CRNN has 342,000 parameters, it is composed of the following:
  • Input layer with dimensions: 200 (sequence length) × 15 (spectrogram height) × (width) [(Input(200, 15, 4)]
  • 2D Convolution layer with dimensions: 30 (number of filters) × 3 × 3 (filter size) and ReLU activation function [Conv2D(30, 3, 3, ReLU)]
  • 2D Convolution layer with dimensions: 60 × 4 × 2 and ReLU activation function [Conv2D(60, 4, 2, ReLU)]
  • Layer that flattens the input attributes, yielding a tensor with dimensions of 200 × 60 [(Flatten(200, 600)]
  • Dropout layer with 0.4 dropout probability [Dropout(0.4)]
  • Bidirectional GRUs layer with 80 neurons and ReLU activation function [Bidirectional(GRU, 80, ReLU)]
  • Dropout layer with 0.4 dropout probability [Dropout(0.4)]
  • Fully connected (dense) layer with 1 neuron and sigmoid activation function [Dense(1, sigmoid)]
  • Output layer with 200 × 1 dimensions

2.8.2. Convolutional and Dense Models

The second model described is a model in which the classification of each frame was based on the combined attribute vectors from several frames. For example, when combining seven frames, the frame was classified based on the actual classified frame, plus three frames on the left and three frames to the right of the classified frame. In this way, it was possible to gain a broader context during classification, without using recursive networks that are difficult to train. As with the recursive model, convolution layers and max pooling layers were used to extract the attribute vector from the frame. The network diagram is shown in Figure 7.
CDNN has 115,000 training parameters, and is a build of the following specifications:
  • Input layer with dimensions: 200 (sequence length) × 15 (spectrogram height) × (width) [(Input(200, 15, 4)]
  • Layer that flattens the input attributes, yielding a tensor with dimensions of 200 × 60 [(Flatten(200, 60)]
  • Fully connected (dense) layer with 120 neurons and ReLU activation function [Dense(120, ReLU)]
  • Layer concatenating 9 adjacent vectors [Concat(9)]
  • Dropout layer with 0.4 dropout probability [Dropout(0.4)]
  • Fully connected (dense) layer with 100 neurons and ReLU activation function [Dense(100, ReLU)]
  • Fully connected (dense) layer with 1 neuron and sigmoid activation function [Dense(1, sigmoid)]
  • Output layer with 200 × 1 dimensions

2.9. Data Augmentation

The purpose of this study was to investigate the effect of data augmentation on model quality. Data augmentation is a method of artificially increasing the amount of training data available. It involves generating new data by modifying the existing data, using techniques such as image shifting or adding noise. Due to the specifics of the spectrogram and the very narrow width of a single frame, augmentation by sliding is difficult to apply. For this reason, data augmentation was performed by adding noise to the spectrogram. We used noise with normal distribution.
The results are presented in Table 6. In the case of too much noise, the sounds are jammed. As a result, augmentation has the opposite effect, as the task of recognizing intestinal sound becomes too difficult for the web. Sufficiently low noise allows improving the results slightly. However, it is debatable whether the difference is large enough to make the use of augmentation in this form profitable. The problem with noise augmentation seems to be that some intestinal sounds are very faint, so adding noise makes them virtually impossible to detect and does not help train the net.

2.10. Software Development Details

Software was written in Python version 3.8 and consists of 964 lines of code. We used additional libraries: librosa for preprocessing audio and TensorFlow for training models. We trained the network, ran all tests and performed numerical experiments on a PC with GTX 1080 Ti GPU (4GB RAM) and Ubuntu 20.04 64bit operating system.
The client-server system to analyse WAV recordings online is composed of 3 main components: analysis model described above, back-end server implemented in Python and Flask and front-end website designed with HTML and JavaScript.

3. Results

In Table 7, we present results for two machine learning models: a convolutional model with a recursive model called the CRNN [21], and a model combining a convolutional network and dense network, called the CDNN. The detailed description is given in the Methods section.
The results on testing data are presented in Table 8. These results use whole dataset, 321,000 records from 19 subjects.
We also obtained an annotated recording from another patients. This recording was used to additional validation. We used models depicted previously. The ratio of the gut sound frames to all frames in this recording is only 2.46 % .
The results are presented in Table 9. The large decrease in precision is understandable due to fewer recorded intestinal sounds compared to the recordings on which the models were trained. The annotated sounds were clear, enabling the models to be more sensitive. At the same time, after manual analysis of the result files, it seems that a large number of false positive errors came from the inaccurate determination of frames marking the beginning and end of a sound and not from the model’s failure to detect the sound itself. The other problem with both models was the occasional treatment of a single intestinal sound as two different ones, resulting in the classification of the signal between as a noise.
We performed additional analysis on a different set of one-minute, non-annotated recordings. The CRNN model detected 146 sounds and the CDNN model 121 sounds; 34 were detected only by the CRNN model, and 9 were detected only by the CDNN model. Based on a review of the entire file, we noticed that most of the predictions were identical in both models, and that the most common differences were the shift of the beginning or ending of the sound.
The recursive model is bigger and performed better. In particular, the greater sensitivity of the combining model while maintaining similar precision is noticeable. This observation is also evidenced by the greater number of sounds found by the recursive model on an unannotated recording. In our opinion, the recursive layers are better for analyzing sequences. In addition, the recursive model can analyze the entire sequence and the model that connects only a fragment of it, which can be crucial for long types of intestinal sounds.
We developed the system, allowing medical doctors to upload and analyze samples with recorded intestine activity from patients. This tool provides a real-time analysis of the recording and statistics related to short sounds and graphical visualization of detected sounds overlaid on the recording spectrogram. In particular, the results include recording length, number of bowel sounds identified, number of bowel sounds per minute, frequency analysis in three-minute periods (mean, median, quadrille of bowel sounds per minute), duration analysis depicted individual bowel sound (histogram, percentage of bowel sounds of length 10 ms, 20 ms, 30 ms, etc.), power analysis (root mean square of the successive differences, the standard deviation of bowel sound intervals), Porta’s index, Guzik’s index, a plot of number bowel sounds per minute vs. time, a plot of the histogram of bowel sounds per minute and a plot of the histogram of bowel sound duration.
This system uses a client-server architecture. The user uploads a recording in WAV format using web-page http://bowelsound.ii.pw.edu.pl (accessed on 13 November 2021), then the analysis results are presented.

4. Discussion

A comparison of the results obtained with the results from other methods of detecting intestinal sounds, described in Section 1, was difficult due to the lack of a generally available set of recordings and different definitions of intestinal sounds. We present results provided by authors in Table 10. However, based on the experiments, it can be concluded that the detection of intestinal sounds by analyzing spectrograms using convolutional and recursive neural networks is possible, provides useful results and is therefore worth further research.

5. Conclusions

The proposed design aims to provide a solution for the analysis (classification) of BS, which would first be implemented for research purposes and then in practice. Our hypothesis states that it is possible to develop such an algorithm and employ it as a classifier of intestinal sounds.
As presented in Section 3, our algorithm achieved 97.7% accuracy, 99.0% specificity, and 77.3% sensitivity with new, high-quality labeled dataset of 321,000 records.
The methods of automatic analysis of intestinal sounds developed so far have used many different algorithmic approaches. A common feature was high efficiency but, unfortunately, a surprising lack of broader application in practice. The lack of application was due to various reasons. Practical microphones and recorders, as well as practical clinical protocols, were lacking. As our experience shows, this simple problem is a significant challenge that can only be overcome through interdisciplinary cooperation. Past researchers have systematically ignored the night-time period, which is the most interesting from a neurogastroenterological point of view. Instead, they have kept short records during the day or the first meal in the morning. The research goals in the past were preliminary and, therefore, sometimes abstract and in need of a broader application perspective and motor context. However, there is no doubt that automatic identification and analysis of intestinal sounds will be used in practice.
The main advantage of the presented approach is high quality in terms of accuracy, specificity and sensitivity. Moreover, we eliminate the problem of recording privacy (no conversations, enables data sharing). Long overnight recordings that are more likely to capture phase III of the migrating motor complex and are less influenced by eating habits, and the minimalization of external noise otherwise resulting from the daytime activity. The principal limitations of this scheme involve single sensor recordings, no information on the reaction of the gastrointestinal system to the first meal, and the lack of wireless capabilities. Nevertheless, at little cost and entirely non-invasively, highly useful information on BS may be obtained. The analysis using neural networks is computationally more expensive than some basic approaches used in the past. Still, the technology is advancing, and our model is easily deployed on new edge devices. Moreover, rapid analysis of the obtained recording should be achievable with a typical personal computer.
Our tool also has an interface designed for quick and easy tagging of the location of detected sounds. If connected to the recording device, we believe that the tool could provide a significant amount of important information related to the patient’s digestive system in real-time. Without a doubt, it will encourage more medical practitioners to record and tag their patients’ intestine activity, which, in turn, would enlarge the training dataset for our algorithm.

Author Contributions

J.F., K.R., R.N. and J.K.N. developed the solution. J.F., K.R. and R.N. project the software and numerical experiments. J.F. and K.R. implemented and tested software. J.K.N. recorded and annotated the data. J.F., K.R., J.K.N. and R.N. interpreted the results and prepared the manuscript. O.Y. and J.W. supported the work. R.N. provided funding. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Warsaw University of Technology grant BIOTECHMED.

Institutional Review Board Statement

The study was conducted according to the revised Declaration of Helsinki and approved by the Bioethical Committee at Poznan University of Medical Sciences, Poznan, Poland (683/18).

Informed Consent Statement

Written informed consent was obtained from all patients involved in the study and/or their parents, as adequate according to local regulations.

Data Availability Statement

Dataset is freely available on Kaggle, doi:10.34740/kaggle/dsv/2789758, https://www.kaggle.com/robertnowak/bowel-sounds, accessed on 13 November 2021, under CC BY-NC 4.0 License.

Acknowledgments

We would like to thank Przemysław Szelejewski (Poznan) and Stefan Stróżecki (Bydgoszcz) for their work on developing the contact microphone. The authors would like to thank patients for their participation in the study and Hanna Wielińska-Wiśniewska (Poznan) for collecting the data.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Sprung, H. Measurement results with the ingestable intestinal sound in the gastrointestinal canal. Langenbecks Arch. Klin. Ver. Mit Dtsch. Z. Chir. 1959, 291, 80–97. [Google Scholar]
  2. Dalle, D.; Devroede, G.; Thibault, R.; Perrault, J. Computer analysis of bowel sounds. Comput. Biol. Med. 1975, 4, 247–256. [Google Scholar] [CrossRef]
  3. Craine, B.L.; Silpa, M.; O’Toole, C.J. Computerized auscultation applied to irritable bowel syndrome. Dig. Dis. Sci. 1999, 44, 1887–1892. [Google Scholar] [CrossRef] [PubMed]
  4. Hadjileontiadis, L.J. Wavelet-based enhancement of lung and bowel sounds using fractal dimension thresholding-Part I: Methodology. IEEE Trans. Biomed. Eng. 2005, 52, 1143–1148. [Google Scholar] [CrossRef] [PubMed]
  5. Dimoulas, C.; Kalliris, G.; Papanikolaou, G.; Petridis, V.; Kalampakas, A. Bowel-sound pattern analysis using wavelets and neural networks with application to long-term, unsupervised, gastrointestinal motility monitoring. Expert Syst. Appl. 2008, 34, 26–41. [Google Scholar] [CrossRef]
  6. Ranta, R.; Louis-Dorr, V.; Heinrich, C.; Wolf, D.; Guillemin, F. Principal component analysis and interpretation of bowel sounds. In Proceedings of the 26th Annual International Conference of the IEEE Engineering in Medicine and Biology Society, San Francisco, CA, USA, 1–4 September 2004; Volume 1, pp. 227–230. [Google Scholar]
  7. Kodani, K.; Sakata, O. Automatic Bowel Sound Detection under Cloth Rubbing Noise. In Proceedings of the 2020 IEEE REGION 10 CONFERENCE (TENCON), Osaka, Japan, 16–19 November 2020; pp. 779–784. [Google Scholar]
  8. Kim, K.S.; Seo, J.H.; Song, C.G. Non-invasive algorithm for bowel motility estimation using a back-propagation neural network model of bowel sounds. Biomed. Eng. Online 2011, 10, 1–10. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  9. Ulusar, U.D. Recovery of gastrointestinal tract motility detection using Naive Bayesian and minimum statistics. Comput. Biol. Med. 2014, 51, 223–228. [Google Scholar] [CrossRef] [PubMed]
  10. Kölle, K.; Fougner, A.L.; Ellingsen, R.; Carlsen, S.M.; Stavdahl, Ø. Feasibility of early meal detection based on abdominal sound. IEEE J. Transl. Eng. Health Med. 2019, 7, 1–12. [Google Scholar] [CrossRef] [PubMed]
  11. Yin, Y.; Jiang, H.; Feng, S.; Liu, J.; Chen, P.; Zhu, B.; Wang, Z. Bowel sound recognition using SVM classification in a wearable health monitoring system. Sci. China Inf. Sci. 2018, 61, 1–3. [Google Scholar] [CrossRef]
  12. Huang, Y.; Song, I.; Rana, P.; Koh, G. Fast diagnosis of bowel activities. In Proceedings of the 2017 International Joint Conference on Neural Networks (IJCNN), Anchorage, AK, USA, 14–19 May 2017; pp. 3042–3049. [Google Scholar]
  13. Liu, J.; Yin, Y.; Jiang, H.; Kan, H.; Zhang, Z.; Chen, P.; Zhu, B.; Wang, Z. Bowel Sound Detection Based on MFCC Feature and LSTM Neural Network; IEEE: Cleveland, OH, USA, 2018; pp. 1–4. [Google Scholar] [CrossRef]
  14. Du, X.; Allwood, G.; Webberley, K.M.; Osseiran, A.; Marshall, B.J. Bowel sounds identification and migrating motor complex detection with low-cost piezoelectric acoustic sensing device. Sensors 2018, 18, 4240. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Nowak, J.; Nowak, R.; Radzikowski, K.; Grułkowski, I.; Walkowiak, J. Automated Bowel Sound Analysis: An Overview. Sensors 2021, 21, 5294. [Google Scholar] [CrossRef] [PubMed]
  16. Allwood, G.; Du, X.; Webberley, K.M.; Osseiran, A.; Marshall, B.J. Advances in acoustic signal processing techniques for enhanced bowel sound analysis. IEEE Rev. Biomed. Eng. 2018, 12, 240–253. [Google Scholar] [CrossRef] [PubMed]
  17. Mansy, H.A.; Sandler, R.H. Bowel-sound signal enhancement using adaptive filtering. IEEE Eng. Med. Biol. Mag. 1997, 16, 105–117. [Google Scholar] [CrossRef] [PubMed]
  18. Ranta, R.; Louis-Dorr, V.; Heinrich, C.; Wolf, D.; Guillemin, F. A complete toolbox for abdominal sounds signal processing and analysis. In Proceedings of the 3rd European Medical and Biological Engineering Conference IFMBE-EMBEC, Portorož, Slovenia, 29 November–3 December 2005; Volume 5. [Google Scholar]
  19. Zhao, K.; Feng, S.; Jiang, H.; Wang, Z.; Chen, P.; Zhu, B.; Duan, X. A Binarized CNN-based Bowel Sound Recognition Algorithm with Time-domain Histogram Features for Wearable Healthcare Systems. IEEE Trans. Circuits Syst. II Express Briefs 2021, 1. [Google Scholar] [CrossRef]
  20. Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. arXiv 2017, arXiv:1412.6980. [Google Scholar]
  21. Kukanov, I.; Hautamäki, V.; Lee, K.A. Maximal figure-of-merit embedding for multi-label audio classification. In Proceedings of the 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018; pp. 136–140. [Google Scholar]
  22. Emoto, T.; Abeyratne, U.; Gojima, Y.; Nanba, K.; Sogabe, M.; Okahisa, T.; Akutagawa, M.; Konaka, S.; Kinouchi, Y. Evaluation of human bowel motility using non-contact microphones. Biomed. Phys. Eng. Express 2016, 2, 045012. [Google Scholar] [CrossRef]
  23. Hadjileontiadis, L.; Rekanos, I. Detection of explosive lung and bowel sounds by means of fractal dimension. IEEE Signal Process. Lett. 2003, 10, 311–314. [Google Scholar] [CrossRef]
  24. Inderjeeth, A.J.; Webberley, K.; Muir, J.; Marshall, B. The potential of computerised analysis of bowel sounds for diagnosis of gastrointestinal conditions: A systematic review. Syst. Rev. 2018, 7, 1–18. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  25. Kumar, T.; Soiland, E.; Stavdahl, O.; Fougner, A. Pilot Study of Early Meal Onset Detection from Abdominal Sounds. In Proceedings of the 2019 E-Health and Bioengineering Conference (EHB), Iasi, Romania, 21–23 November 2019. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Flowchart of the method developed. The first optional step of the algorithm is converting audio into a single channel WAV format with 24 bit, 44.1 kHz sampling and represent it as a collection of 2-s recordings. Next, we remove the sounds outside interesting spectra using Fast Fourier Transform and Low Pass Filter. The signal is then converted into a spectrogram and the spectrogram obtained is divided into a sequence of frames. Finally, every frame (spectrogram) is classified using machine learning algorithms, with the returning binary value representing the frame as either bowel sound or noise. We calculate the output from adjacent frames.
Figure 1. Flowchart of the method developed. The first optional step of the algorithm is converting audio into a single channel WAV format with 24 bit, 44.1 kHz sampling and represent it as a collection of 2-s recordings. Next, we remove the sounds outside interesting spectra using Fast Fourier Transform and Low Pass Filter. The signal is then converted into a spectrogram and the spectrogram obtained is divided into a sequence of frames. Finally, every frame (spectrogram) is classified using machine learning algorithms, with the returning binary value representing the frame as either bowel sound or noise. We calculate the output from adjacent frames.
Sensors 21 07602 g001
Figure 2. Smoothed learning curves representing the accuracy of the linking model on the validation set, depending on the maximum frequency of the spectrogram max_freq.
Figure 2. Smoothed learning curves representing the accuracy of the linking model on the validation set, depending on the maximum frequency of the spectrogram max_freq.
Sensors 21 07602 g002
Figure 3. Smoothed learning curves representing the accuracy of the linking model on the validation set, depending on the window width fft for 3 different values.
Figure 3. Smoothed learning curves representing the accuracy of the linking model on the validation set, depending on the window width fft for 3 different values.
Sensors 21 07602 g003
Figure 4. Hann and Hamming window. The x-axis represents a number of sample n in signal.
Figure 4. Hann and Hamming window. The x-axis represents a number of sample n in signal.
Sensors 21 07602 g004
Figure 5. Smoothed learning curves showing the accuracy of the connecting model on the validation set, depending on the number of frames.
Figure 5. Smoothed learning curves showing the accuracy of the connecting model on the validation set, depending on the number of frames.
Sensors 21 07602 g005
Figure 6. CRNN model, combination of convolutional layers (a feature extractors) and recursive layers.
Figure 6. CRNN model, combination of convolutional layers (a feature extractors) and recursive layers.
Sensors 21 07602 g006
Figure 7. Classification model combining a convolutional network and dense network, to classify using adjacent frames results, called the CDNN.
Figure 7. Classification model combining a convolutional network and dense network, to classify using adjacent frames results, called the CDNN.
Sensors 21 07602 g007
Table 1. Base algorithm results. The cut-off value is the mean value of the amplitudes of the standardized spectrogram, above which the algorithm considers the frame a bowel sound.
Table 1. Base algorithm results. The cut-off value is the mean value of the amplitudes of the standardized spectrogram, above which the algorithm considers the frame a bowel sound.
Cut-Off ValueACC [%]Prec. [%]Sensitivity [%]Specificity [%]
0.2085.2097.880.0299.99
0.1086.1394.130.0899.90
0.0886.4688.3911.8899.72
0.0786.5582.4913.8399.48
0.0686.4773.5116.2798.96
0.0586.1262.9119.6197.94
Table 2. Results of the linking model on the validation set, depending on the maximum frequency of the spectrogram max freq. The data come from the best eras. They were obtained by the fivefold cross-validation method.
Table 2. Results of the linking model on the validation set, depending on the maximum frequency of the spectrogram max freq. The data come from the best eras. They were obtained by the fivefold cross-validation method.
Max Freq [Hz]ACC [%]Prec. [%]Sensitivity [%]Specificity [%]
50090.9283.0250.0498.18
100092.6986.2161.3898.25
150092.9686.5563.1998.25
200092.9387.1962.3498.37
300092.7985.7162.7498.14
Table 3. The results of the link model on the validation set, depending on the window width fft. The data is from the best of the era. They were obtained by the fivefold cross-validation method.
Table 3. The results of the link model on the validation set, depending on the window width fft. The data is from the best of the era. They were obtained by the fivefold cross-validation method.
fftACC [%]Prec. [%]Sensitivity [%]Specificity [%]
176491.8484.5056.2198.17
88292.5086.9359.2498.42
44192.9285.7263.6698.11
22092.3386.2958.4998.35
11091.6184.5754.2998.24
Table 4. The results of the linking model on the validation set, depending on the type of window function used. The data is from the best of the era. They were obtained by the fivefold cross-validation method.
Table 4. The results of the linking model on the validation set, depending on the type of window function used. The data is from the best of the era. They were obtained by the fivefold cross-validation method.
SmoothingACC [%]Prec. [%]Sensitivity [%]Specificity [%]
Hann92.9987.6762.3298.44
Hamming92.7887.0861.8398.39
Table 5. The results of the linking model on the validation set, depending on the number of frames. The data come from the best eras. They were obtained by the fivefold cross-validation method.
Table 5. The results of the linking model on the validation set, depending on the number of frames. The data come from the best eras. They were obtained by the fivefold cross-validation method.
Num FramesACC [%]Prec. [%]Sensitivity [%]Specificity [%]
191.0786.4548.3898.65
392.0686.3056.3798.41
592.6185.8361.1398.21
792.8487.5461.3398.44
992.9686.5563.1998.25
1193.0087.5162.5598.41
1393.0487.0863.3498.33
1593.0486.5863.8298.24
Table 6. Results of the linking model on the validation set, depending on the variance of Gaussian noise. From each data sequence, 5 new sequences were created by adding noise with variable standard deviation σ . The data comes from the best eras. They were obtained by the fivefold cross-validation method.
Table 6. Results of the linking model on the validation set, depending on the variance of Gaussian noise. From each data sequence, 5 new sequences were created by adding noise with variable standard deviation σ . The data comes from the best eras. They were obtained by the fivefold cross-validation method.
σ ACC [%]Prec. [%]Sensitivity [%]Specificity [%]
no92.9087.7161.5998.47
0.0192.9886.0563.8598.16
0.0292.9186.9362.4898.33
0.0392.9787.0262.7998.33
0.0592.9286.9362.4898.33
0.1092.5987.2859.5698.46
Table 7. Results of the models on the validation set, depending on the type of architecture. The data come from the best epochs. They were obtained by the fivefold cross-validation method.
Table 7. Results of the models on the validation set, depending on the type of architecture. The data come from the best epochs. They were obtained by the fivefold cross-validation method.
ModelACC [%]Prec. [%]Sensitivity [%]Specificity [%]
CRNN93.2485.5666.4398.01
CDNN92.9786.5563.2198.26
Table 8. Results of the best recursive and combining model on the test set. The models were trained on a set of all 5 validation subsets. The ratio of the number of intestinal sound frames to the total number of frames in the test set is 15%.
Table 8. Results of the best recursive and combining model on the test set. The models were trained on a set of all 5 validation subsets. The ratio of the number of intestinal sound frames to the total number of frames in the test set is 15%.
ModelACC [%]Prec. [%]Sensitivity [%]Specificity [%]
CRNN97.782.777.399.0
CDNN97.483.371.199.1
Table 9. Results of additional testing on the new annotated audio recording.
Table 9. Results of additional testing on the new annotated audio recording.
ModelACC [%]Prec. [%]Sensitivity [%]Specificity [%]
CRNN98.1057.5785.6398.41
CDNN97.8754.2084.7398.20
Table 10. Comparison of our algorithm with others (in alphabetical order). We present results provided by authors.
Table 10. Comparison of our algorithm with others (in alphabetical order). We present results provided by authors.
AuthorACCYearAlgorithm
Emoto et al. [22]91%2013–2018FFT, moving average
Hadjileontiadis et al. [23]95%1999–2011perceptron
Inderjeeth et al. [24]87%2018–2020machine learning
Kumar et al. [25]75%2019SVM
Liu et al. [13]92%2018–2019CNN
Ulusar et al. [9]94%2014Bayesian classifier
This97% CNN + RNN
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Ficek, J.; Radzikowski, K.; Nowak, J.K.; Yoshie, O.; Walkowiak, J.; Nowak, R. Analysis of Gastrointestinal Acoustic Activity Using Deep Neural Networks. Sensors 2021, 21, 7602. https://doi.org/10.3390/s21227602

AMA Style

Ficek J, Radzikowski K, Nowak JK, Yoshie O, Walkowiak J, Nowak R. Analysis of Gastrointestinal Acoustic Activity Using Deep Neural Networks. Sensors. 2021; 21(22):7602. https://doi.org/10.3390/s21227602

Chicago/Turabian Style

Ficek, Jakub, Kacper Radzikowski, Jan Krzysztof Nowak, Osamu Yoshie, Jaroslaw Walkowiak, and Robert Nowak. 2021. "Analysis of Gastrointestinal Acoustic Activity Using Deep Neural Networks" Sensors 21, no. 22: 7602. https://doi.org/10.3390/s21227602

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop