Next Article in Journal
An Underwater Source Localization Method Using Bearing Measurements
Previous Article in Journal
Context-Based e2e Autonomous Operation in B5G Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Breath Measurement Method for Synchronized Reproduction of Biological Tones in an Augmented Reality Auscultation Training System

1
Department of Medical Engineering, Graduate School of Science and Engineering, Chiba University, 1-33 Yayoicho, Inage-ku, Chiba-shi 263-8522, Chiba, Japan
2
Department of Cardiovascular Medicine, Graduate School of Medicine, Chiba University, 1-8-1 Inohana, Chuo-ku, Chiba-shi 260-8670, Chiba, Japan
3
Department of Respirology, Graduate School of Medicine, Chiba University, 1-8-1 Inohana, Chuo-ku, Chiba-shi 260-8670, Chiba, Japan
4
Department of Medical Education, Graduate School of Medicine, Chiba University, 1-8-1 Inohana, Chuo-ku, Chiba-shi 260-8670, Chiba, Japan
5
Chiba University Hospital, 1-8-1 Inohana, Chuo-ku, Chiba-shi 260-8677, Chiba, Japan
6
Chiba University, 1-33 Yayoicho, Inage-ku, Chiba-shi 263-8522, Chiba, Japan
7
Center for Frontier Medical Engineering, Chiba University, 1-33 Yayoicho, Inage-ku, Chiba-shi 263-8522, Chiba, Japan
*
Authors to whom correspondence should be addressed.
Sensors 2024, 24(5), 1626; https://doi.org/10.3390/s24051626
Submission received: 4 February 2024 / Revised: 22 February 2024 / Accepted: 23 February 2024 / Published: 1 March 2024
(This article belongs to the Section Biomedical Sensors)

Abstract

:
An educational augmented reality auscultation system (EARS) is proposed to enhance the reality of auscultation training using a simulated patient. The conventional EARS cannot accurately reproduce breath sounds according to the breathing of a simulated patient because the system instructs the breathing rhythm. In this study, we propose breath measurement methods that can be integrated into the chest piece of a stethoscope. We investigate methods using the thoracic variations and frequency characteristics of breath sounds. An accelerometer, a magnetic sensor, a gyro sensor, a pressure sensor, and a microphone were selected as the sensors. For measurement with the magnetic sensor, we proposed a method by detecting the breathing waveform in terms of changes in the magnetic field accompanying the surface deformation of the stethoscope based on thoracic variations using a magnet. During breath sound measurement, the frequency spectra of the breath sounds acquired by the built-in microphone were calculated. The breathing waveforms were obtained from the difference in characteristics between the breath sounds during exhalation and inhalation. The result showed the average value of the correlation coefficient with the reference value reached 0.45, indicating the effectiveness of this method as a breath measurement method. And the evaluations suggest more accurate breathing waveforms can be obtained by selecting the measurement method according to breathing method and measurement point.

1. Introduction

Auscultation of breath and heart sounds constitutes one of the clinical skills in physical examination; auscultation is a medical technique used to listen to lung airflow and heart sounds to obtain biological information noninvasively and quickly [1]. Since these sounds vary greatly depending on the auscultation position and presence of diseases, auscultation requires sufficient training for clinical practice [2]. Recent studies have demonstrated a decline in auscultation skills in clinical practice, and there is growing interest among practitioners to improve auscultation training [3,4,5,6]. Against this backdrop, current medical educational institutions conduct the objective structured clinical examination (OSCE) before students advance to clinical practice [7,8]. This examination aims to evaluate the proficiency of clinical skills such as “basic examination skills” and “communication skills with patients”, which are difficult to evaluate by paper tests.
In current auscultation training and OSCE, medical interviews and auscultation techniques are performed on a simulated or standardized patient (SP) who reproduces the physical findings of the disease. Since the SP is healthy, the numbers of breath and heart sounds that can be reproduced are limited. Hence, many medical education institutions use a method that combines medical interviews with auscultation procedures on a mannequin-type auscultation simulator [9,10]. Simulators have been shown to enhance physical examination skills of learners, and learners have also been shown to value simulation-based teaching very highly [11,12]. However, this method requires the trainees to turn around and face the simulator when shifting from the medical interview to the auscultation procedure, which may reduce the reality and training efficiency. In addition, the simulator is very expensive and requires ample space for storage and use, making it difficult to implement in facilities that are not large.
To overcome these problems, Nakaguchi et al. [13] proposed a virtual auscultation simulator for auscultation training on an SP and showed its effectiveness in reproducing training by comparison with a mannequin-type simulator. In addition, Sekiguchi et al. [14,15] proposed the educational augmented reality auscultation system (EARS), an augmented-reality-based auscultation training system, using deep learning; this system reproduces the breath and heart sounds of various diseases on a healthy SP and enables medical interviews as well as auscultation procedures to be performed on the SP. The trainee performs auscultation on the SP using a special stethoscope, similar to a regular medical examination; this stethoscope is divided into the chest piece and camera unit parts (Figure 1). The chest piece part has a built-in contact sensor that determines whether the patient is being auscultated by the presence or absence of contact with the body. The camera in the camera unit part acquires the positions of the chest piece during auscultation along with those of the SP’s head, shoulders, and hips during non-auscultation to calculate the auscultation position on the SP’s body based on the position information. EARS also plays the sounds of selected diseases according to the calculated auscultation positions.
However, when using the conventional system, the SP must breathe according to the timing shown on the breath indicator mounted on the camera unit of the stethoscope to match the rhythm of thoracic variations and breath sounds played. This burdens the SP heavily and makes it impossible to respond to deep breaths and breath holding. To solve this, it is necessary to measure the breathing of the SP during auscultation training.
Breath measurement methods can be broadly classified into contact and noncontact types. As a contact-type breath measurement method, a belt with an inertial or a piezoelectric sensor is attached to the chest [16,17]. This method measures the thoracic variations caused by breathing using sensors built into the belt. Another contact-type method involves attaching a microphone to the chest and investigating the exhalation and inhalation sounds of breathing to estimate respiration [18]. The disadvantage of such contact-type methods is the increased preparation time required for attaching the sensor. As a noncontact breath measurement method, the movements of the thorax are acquired using a depth camera and an ultrasonic proximity sensor [19,20]. This method involves monitoring breathing motions by measuring the distance to the thorax from the camera or sensor. However, the main disadvantages of this method are that the accuracy of acquisition depends on the positional relationship between the SP and camera or sensor, which requires time to setup, and that the accuracies of the breath measurements depend on the movements of the SP and wrinkles in the clothing accompanying the movements. There is also a method to measure breathing based on temperature changes around the nasal cavity of the SP using a thermal imaging camera [21,22]; however, this method has the disadvantage that it can only be applied when the SP is facing the front of the camera. These noncontact types have the disadvantage of requiring unique cameras, and when mounted, the EARS device becomes expensive.
In the present study, we propose a method to measure breathing by contact during auscultation training to realize the reproduction of breath sounds based on the breathing of the SP in EARS. To reduce the time and effort required to attach the sensor to the SP, which is a drawback of contact-type breath measurement, a method that can be integrated into the chest piece of a stethoscope is investigated. We propose a new principle for real-time acquisition of breathing waveforms, then evaluate the method and discuss how it can be used in a system. The remainder of this manuscript is organized as follows. Section 2 describes the proposed breath measurement methods. Section 3 presents the experimental setup and measurement accuracy of each method. Section 4 presents analyses of the measurement accuracies and discusses their prospects. Finally, Section 5 presents the conclusions of this study.

2. Materials and Methods

In this study, we investigate two methods for measuring breathing using thoracic variations and breathing sounds. For measurement devices, we developed ones that can be measured using only a chest piece and that are close to the size and weight of an actual chest piece to maintain the realism of the actual auscultation procedure.

2.1. Breath Measurement Using Thoracic Variations

Humans breathe by taking oxygen into and expelling carbon dioxide from the alveoli. Breathing is mainly achieved by the diaphragm and external intercostal muscles, which expand and contract the lungs. The movements of the diaphragm and external intercostal muscles cause thoracic variations [23]. During inhalation, the diaphragm and external intercostal muscles contract. When the diaphragm contracts and descends, the contraction of the external intercostal muscles expands the thorax back and forth so that the lungs expand, resulting in inhalation. During exhalation, the diaphragm and external intercostal muscles relax. The diaphragm returns to its original position, while the thorax narrows and lungs contract by elastic contraction. In this study, we propose a method to measure breathing motions using such movements of the thorax.

2.1.1. Thoracic Variability Measurement Device Design and Data Collection

A stethoscope-type measurement device was fabricated using a 3D printer to measure the thoracic variations; this device is shown in Figure 2. With this device, breath measurements can be performed by placing the device on the SP’s chest as in the actual auscultation technique. The side and bell parts of the chest piece are made of polylactic acid resin, and the diaphragm surface is made of a deformable thermoplastic polyurethane resin. A pressure sensor FSR402 (Interlink Electronics Inc., Camarillo, CA, USA), an indenter, and a magnet are placed on the diaphragm surface and are connected to a breadboard and a 9-axis inertial sensor Adafruit Precision NXP 9-DOF Breakout Board—FXOS8700 + FXAS21002 (Adafruit Industries, New York City, NY, USA). The 9-axis inertial sensor includes an accelerometer, a gyro sensor, and a magnetic sensor. The characteristics of these sensors are shown in Table 1 and Table 2. Arduino UNO (Arduino Holding, Ivrea, Italy) was used to interface the sensors with the PC.

2.1.2. Measurement by Parallel Shift Movement

An accelerometer was used to measure the movements of the thorax in terms of the amount of parallel shift. In addition, a measurement mechanism using a magnetic sensor and a magnet was developed to measure the amount of movement of the stethoscope surface caused by the thoracic movements. In this mechanism, a small permanent magnet is placed on the stethoscope surface, which is the contact surface, and the magnetic sensor measures the slight variations in the magnet due to deformation of the surface from changes in the magnetic field. The measurement principle is shown in Figure 3. A lowpass filter was then applied to each sensor output and processed high-frequency noise by the circuitry and unstable signal components.

2.1.3. Measurement by Angular Displacement

Accelerometer and gyro sensors were used to measure the movements of the thorax as rotational motions. The sensor values were corrected using a lowpass filter and a Kalman filter.

2.1.4. Pressure Measurement

A pressure sensor was used to measure the pressure on the stethoscope’s chest piece in contact with the measurement surface caused by the back-and-forth movements of the thorax due to breathing motions. The pressure sensor used here is the one described in Section 2.1.1. Here, an indenter for noise reduction was created with a 3D printer and placed on the back of the pressure sensor, as shown in Figure 2.

2.2. Breath Measurements Using Breath Sounds

The overall sounds produced by breathing are called lung sounds, which are classified into two categories as breath sounds (airflow sounds during ventilation of the airways and alveoli) and sub-noise (abnormal sounds generated in pathological conditions) [24]. In this study, we focus mainly on the breathing sounds. Three types of breath sounds can be heard under normal conditions: bronchial, bronchoalveolar, and alveolar. Bronchial breath sounds are louder than those heard from the other parts of the body and are particularly louder during exhalation than inhalation, with a clear pause between inhalation and exhalation. The frequency components of the sound are around 600 Hz for inhalation and 400 Hz for exhalation. Bronchoalveolar breath sounds can be heard more clearly than alveolar breath sounds during both inhalation and exhalation. The loudness of the breath sounds is the same for inhalation and exhalation, or slightly louder for inhalation, and the frequency components are the same as those for bronchial breath sounds, which are around 600 Hz for inhalation and around 400 Hz for exhalation. Alveolar breath sounds are clearly audible upon inhalation, but the exhalation sounds are quiet and difficult to hear. The ratio of inhalation to exhalation time is approximately 1:2. The frequency components of the inhalation sounds are around 400 Hz and those of the exhalation sounds are around 200 Hz [25,26]. We propose a method to measure the breathing phase using these differences between the exhalation and inhalation sounds.

2.2.1. Related Research

Li et al. [27] proposed a breath sound analysis algorithm for real-time whistle sound detection; this algorithm extracts the characteristics of breath sounds in both time and frequency domains and classifies normal breath sounds and whistle sounds by taking advantage of the frequency range of the whistle sounds (250–800 Hz).
The steps in this method are as follows:
  • The breath sounds are bandpass filtered in a passband of 150–1000 Hz. This step is used to remove heart sounds and signals from muscle interference [28].
  • The data are subjected to a Fourier transform using a Hanning window (short-time Fourier transform). Spectral integration (SI) and normalized spectral integration (NSI) are then calculated to obtain the characteristics of the breath sounds in the frequency domain. The NSI is the ratio of the value of each SI when the sum of the calculated SI is set to 1.
  • The NSI values are used to classify normal breath sounds from whistle sounds using Fisher linear discriminant analysis.

2.2.2. Proposed Methodology for this Study

In this study, we propose a breathing waveform acquisition method using the ratio of SIs as well as the SI and NSI described above, as outlined in Figure 4. First, the breath sounds are denoised using a moving average filter of size 10 samples, and a short-time Fourier transform is then performed using a Hanning window. The window width is set to 4096 points and overlapped so that the frameshift is 50 ms. The ratio of spectral integrals is then calculated from the obtained spectra. As shown in Figure 5, in this study, the ratio of SIs in the range of 500–800 Hz out of those in the range of 100–1000 Hz, which is the frequency range of breath sounds, is calculated as the values of the breathing waveform. The calculated values are then multiplied with a moving average filter of size 10 samples and used as the final breathing waveform acquired by this method.

2.2.3. Breath Sound Sampling Device Design and Data Collection

A stethoscope-type breath sound sampling device with a microphone inserted into the stethoscope tube was fabricated to measure the breath sounds. The fabricated device is shown in Figure 6. The stethoscope used was a Flare Phonet No. 137II (KENZMEDICO CO. LTD., Saitama, Japan), and the microphone used was ECM-PC60 (Sony Group Corporation, Tokyo, Japan). The microphone characteristics are shown in Table 3; the microphone was connected to a laptop computer via a conversion adapter, and its sampling rate was 44.1 kHz.
The chest piece has two types of listening surfaces, namely the diaphragm and bell surfaces. In this experiment, the diaphragm surface was used to measure the breath sounds.

3. Results

3.1. Experimental Procedures

A total of 14 subjects (7 females and 7 males) in their early 20s without any health problems were recruited for the study. Since auscultation is performed on healthy subjects in the EARS system, we selected healthy subjects for this breath measurement. Informed consent was obtained from all subjects involved in the study. Based on actual EARS use, measurements were obtained while the subjects were wearing clothing. Five measurement points (four in the front and one in the back) were selected from the basic auscultation positions for the breath sounds (Figure 7). In this study, measurements were taken only on the right side of the body, which is less affected by the heartbeat. To reproduce an actual medical interview, each subject was instructed to take either normal or deep breaths for each of the measurements. The amount of movement of the chest and abdomen varies depending on the breathing method, which may result in differences in accuracy depending on the measurement point and method [29]. Three 20-s measurements were acquired per subject at each measurement position and for each breathing method. All subjects wore the same cotton T-shirt to minimize any influence from clothing. The experiments were conducted in a sitting position to simulate a medical interview.
The signal from the Go Direct Respiration Belt (Vernier Science Education, Beaverton, OR, USA) was used as the reference for the breathing waveform. The specifications of the sensor are shown in Table 4. The sensor was placed above the subject’s dovetail (blue circle in Figure 7). From now on, the signals obtained with this respiration measurement belt are referred to as the biometric sensor signals.

3.2. Evaluation Method

The normalized cross-correlation function was used as the evaluation index. The equation of the cross-correlation R for two time-series waveforms ( x , y ) and number of data N is shown below.
R ( t ) = 1 N t t = 1 N t x t y ( t + t )
After normalizing the reference biometric sensor signal and breathing waveform signals from the data acquired by the thoracic variation measurement device and breath sound sampling device, cross-correlation analysis was performed to obtain the correlation coefficient of the waveform with the reference signal. Then, an analysis of variance was performed on the obtained correlation coefficients to test the differences in accuracies of acquiring breathing waveforms by different measurement methods. For the analysis of variance, anovakun ver4.8.6—a script for analysis of variance on the statistical software R—was used. The version of the statistical software R used was 4.0.0.

3.3. Experimental Results

Figure 8 shows the breathing waveforms obtained for each measurement method, and Figure 9 compares the cross-correlation coefficients among the measurement methods. It was confirmed that the average value of the correlation coefficient for each measurement method was more than 0.4, which is considered to be correlated. However, the final target value for the correlation coefficient is 0.7, which is considered a strong correlation, so the results indicate that we have not yet reached that level.
A three-factor analysis of variance for mixed designs was conducted for the number of interrelationships based on the breathing method (within-participant items: 2 levels), measurement position (within-participant items: 5 levels), and measurement technique (within-participant items: 5 levels). These results are shown in Table 5; the results show that the second-order interactions were insignificant. However, significant differences were obtained in the first-order interactions between the breathing and measurement method factors and also in the first-order interactions between the measurement position and measurement method factors. Table 6 shows the means and standard deviations for the corresponding breathing method and measurement technique factors. Table 7 shows the means and standard deviations for the measurement position and technique factors. Table 8 shows the result of simple main effects for each interaction, and Table 9 shows the results of Bonferroni’s multiple comparisons for those factors for which the main effects were significant. In the table, SS represents the sum of squares, Df represents degrees of freedom, and MS means mean squares.

4. Discussion

The breathing method and measurement position are factors that change the thoracic variations and types of breath sounds. The analysis of variance showed that differences in the accuracy of acquiring breathing waveforms between the measurement methods based on thoracic variations were caused by differences in the breathing method and measurement position.
In terms of the breathing method, deep breaths tended to be abdominal breathing, compared to normal breaths, and we considered a significant difference in the accuracy of breathing waveform acquisition [29]. Breathing can be broadly classified into two types: as costal and abdominal breathing. Costal breathing is a method in which mainly the thorax moves, and the muscles that work during inhalation are the external intercostals and respiratory accessory muscles, while the muscles working during exhalation are the internal intercostals. During inhalation, the external intercostals and several respiratory accessory muscles are strongly contracted to raise the thorax anteriorly and upward to assist inhalation. During exhalation, the internal intercostals contract to contract the thorax and assist in exhalation. Abdominal breathing is a method in which the movement of the diaphragm is the main component. The diaphragm is the muscle that works during inhalation, and the abdominal muscles work during exhalation. During inhalation, the diaphragm contracts strongly, expanding the thorax in a large vertical direction. During inhalation, the diaphragm contracts strongly, and the thorax expands vertically. Thus, the differences in the movements of the thorax depending on the breathing method are considered to be the reason for the significant differences among the three methods for detecting the movements of the thorax. The results of the analysis of variance showed that the interactions between the breathing and measurement methods were significant and that these effects were considerable. Considering this, we selected the pressure sensor for normal breath and magnetic sensor for deep breath as the measurement methods and analyzed the results again, as shown in Table 10. The accuracy of breath measurement was improved compared to the case where no method was selected, suggesting that the selection of the measurement method according to the breathing method is effective.
The correlation coefficients were significantly different at the measurement points due to differences in thoracic variations by position and breath sounds. The correlation coefficients for the magnetic sensor measurements were significantly higher at measurement points (4) and (2) than at the other points. This is likely because the measurement points are closer to the upper thorax and abdomen than the other points, and thorax variation is more significant in these points. In addition, the correlation coefficients were significantly higher at the measurement point (2) than at the other measurement points for the accelerometer and gyro sensor measurements. This is because the measurement point (2) is close to the top of the thorax, and the stethoscope is placed at an angle to the body compared to other points, so the thoracic variations are easily expressed as angular displacements. Thus, the accuracy of breathing waveform acquisition may change depending on the difference in thoracic variation depending on position. These results are used to select the most appropriate breath measurement method according to the measurement position. The correlation coefficient for breath sounds was lower near measurement point (4) than at other points, but we consider that this was due to the fact that it was in the peripheral region of the bronchi. This is because the breath sounds are generated only in the bronchi from the oral cavity to the seventh to ninth branches, which are turbulent regions, and the breath sounds therefore tend to become lower than those in other regions as they get closer to the end region of the bronchi, being easily affected by noise such as clothing. In addition, in the peripheral region of the bronchi, the percentage of alveolar sounds is higher than bronchial sounds in the collected breath sounds. Therefore, as described in Section 2.2, the frequency range of exhalation and inhalation sounds tended to be lower than that of the bronchial sounds, making it difficult to obtain breathing waveforms in the frequency range selected for this experiment. In this work, the ratio of SI in the range of 500–800 Hz was acquired as the breathing waveform. However, if the ratio of SI in the range of 300–600 Hz was acquired as the breathing waveforms for the alveolar sounds, the results would be as shown in Table 11. Therefore, acquiring more accurate breathing waveforms for breath sounds is possible by changing the preprocessing according to the position as well as the frequency band of the exhalation–inhalation discrimination.
One limitation of this research is that the quality of circuit design and component mounting needs to be improved because this research aims to confirm the principle of a new respiration sensing method. Collaborating with hardware design experts for future practical use is necessary.

5. Conclusions

In this study, we propose breath measurement methods for synchronized reproduction of breath sounds in a stethoscope-based training system. Two methods are proposed here, namely those based on thoracic variations and breath sounds. For the breath measurement method using thoracic variations, a pressure sensor, an accelerometer, a gyro sensor, and a magnetic sensor were selected for the measurements, and a stethoscope-type measurement device was created using a 3D printer. For the breath measurement method using breath sounds, a measurement device using a stethoscope and a microphone was created, and a method for acquiring breathing waveforms was proposed to discriminate exhalation and inhalation sounds using spectral integration and conversion to waveforms.
The evaluation experiments showed correlations with the reference values of the breathing waveform, but strong correlations could not be obtained. For each method, the value of the correlation coefficient was higher when the measurement method with the highest correlation coefficient was selected, suggesting that the breath measurement performance can be improved by selecting an appropriate method.
The contribution of this paper is that we devised a new principle that enables an effective and real-time acquisition of breathing, under the constraint that it can be built into the chest piece, and evaluated its effectiveness. In the future, we intend to systematically study selection of the measurement method according to measurement conditions. We also aim to examine the optimal method of dealing with noise and signal preprocessing, which this study could not examine to obtain a stronger correlation with reference values. Finally, we would like to consider implementation of a breath sound synchronization reproduction method for auscultation training systems.

Author Contributions

Conceptualization, Y.K. and T.N.; Data curation, Y.K.; Formal analysis, Y.K.; Funding acquisition, T.N.; Investigation, Y.K.; Methodology, Y.K.; Project administration, T.N.; Resources, T.N.; Software, Y.K.; Supervision, K.M., H.K., S.I., M.A., M.T., Y.N. and T.N.; Visualization, Y.K.; Writing—original draft, Y.K.; Writing—review & editing, T.N. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Informed consent was obtained from all subjects involved in this study.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Ward, J.J.; Wattier, B.A. Technology for Enhancing Chest Auscultation in Clinical Simulation. Respir. Care 2011, 56, 834–845. [Google Scholar] [CrossRef]
  2. Sarkar, M.; Madabhavi, I.; Niranjan, N.; Dogra, M. Auscultation of the Respiratory System. Ann. Thorac. Med. 2015, 10, 158–168. [Google Scholar] [CrossRef]
  3. Vukanovic-Criley, J.M.; Criley, S.; Warde, C.M.; Boker, J.R.; Guevara-Matheus, L.; Churchill, W.H.; Nelson, W.P.; Criley, J.M. Competency in Cardiac Examination Skills in Medical Students, Trainees, Physicians, and Faculty: A Multicenter Study. Arch. Intern. Med. 2006, 166, 610–616. [Google Scholar] [CrossRef]
  4. Woywodt, A.; Herrmann, A.; Kielstein, J.T.; Haller, H.; Haubitz, M.; Purnhagen, H. A Novel Multimedia Tool to Improve Bedside Teaching of Cardiac Auscultation. Postgrad. Med. J. 2004, 80, 355–357. [Google Scholar] [CrossRef]
  5. Kumar, K.; Thompson, W.R. Evaluation of Cardiac Auscultation Skills in Pediatric Residents. Clin. Pediatr. 2013, 52, 66–73. [Google Scholar] [CrossRef]
  6. Binka, E.; Lewin, L.; Gaskin, P. Small Steps in Impacting Clinical Auscultation of Medical Students. Glob. Pediatr. Health 2016, 3, 2333794X16669013. [Google Scholar] [CrossRef] [PubMed]
  7. Harden, R.M.; Gleeson, F.A. Assessment of Clinical Competence Using an Objective Structured Clinical Examination (OSCE). Med. Educ. 1979, 13, 41–54. [Google Scholar] [CrossRef] [PubMed]
  8. Newble, D.I.; Swanson, D.B. Psychometric Characteristics of the Objective Structured Clinical Examination. Med. Educ. 1988, 22, 325–334. [Google Scholar] [CrossRef] [PubMed]
  9. Takashina, T.; Masuzawa, T.; Fukui, Y. A New Cardiac Auscultation Simulator. Clin. Cardiol. 1990, 13, 869–872. [Google Scholar] [CrossRef] [PubMed]
  10. Kagaya, Y.; Tabata, M.; Arata, Y.; Kameoka, J.; Ishii, S. Variation in Effectiveness of a Cardiac Auscultation Training Class with a Cardiology Patient Simulator among Heart Sounds and Murmurs. J. Cardiol. 2017, 70, 192–198. [Google Scholar] [CrossRef] [PubMed]
  11. Kern, D.H.; Mainous, A.G.; Carey, M.; Beddingfield, A. Simulation-Based Teaching to Improve Cardiovascular Exam Skills Performance Among Third-Year Medical Students. Teach. Learn. Med. 2011, 23, 15–20. [Google Scholar] [CrossRef]
  12. Issenberg, S.B.; McGaghie, W.C.; Petrusa, E.R.; Lee Gordon, D.; Scalese, R.J. Features and Uses of High-Fidelity Medical Simulations That Lead to Effective Learning: A BEME Systematic Review. Med. Teach. 2005, 27, 10–28. [Google Scholar] [CrossRef]
  13. Nakaguchi, T.; Ueda, T.; Muratake, T.; Kasahara, Y.; Iwata, Y.; Tanabe, M. Development of VR-Based Auscultation Training System Using Simulated Patient. J. Med. Imaging Health Inform. 2013, 3, 89–93. [Google Scholar] [CrossRef]
  14. Sekiguchi, Y.; Nakaguchi, T.; Muratake, T.; Miura, K.; Kawada, N.; Yoshimura, Y.; Ito, S.; Asahina, M.; Tanabe, M. Development of camera built-in stethoscope device for AR-based auscultation training simulator. Jpn. J. Med. Virtual Real. 2020, 16, 15–26. [Google Scholar] [CrossRef]
  15. Sekiguchi, Y.; Nakaguchi, T.; Kono, Y.; Miura, K.; Kasai, H.; Kawada, N.; Yoshimura, Y.; Ito, S.; Asahina, M.; Tanabe, M. Development of an Augmented Reality Auscultation Training System Using Deep Learning and Its Application to Medical Education Tests. Trans. Virtual Real. Soc. Jpn. 2021, 26, 160–168. [Google Scholar] [CrossRef]
  16. Preejith, S.P.; Jeelani, A.; Maniyar, P.; Joseph, J.; Sivaprakasam, M. Accelerometer based system for continuous respiratory rate monitoring. In Proceedings of the 2017 IEEE International Symposium on Medical Measurements and Applications (MeMeA), Rochester, MN, USA, 8–10 May 2017; pp. 171–176. [Google Scholar] [CrossRef]
  17. Park, S.; Shin, H. Feasibility Study for Unconstrained Respiration Monitoring Based on Multi-Way Approach Using an Acceleration and Force Sensing Module. IEEE Sens. J. 2017, 17, 3482–3489. [Google Scholar] [CrossRef]
  18. Massaroni, C.; Nicolò, A.; Lo Presti, D.; Sacchetti, M.; Silvestri, S.; Schena, E. Contact-Based Methods for Measuring Respiratory Rate. Sensors 2019, 19, 908. [Google Scholar] [CrossRef]
  19. Yu, M.-C.; Liou, J.-L.; Kuo, S.-W.; Lee, M.-S.; Hung, Y.-P. Noncontact Respiratory Measurement of Volume Change Using Depth Camera. In Proceedings of the Annual International Conference of the IEEE Engineering in Medicine and Biology Society, San Diego, CA, USA, 28 August–1 September 2012; pp. 2371–2374. [Google Scholar] [CrossRef]
  20. Min, S.D.; Kim, J.K.; Shin, H.S.; Yun, Y.H.; Lee, C.K.; Lee, M. Noncontact Respiration Rate Measurement System Using an Ultrasonic Proximity Sensor. IEEE Sens. J. 2010, 10, 1732–1739. [Google Scholar] [CrossRef]
  21. Abbas, A.K.; Heimann, K.; Jergus, K.; Orlikowsky, T.; Leonhardt, S. Neonatal Non-Contact Respiratory Monitoring Based on Real-Time Infrared Thermography. Biomed. Eng. Online 2011, 10, 93. [Google Scholar] [CrossRef]
  22. Pereira, C.B.; Yu, X.; Czaplik, M.; Rossaint, R.; Blazek, V.; Leonhardt, S. Remote Monitoring of Breathing Dynamics Using Infrared Thermography. Biomed. Opt. Express 2015, 6, 4378–4394. [Google Scholar] [CrossRef] [PubMed]
  23. Moll, J.M.; Wright, V. An objective clinical study of chest expansion. Ann. Rheum. Dis. 1972, 31, 1–8. [Google Scholar] [CrossRef] [PubMed]
  24. Reichert, S.; Gass, R.; Brandt, C.; Andrès, E. Analysis of Respiratory Sounds: State of the Art. Clin. Med. Circ. Respirat. Pulm. Med. 2008, 2, 45–58. [Google Scholar] [CrossRef] [PubMed]
  25. Arita, H. Dictionary of Breathing; Asakura Publishing Co., Ltd.: Tokyo, Japan, 2006; p. 728. [Google Scholar]
  26. Pasterkamp, H.; Kraman, S.S.; Wodicka, G.R. Respiratory sounds: Advances beyond the stethoscope. Am. J. Respir. Crit. Care Med. 1997, 156, 974–987. [Google Scholar] [CrossRef] [PubMed]
  27. Li, S.-H.; Lin, B.-S.; Tsai, C.-H.; Yang, C.-T.; Lin, B.-S. Design of Wearable Breathing Sound Monitoring System for Real-Time Wheeze Detection. Sensors 2017, 17, 171. [Google Scholar] [CrossRef] [PubMed]
  28. Charbonneau, G.; Ademovic, E.; Cheetham, B.M.G.; Malmberg, L.P.; Vanderschoot, J.; Sovijarvi, A.R.A. Basic techniques for respiratory sound analysis. Eur. Respir. Rev. 2000, 10, 625–635. [Google Scholar]
  29. Kaneko, H.; Horie, J. Breathing Movements of the Chest and Abdominal Wall in Healthy Subjects. Respir. Care. 2012, 57, 1442–1451. [Google Scholar] [CrossRef]
Figure 1. Conventional stethoscope used in EARS.
Figure 1. Conventional stethoscope used in EARS.
Sensors 24 01626 g001
Figure 2. Thoracic variability measurement device.
Figure 2. Thoracic variability measurement device.
Sensors 24 01626 g002
Figure 3. Principle of measurement by the magnetic sensor. When the surface of the stethoscope is indented due to thoracic variations, the magnetic field emitted by the magnet shifts and changes the value of the magnetic sensor.
Figure 3. Principle of measurement by the magnetic sensor. When the surface of the stethoscope is indented due to thoracic variations, the magnetic field emitted by the magnet shifts and changes the value of the magnetic sensor.
Sensors 24 01626 g003
Figure 4. Flowchart for breathing waveform acquisition using breath sounds.
Figure 4. Flowchart for breathing waveform acquisition using breath sounds.
Sensors 24 01626 g004
Figure 5. Spectral integral ratio range of breath sounds calculated using the proposed method.
Figure 5. Spectral integral ratio range of breath sounds calculated using the proposed method.
Sensors 24 01626 g005
Figure 6. Breath sounds sampling device.
Figure 6. Breath sounds sampling device.
Sensors 24 01626 g006
Figure 7. Areas in red were measured using the proposed method, and areas in blue were measured with the Go Direct Respiration Belt for the reference value.
Figure 7. Areas in red were measured using the proposed method, and areas in blue were measured with the Go Direct Respiration Belt for the reference value.
Sensors 24 01626 g007
Figure 8. Breathing waveform for each measurement method.
Figure 8. Breathing waveform for each measurement method.
Sensors 24 01626 g008
Figure 9. Comparison of breathing waveform acquisition accuracies by measurement method. The area shown in red color indicates where the correlations are considered to exist (>0.4).
Figure 9. Comparison of breathing waveform acquisition accuracies by measurement method. The area shown in red color indicates where the correlations are considered to exist (>0.4).
Sensors 24 01626 g009
Table 1. Pressure sensor specifications.
Table 1. Pressure sensor specifications.
Thickness0.20–1.25 mm
Pressure sensitive range0.2–10.0 N
Minimum sensitivity20–100 g
Table 2. Specifications of the 9-axis inertial sensor.
Table 2. Specifications of the 9-axis inertial sensor.
FXOS8700
3-axis accelerometer
Resolution14 bit
Detection range−2 to 2 g
FXOS8700
3-axis magnetic Sensor
Resolution16 bit
Detection range−1200 to 1200 µT
FXAS21002
3-axis gyro sensor
Resolution16 bit
Detection range−250 to 250 degree/s
Table 3. Microphone specifications.
Table 3. Microphone specifications.
DirectivityOmnidirectional
Frequency response50–15,000 Hz
Frontal sensitivity−38 ± 3.5 dB
Table 4. Go Direct Respiration Belt specifications.
Table 4. Go Direct Respiration Belt specifications.
Detection range0–50 N
Resolution0.01 N
Sampling rate20 Hz
Table 5. Analysis of variance table within three-factor participants (A: breathing method, B: measurement point, C: measurement method, SS: sum of squares, Df: degrees of freedom, MS: mean square), ns: not significant, * p < 0.05, ** p < 0.01.
Table 5. Analysis of variance table within three-factor participants (A: breathing method, B: measurement point, C: measurement method, SS: sum of squares, Df: degrees of freedom, MS: mean square), ns: not significant, * p < 0.05, ** p < 0.01.
Variable NameSSDfMSF Valuep-Value Partial   η 2
A0.9681.000.96827.7030.000**0.681
B0.8744.000.2196.2180.000**0.324
C0.6482.130.3047.0660.003**0.352
A × B0.0124.000.0030.0720.990ns0.006
A × C0.4852.530.1928.4710.000**0.395
B × C0.54516.000.0342.0120.014*0.134
A × B × C0.26216.000.0161.1460.315ns0.081
Table 6. Mean values and standard deviations of correlation coefficients for the breathing method and measurement method factors under each condition (A: breathing method, A1: normal breath, A2: deep breath, C: measurement method, C1: pressure sensor, C2: accelerometer, C3: magnetic sensor, C4: accelerometer + gyro sensor, C5: microphone).
Table 6. Mean values and standard deviations of correlation coefficients for the breathing method and measurement method factors under each condition (A: breathing method, A1: normal breath, A2: deep breath, C: measurement method, C1: pressure sensor, C2: accelerometer, C3: magnetic sensor, C4: accelerometer + gyro sensor, C5: microphone).
Condition ACondition CNumber of DataMean ValueSD
A1C1700.4850.164
A1C2700.3960.159
A1C3700.4370.141
A1C4700.3910.156
A1C5700.4430.142
A2C1700.5330.154
A2C2700.5010.179
A2C3700.5620.156
A2C4700.5020.179
A2C5700.4260.113
Table 7. Mean values and standard deviations of correlation coefficients for the position and method factors under each condition (B: measurement point, C: measurement method, C1: pressure sensor, C2: accelerometer, C3: magnetic sensor, C4: accelerometer + gyro sensor, C5: microphone).
Table 7. Mean values and standard deviations of correlation coefficients for the position and method factors under each condition (B: measurement point, C: measurement method, C1: pressure sensor, C2: accelerometer, C3: magnetic sensor, C4: accelerometer + gyro sensor, C5: microphone).
Condition BCondition CNumber of DataMean ValueSD
B1C1280.4500.128
B1C2280.4110.192
B1C3280.4050.154
B1C4280.4060.187
B1C5280.4530.141
B2C1280.5110.176
B2C2280.5240.193
B2C3280.5490.152
B2C4280.5150.195
B2C5280.4330.151
B3C1280.5020.177
B3C2280.3890.162
B3C3280.4680.138
B3C4280.3940.169
B3C5280.4270.126
B4C1280.5590.150
B4C2280.4810.170
B4C3280.5900.160
B4C4280.4920.170
B4C5280.4340.092
B5C1280.5230.157
B5C2280.4380.139
B5C3280.4850.142
B5C4280.4250.134
B5C5280.4250.132
Table 8. Test results of the simple main effects for the interactions between the breathing and measurement method factors as well as for interactions between the measurement position and measurement method factors. (A: breathing method, A1: normal breath, A2: deep breath, B: measurement point, C: measurement method, C1: pressure sensor, C2: accelerometer, C3: magnetic sensor, C4: accelerometer + gyro sensor, C5: microphone, SS: sum of squares, Df: degrees of freedom, MS: mean square) ns: not significant, * p < 0.05 ** p < 0.01.
Table 8. Test results of the simple main effects for the interactions between the breathing and measurement method factors as well as for interactions between the measurement position and measurement method factors. (A: breathing method, A1: normal breath, A2: deep breath, B: measurement point, C: measurement method, C1: pressure sensor, C2: accelerometer, C3: magnetic sensor, C4: accelerometer + gyro sensor, C5: microphone, SS: sum of squares, Df: degrees of freedom, MS: mean square) ns: not significant, * p < 0.05 ** p < 0.01.
SliceVariable NameSSDfMSF Valuep-Value Partial   η 2
C = C1A0.0811.0000.0814.0320.066ns0.237
C = C2A0.3881.0000.38823.3400.000**0.642
C = C3A0.5461.0000.54634.9270.000**0.729
C = C4A0.4271.0000.42723.8610.000**0.647
C = C5A0.0101.0000.0100.4630.508ns0.034
A = A1C0.5451.9700.2095.4200.011*0.294
A = A2C0.2622.6900.2689.8860.000**0.432
C = C1B0.1764.0000.0441.8330.136ns0.124
C = C2B0.3264.0000.0823.6800.010*0.221
C = C3B0.5824.0000.1457.2970.000**0.360
C = C4B0.3214.0000.0803.8690.008**0.230
C = C5B0.0144.0000.0040.2220.925ns0.017
B = B1C0.0661.9600.0340.7530.479ns0.055
B = B2C0.2132.6400.0812.2450.108ns0.147
B = B3C0.2612.1600.1213.8460.031*0.228
B = B4C0.4411.5400.2877.6000.006**0.369
B = B5C0.2111.9600.1073.9430.033*0.233
Table 9. Results of multiple comparisons (Bonferroni method) (A: breathing method, A1: normal breath, A2: deep breath, B: measurement position, C: measurement method, C1: pressure sensor, C2: accelerometer, C3: magnetic sensor, C4: accelerometer + gyro sensor, C5: microphone, Df: degrees of freedom).
Table 9. Results of multiple comparisons (Bonferroni method) (A: breathing method, A1: normal breath, A2: deep breath, B: measurement position, C: measurement method, C1: pressure sensor, C2: accelerometer, C3: magnetic sensor, C4: accelerometer + gyro sensor, C5: microphone, Df: degrees of freedom).
Level PairTermsDifft-ValueDfp-ValueADJUSTMENT p-ValueOrder
C1–C4A10.0933.55513.0000.0040.035C1 > C4
C1–C2A10.0893.43213.0000.0050.035C1 > C2
C3–C5A20.1365.54213.0000.0000.001C3 > C5
C1–C5A20.1074.51913.0000.0010.004C1 > C5
B1–B4C3−0.1855.08813.0000.0000.002B1 < B4
B1–B2C3−0.1444.05313.0000.0010.008B1 < B2
B3–B4C3−0.1223.14313.0000.0080.047B3 < B4
B2–B5C40.0893.54213.0000.0040.036B2 > B5
B1–B2C4−0.1083.31313.0000.0060.036B1 < B2
C3–C5B40.1566.89313.0000.0000.000C3 > C5
C1–C5B40.1255.14413.0000.0000.001C1 > C5
Table 10. Average correlation coefficients by method selection.
Table 10. Average correlation coefficients by method selection.
Before method selection0.46 ± 0.14
After method selection0.52 ± 0.14
Table 11. Breathing waveform calculations by changing the frequency band Mean value of correlation coefficient.
Table 11. Breathing waveform calculations by changing the frequency band Mean value of correlation coefficient.
Measurement point (1)300–600 Hz0.40 ± 0.09
500–800 Hz0.43 ± 0.10
Measurement point (2)300–600 Hz0.42 ± 0.09
500–800 Hz0.48 ± 0.14
Measurement point (3)300–600 Hz0.44 ± 0.15
500–800 Hz0.45 ± 0.16
Measurement point (4)300–600 Hz0.47 ± 0.17
500–800 Hz0.43 ± 0.12
Measurement point (5)300–600 Hz0.42 ± 0.13
500–800 Hz0.47 ± 0.13
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kono, Y.; Miura, K.; Kasai, H.; Ito, S.; Asahina, M.; Tanabe, M.; Nomura, Y.; Nakaguchi, T. Breath Measurement Method for Synchronized Reproduction of Biological Tones in an Augmented Reality Auscultation Training System. Sensors 2024, 24, 1626. https://doi.org/10.3390/s24051626

AMA Style

Kono Y, Miura K, Kasai H, Ito S, Asahina M, Tanabe M, Nomura Y, Nakaguchi T. Breath Measurement Method for Synchronized Reproduction of Biological Tones in an Augmented Reality Auscultation Training System. Sensors. 2024; 24(5):1626. https://doi.org/10.3390/s24051626

Chicago/Turabian Style

Kono, Yukiko, Keiichiro Miura, Hajime Kasai, Shoichi Ito, Mayumi Asahina, Masahiro Tanabe, Yukihiro Nomura, and Toshiya Nakaguchi. 2024. "Breath Measurement Method for Synchronized Reproduction of Biological Tones in an Augmented Reality Auscultation Training System" Sensors 24, no. 5: 1626. https://doi.org/10.3390/s24051626

APA Style

Kono, Y., Miura, K., Kasai, H., Ito, S., Asahina, M., Tanabe, M., Nomura, Y., & Nakaguchi, T. (2024). Breath Measurement Method for Synchronized Reproduction of Biological Tones in an Augmented Reality Auscultation Training System. Sensors, 24(5), 1626. https://doi.org/10.3390/s24051626

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop