Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (55)

Search Parameters:
Keywords = heart-sound classification

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
33 pages, 13568 KB  
Article
Are Artificial Intelligence Models Listening Like Cardiologists? Bridging the Gap Between Artificial Intelligence and Clinical Reasoning in Heart-Sound Classification Using Explainable Artificial Intelligence
by Sami Alrabie and Ahmed Barnawi
Bioengineering 2025, 12(6), 558; https://doi.org/10.3390/bioengineering12060558 - 22 May 2025
Viewed by 872
Abstract
In recent years, deep learning has shown promise in automating heart-sound classification. Although this approach is fast, non-invasive, and cost-effective, its diagnostic accuracy still mainly depends on the clinician’s expertise, making it particularly challenging to detect rare or complex conditions. This study is [...] Read more.
In recent years, deep learning has shown promise in automating heart-sound classification. Although this approach is fast, non-invasive, and cost-effective, its diagnostic accuracy still mainly depends on the clinician’s expertise, making it particularly challenging to detect rare or complex conditions. This study is motivated by two key concerns in the field of heart-sound classification. First, we observed that automatic heart-sound segmentation algorithms—commonly used for data augmentation—produce varying outcomes, raising concerns about the accuracy of both the segmentation process and the resulting classification performance. Second, we noticed inconsistent accuracy scores across different pretrained models, prompting the need for interpretable explanations to validate these results. We argue that without interpretability to support reported metrics, accuracy scores can be misleading because of ambiguity in how training data interact with pretrained models. Specifically, it remains unclear whether these models classify spectrogram images—generated from heart-sound signals—in a way that aligns with clinical reasoning, where experts focus on specific components of the heart cycle, such as S1, systole, S2, and diastole. To address this, we applied explainable AI (XAI) techniques with two primary objectives: (1) to assess whether the model truly focuses on clinically relevant features, thereby allowing classification results to be verified and trusted, and (2) to investigate whether incorporating attention mechanisms can improve both the performance and the model’s focus on meaningful segments of the signal. To the best of our knowledge, this is the first study conducted on a manually segmented dataset, which objectively evaluates the model’s behavior using XAI and explores performance enhancement by combining attention mechanisms with pretrained models. We employ the Grad-CAM method to visualize the model’s attention and gain insights into the decision-making process. The experimental results show that integrating multi-head attention significantly improves both the classification accuracy and interpretability. Notably, ResNet50 with multi-head attention achieved an accuracy of 97.3%, outperforming those of both the baseline and SE-enhanced models. Moreover, the mean intersection over union (mIoU) for interpretability increased from 75.7% to 82.0%, indicating the model’s improved focus on diagnostically relevant regions. Full article
Show Figures

Figure 1

21 pages, 1228 KB  
Article
Automatic Feature Selection for Imbalanced Echocardiogram Data Using Event-Based Self-Similarity
by Huang-Nan Huang, Hong-Min Chen, Wei-Wen Lin, Rita Wiryasaputra, Yung-Cheng Chen, Yu-Huei Wang and Chao-Tung Yang
Diagnostics 2025, 15(8), 976; https://doi.org/10.3390/diagnostics15080976 - 11 Apr 2025
Viewed by 731
Abstract
Background and Objective: Using echocardiogram data for cardiovascular disease (CVD) can lead to difficulties due to imbalanced datasets, leading to biased predictions. Machine learning models can enhance prognosis accuracy, but their effectiveness is influenced by optimal feature selection and robust classification techniques. This [...] Read more.
Background and Objective: Using echocardiogram data for cardiovascular disease (CVD) can lead to difficulties due to imbalanced datasets, leading to biased predictions. Machine learning models can enhance prognosis accuracy, but their effectiveness is influenced by optimal feature selection and robust classification techniques. This study introduces an event-based self-similarity approach to enhance automatic feature selection approach for imbalanced echocardiogram data. Critical features correlated with disease progression were identified by leveraging self-similarity patterns. This study used an echocardiogram dataset, visual presentations of high-frequency sound wave signals, and data of patients with heart disease who are treated using three treatment methods: catheter ablation, ventricular defibrillator, and drug control—over the course of three years. Methods: The dataset was classified into nine categories and Recursive Feature Elimination (RFE) was applied to identify the most relevant features, reducing model complexity while maintaining diagnostic accuracy. Machine learning classification models, including XGBoost and CATBoost, were trained and evaluated. Results: Both models achieved comparable accuracy values, 84.3% and 88.4%, respectively, under different normalization techniques. To further optimize performance, the models were combined into a voting ensemble, improving feature selection and predictive accuracy. Four essential features—age, aorta (AO), left ventricular (LV), and left atrium (LA)—were identified as critical for prognosis and were found in Random Forest (RF)-voting ensemble classifier. The results underscore the importance of feature selection techniques in handling imbalanced datasets, improving classification robustness, and reducing bias in automated prognosis systems. Conclusions: Our findings highlight the potential of machine learning-driven echocardiogram analysis to enhance patient care by providing accurate, data-driven assessments. Full article
(This article belongs to the Section Medical Imaging and Theranostics)
Show Figures

Figure 1

22 pages, 7716 KB  
Article
A Deep-Learning Approach to Heart Sound Classification Based on Combined Time-Frequency Representations
by Leonel Orozco-Reyes, Miguel A. Alonso-Arévalo, Eloísa García-Canseco, Roilhi F. Ibarra-Hernández and Roberto Conte-Galván
Technologies 2025, 13(4), 147; https://doi.org/10.3390/technologies13040147 - 7 Apr 2025
Cited by 2 | Viewed by 2193
Abstract
Worldwide, heart disease is the leading cause of mortality. Cardiac auscultation, when conducted by a trained professional, is a non-invasive, cost-effective, and readily available method for the initial assessment of cardiac health. Automated heart sound analysis offers a promising and accessible approach to [...] Read more.
Worldwide, heart disease is the leading cause of mortality. Cardiac auscultation, when conducted by a trained professional, is a non-invasive, cost-effective, and readily available method for the initial assessment of cardiac health. Automated heart sound analysis offers a promising and accessible approach to supporting cardiac diagnosis. This work introduces a novel method for classifying heart sounds as normal or abnormal by leveraging time-frequency representations. Our approach combines three distinct time-frequency representations—short-time Fourier transform (STFT), mel-scale spectrogram, and wavelet synchrosqueezed transform (WSST)—to create images that enhance classification performance. These images are used to train five convolutional neural networks (CNNs): AlexNet, VGG-16, ResNet50, a CNN specialized in STFT images, and our proposed CNN model. The method was trained and tested using three public heart sound datasets: PhysioNet/CinC Challenge 2016, CirCor DigiScope Phonocardiogram Dataset 2022, and another open database. While individual representations achieve maximum accuracy of ≈85.9%, combining STFT, mel, and WSST boosts accuracy to ≈99%. By integrating complementary time-frequency features, our approach demonstrates robust heart sound analysis, achieving consistent classification performance across diverse CNN architectures, thus ensuring reliability and generalizability. Full article
Show Figures

Figure 1

18 pages, 4101 KB  
Article
Heart Sound Classification Based on Multi-Scale Feature Fusion and Channel Attention Module
by Mingzhe Li, Zhaoming He and Hao Wang
Bioengineering 2025, 12(3), 290; https://doi.org/10.3390/bioengineering12030290 - 14 Mar 2025
Cited by 1 | Viewed by 1281
Abstract
Intelligent heart sound diagnosis based on Convolutional Neural Networks (CNN) has been attracting increasing attention due to its accuracy and efficiency, which have been improved by recent studies. However, the performance of CNN models, heavily influenced by their parameters and structures, still has [...] Read more.
Intelligent heart sound diagnosis based on Convolutional Neural Networks (CNN) has been attracting increasing attention due to its accuracy and efficiency, which have been improved by recent studies. However, the performance of CNN models, heavily influenced by their parameters and structures, still has room for improvement. In this paper, we propose a heart sound classification model named CAFusionNet, which fuses features from different layers with varying resolution ratios and receptive field sizes. Key features related to heart valve diseases are weighted by a channel attention block at each layer. To address the issue of limited dataset size, we apply a homogeneous transfer learning approach. CAFusionNet outperforms existing models on a dataset comprising public data combined with our proprietary dataset, achieving an accuracy of 0.9323. Compared to traditional deep learning methods, the transfer learning algorithm achieves an accuracy of 0.9665 in the triple classification task. Output data and visualized heat maps highlight the significance of feature fusion from different layers. The proposed methods significantly enhanced the performance of heart sound classification and demonstrated the importance of feature fusion, as interpreted through visualized heat maps. Full article
(This article belongs to the Section Biosignal Processing)
Show Figures

Figure 1

19 pages, 3377 KB  
Article
AI-Enhanced Detection of Heart Murmurs: Advancing Non-Invasive Cardiovascular Diagnostics
by Maria-Alexandra Zolya, Elena-Laura Popa, Cosmin Baltag, Dragoș-Vasile Bratu, Simona Coman and Sorin-Aurel Moraru
Sensors 2025, 25(6), 1682; https://doi.org/10.3390/s25061682 - 8 Mar 2025
Viewed by 1628
Abstract
Cardiovascular diseases (CVDs) are the leading cause of death worldwide, claiming over 17 million lives annually. Early detection of conditions like heart murmurs, often indicative of heart valve abnormalities, is critical for improving patient outcomes. Traditional diagnostic methods, including physical auscultation and advanced [...] Read more.
Cardiovascular diseases (CVDs) are the leading cause of death worldwide, claiming over 17 million lives annually. Early detection of conditions like heart murmurs, often indicative of heart valve abnormalities, is critical for improving patient outcomes. Traditional diagnostic methods, including physical auscultation and advanced imaging techniques, are constrained by their reliance on specialized clinical expertise, inherent procedural invasiveness, substantial financial costs, and limited accessibility, particularly in resource-limited healthcare environments. This study presents a novel convolutional recurrent neural network (CRNN) model designed for the non-invasive classification of heart murmurs. The model processes heart sound recordings using advanced pre-processing techniques such as z-score normalization, band-pass filtering, and data augmentation (Gaussian noise, time shift, and pitch shift) to enhance robustness. By combining convolutional and recurrent layers, the CRNN captures spatial and temporal features in audio data, achieving an accuracy of 90.5%, precision of 89%, and recall of 87%. These results underscore the potential of machine-learning technologies to revolutionize cardiac diagnostics by offering scalable, accessible solutions for the early detection of cardiovascular conditions. This approach paves the way for broader applications of AI in healthcare, particularly in underserved regions where traditional resources are scarce. Full article
Show Figures

Figure 1

37 pages, 4062 KB  
Article
Heart Sound Classification Using Harmonic and Percussive Spectral Features from Phonocardiograms with a Deep ANN Approach
by Anupinder Singh, Vinay Arora and Mandeep Singh
Appl. Sci. 2024, 14(22), 10201; https://doi.org/10.3390/app142210201 - 6 Nov 2024
Cited by 3 | Viewed by 2050
Abstract
Cardiovascular diseases (CVDs) are a leading cause of mortality worldwide, with a particularly high burden in India. Non-invasive methods like Phonocardiogram (PCG) analysis capture the acoustic activity of the heart. This holds significant potential for the early detection and diagnosis of heart conditions. [...] Read more.
Cardiovascular diseases (CVDs) are a leading cause of mortality worldwide, with a particularly high burden in India. Non-invasive methods like Phonocardiogram (PCG) analysis capture the acoustic activity of the heart. This holds significant potential for the early detection and diagnosis of heart conditions. However, the complexity and variability of PCG signals pose considerable challenges for accurate classification. Traditional methods of PCG signal analysis, including time-domain, frequency-domain, and time-frequency domain techniques, often fall short in capturing the intricate details necessary for reliable diagnosis. This study introduces an innovative approach that leverages harmonic–percussive source separation (HPSS) to extract distinct harmonic and percussive spectral features from PCG signals. These features are then utilized to train a deep feed-forward artificial neural network (ANN), classifying heart conditions as normal or abnormal. The methodology involves advanced digital signal processing techniques applied to PCG recordings from the PhysioNet 2016 dataset. The feature set comprises 164 attributes, including the Chroma STFT, Chroma CENS, Mel-frequency cepstral coefficients (MFCCs), and statistical features. These are refined using the ROC-AUC feature selection method to ensure optimal performance. The deep feed-forward ANN model was rigorously trained and validated on a balanced dataset. Techniques such as noise reduction and outlier detection were used to improve model training. The proposed model achieved a validation accuracy of 93.40% with sensitivity and specificity rates of 82.40% and 80.60%, respectively. These results underscore the effectiveness of harmonic-based features and the robustness of the ANN in heart sound classification. This research highlights the potential for deploying such models in non-invasive cardiac diagnostics, particularly in resource-constrained settings. It also lays the groundwork for future advancements in cardiac signal analysis. Full article
(This article belongs to the Special Issue Machine Learning in Biomedical Applications)
Show Figures

Figure 1

20 pages, 1645 KB  
Article
Classification of Acoustic Tones and Cardiac Murmurs Based on Digital Signal Analysis Leveraging Machine Learning Methods
by Nataliya Shakhovska and Ivan Zagorodniy
Computation 2024, 12(10), 208; https://doi.org/10.3390/computation12100208 - 17 Oct 2024
Cited by 3 | Viewed by 2369
Abstract
Heart murmurs are abnormal heart sounds that can indicate various heart diseases. Although traditional auscultation methods are effective, they depend more on specialists’ knowledge, making it difficult to make an accurate diagnosis. This paper presents a machine learning-based framework for the classification of [...] Read more.
Heart murmurs are abnormal heart sounds that can indicate various heart diseases. Although traditional auscultation methods are effective, they depend more on specialists’ knowledge, making it difficult to make an accurate diagnosis. This paper presents a machine learning-based framework for the classification of acoustic sounds and heart murmurs using digital signal analysis. Using advanced machine learning algorithms, we aim to improve the accuracy, speed, and accessibility of heart murmur detection. The proposed method includes feature extraction from digital auscultatory recordings, preprocessing using signal processing techniques, and classification using state-of-the-art machine learning models. We evaluated the performance of different machine learning algorithms, such as convolutional neural networks (CNNs), random forests (RFs) and support vector machines (SVMs), on a selected heart noise dataset. The results show that our framework achieves high accuracy in differentiating normal heart sounds from different types of heart murmurs and provides a robust tool for clinical decision-making. Full article
(This article belongs to the Special Issue Artificial Intelligence Applications in Public Health)
Show Figures

Figure 1

15 pages, 3187 KB  
Article
Segmentation of Heart Sound Signal Based on Multi-Scale Feature Fusion and Multi-Classification of Congenital Heart Disease
by Yuan Zeng, Mingzhe Li, Zhaoming He and Ling Zhou
Bioengineering 2024, 11(9), 876; https://doi.org/10.3390/bioengineering11090876 - 29 Aug 2024
Cited by 1 | Viewed by 2024
Abstract
Analyzing heart sound signals presents a novel approach for early diagnosis of pediatric congenital heart disease. The existing segmentation algorithms have limitations in accurately distinguishing the first (S1) and second (S2) heart sounds, limiting the diagnostic utility of cardiac cycle data for pediatric [...] Read more.
Analyzing heart sound signals presents a novel approach for early diagnosis of pediatric congenital heart disease. The existing segmentation algorithms have limitations in accurately distinguishing the first (S1) and second (S2) heart sounds, limiting the diagnostic utility of cardiac cycle data for pediatric pathology assessment. This study proposes a time bidirectional long short-term memory network (TBLSTM) based on multi-scale analysis to segment pediatric heart sound signals according to different cardiac cycles. Mel frequency cepstral coefficients and dynamic characteristics of the heart sound fragments were extracted and input into random forest for multi-classification of congenital heart disease. The segmentation model achieved an overall F1 score of 94.15% on the verification set, with specific F1 scores of 90.25% for S1 and 86.04% for S2. In a situation where the number of cardiac cycles in the heart sound fragments was set to six, the results for multi-classification achieved stabilization. The performance metrics for this configuration were as follows: accuracy of 94.43%, sensitivity of 95.58%, and an F1 score of 94.51%. Furthermore, the segmentation model demonstrates robustness in accurately segmenting pediatric heart sound signals across different heart rates and in the presence of noise. Notably, the number of cardiac cycles in heart sound fragments directly impacts the multi-classification of these heart sound signals. Full article
(This article belongs to the Special Issue Computational Models in Cardiovascular System)
Show Figures

Figure 1

27 pages, 626 KB  
Review
Review of Phonocardiogram Signal Analysis: Insights from the PhysioNet/CinC Challenge 2016 Database
by Bing Zhu, Zihong Zhou, Shaode Yu, Xiaokun Liang, Yaoqin Xie and Qiurui Sun
Electronics 2024, 13(16), 3222; https://doi.org/10.3390/electronics13163222 - 14 Aug 2024
Cited by 9 | Viewed by 5202
Abstract
The phonocardiogram (PCG) is a crucial tool for the early detection, continuous monitoring, accurate diagnosis, and efficient management of cardiovascular diseases. It has the potential to revolutionize cardiovascular care and improve patient outcomes. The PhysioNet/CinC Challenge 2016 database, a large and influential resource, [...] Read more.
The phonocardiogram (PCG) is a crucial tool for the early detection, continuous monitoring, accurate diagnosis, and efficient management of cardiovascular diseases. It has the potential to revolutionize cardiovascular care and improve patient outcomes. The PhysioNet/CinC Challenge 2016 database, a large and influential resource, encourages contributions to accurate heart sound state classification (normal versus abnormal), achieving promising benchmark performance (accuracy: 99.80%; sensitivity: 99.70%; specificity: 99.10%; and score: 99.40%). This study reviews recent advances in analytical techniques applied to this database, and 104 publications on PCG signal analysis are retrieved. These techniques encompass heart sound preprocessing, signal segmentation, feature extraction, and heart sound state classification. Specifically, this study summarizes methods such as signal filtering and denoising; heart sound segmentation using hidden Markov models and machine learning; feature extraction in the time, frequency, and time-frequency domains; and state-of-the-art heart sound state recognition techniques. Additionally, it discusses electrocardiogram (ECG) feature extraction and joint PCG and ECG heart sound state recognition. Despite significant technical progress, challenges remain in large-scale high-quality data collection, model interpretability, and generalizability. Future directions include multi-modal signal fusion, standardization and validation, automated interpretation for decision support, real-time monitoring, and longitudinal data analysis. Continued exploration and innovation in heart sound signal analysis are essential for advancing cardiac care, improving patient outcomes, and enhancing user trust and acceptance. Full article
(This article belongs to the Special Issue Signal, Image and Video Processing: Development and Applications)
Show Figures

Figure 1

25 pages, 3406 KB  
Article
Machine Learning Algorithms for Processing and Classifying Unsegmented Phonocardiographic Signals: An Efficient Edge Computing Solution Suitable for Wearable Devices
by Roberto De Fazio, Lorenzo Spongano, Massimo De Vittorio, Luigi Patrono and Paolo Visconti
Sensors 2024, 24(12), 3853; https://doi.org/10.3390/s24123853 - 14 Jun 2024
Cited by 2 | Viewed by 1671
Abstract
The phonocardiogram (PCG) can be used as an affordable way to monitor heart conditions. This study proposes the training and testing of several classifiers based on SVMs (support vector machines), k-NN (k-Nearest Neighbor), and NNs (neural networks) to perform binary (“Normal”/”Pathologic”) and multiclass [...] Read more.
The phonocardiogram (PCG) can be used as an affordable way to monitor heart conditions. This study proposes the training and testing of several classifiers based on SVMs (support vector machines), k-NN (k-Nearest Neighbor), and NNs (neural networks) to perform binary (“Normal”/”Pathologic”) and multiclass (“Normal”, “CAD” (coronary artery disease), “MVP” (mitral valve prolapse), and “Benign” (benign murmurs)) classification of PCG signals, without heart sound segmentation algorithms. Two datasets of 482 and 826 PCG signals from the Physionet/CinC 2016 dataset are used to train the binary and multiclass classifiers, respectively. Each PCG signal is pre-processed, with spike removal, denoising, filtering, and normalization; afterward, it is divided into 5 s frames with a 1 s shift. Subsequently, a feature set is extracted from each frame to train and test the binary and multiclass classifiers. Concerning the binary classification, the trained classifiers yielded accuracies ranging from 92.4 to 98.7% on the test set, with memory occupations from 92.7 kB to 11.1 MB. Regarding the multiclass classification, the trained classifiers achieved accuracies spanning from 95.3 to 98.6% on the test set, occupying a memory portion from 233 kB to 14.1 MB. The NNs trained and tested in this work offer the best trade-off between performance and memory occupation, whereas the trained k-NN models obtained the best performance at the cost of large memory occupation (up to 14.1 MB). The classifiers’ performance slightly depends on the signal quality, since a denoising step is performed during pre-processing. To this end, the signal-to-noise ratio (SNR) was acquired before and after the denoising, indicating an improvement between 15 and 30 dB. The trained and tested models occupy relatively little memory, enabling their implementation in resource-limited systems. Full article
(This article belongs to the Special Issue AI-Based Automated Recognition and Detection in Healthcare)
Show Figures

Figure 1

25 pages, 4050 KB  
Article
Heart Sound Signals Classification with Image Conversion Employed
by Erqiang Deng, Yibei Jia, Guobin Zhu and Erqiang Zhou
Electronics 2024, 13(7), 1179; https://doi.org/10.3390/electronics13071179 - 22 Mar 2024
Cited by 5 | Viewed by 1798
Abstract
The number of patients with cardiovascular diseases worldwide is increasing rapidly, while medical resources are increasingly scarce. Heart sound classification, as the most direct means of discovering cardiovascular diseases, is attracting the attention of researchers around the world. Although great progress has been [...] Read more.
The number of patients with cardiovascular diseases worldwide is increasing rapidly, while medical resources are increasingly scarce. Heart sound classification, as the most direct means of discovering cardiovascular diseases, is attracting the attention of researchers around the world. Although great progress has been made in heart sound classification in recent years, most of them are based on traditional statistical feature methods and temporal dimension features. These traditional temporal dimension feature representation and classification methods cannot achieve good classification accuracy. This paper proposes a new partition attention module and Fusionghost module, and the entire network framework is named PANet. Without segmentation of the heart sound signal, the heart sound signal is converted into a bispectrum and input into the proposed framework for feature extraction and classification tasks. The network makes full use of multi-scale feature extraction and feature map fusion, improving the network feature extraction ability. This paper conducts a comprehensive study of the performance of different network parameters and different module numbers, and compares the performance with the most advanced algorithms currently available. Experiments have shown that for two classification problems (normal or abnormal), the classification accuracy rate on the 2016 PhysioNet/CinC Challenge database reached 97.89%, the sensitivity was 96.96%, and the specificity was 98.85%. Full article
(This article belongs to the Section Computer Science & Engineering)
Show Figures

Figure 1

1 pages, 152 KB  
Retraction
RETRACTED: Aljohani et al. A Novel Deep Learning CNN for Heart Valve Disease Classification Using Valve Sound Detection. Electronics 2023, 12, 846
by Randa I. Aljohani, Hanan A. Hosni Mahmoud, Alaaeldin Hafez and Magdy Bayoumi
Electronics 2024, 13(5), 856; https://doi.org/10.3390/electronics13050856 - 23 Feb 2024
Viewed by 1250
Abstract
The Electronics Editorial Office retracts the article, “A Novel Deep Learning CNN for Heart Valve Disease Classification Using Valve Sound Detection” [...] Full article
22 pages, 4739 KB  
Article
Abnormal Heart Sound Classification and Model Interpretability: A Transfer Learning Approach with Deep Learning
by Milan Marocchi, Leigh Abbott, Yue Rong, Sven Nordholm and Girish Dwivedi
J. Vasc. Dis. 2023, 2(4), 438-459; https://doi.org/10.3390/jvd2040034 - 4 Dec 2023
Cited by 6 | Viewed by 2983
Abstract
Physician detection of heart sound abnormality is complicated by the inherent difficulty of detecting critical abnormalities in the presence of noise. Computer-aided heart auscultation provides a promising alternative for more accurate detection, with recent deep learning approaches exceeding expert accuracy. Although combining phonocardiogram [...] Read more.
Physician detection of heart sound abnormality is complicated by the inherent difficulty of detecting critical abnormalities in the presence of noise. Computer-aided heart auscultation provides a promising alternative for more accurate detection, with recent deep learning approaches exceeding expert accuracy. Although combining phonocardiogram (PCG) data with electrocardiogram (ECG) data provides more information to an abnormal heart sound classifier, the scarce presence of labelled datasets with this combination impedes training. This paper explores fine-tuning deep convolutional neural networks such as ResNet, VGG, and inceptionv3, on images of spectrograms, mel-spectrograms, and scalograms. By fine-tuning deep pre-trained models on image representations of ECG and PCG, we achieve 91.25% accuracy on the training-a dataset of the PhysioNet Computing in Cardiology Challenge 2016, compared to a previous result of 81.48%. Interpretation of the model’s learned features is also provided, with the results indicative of clinical significance. Full article
(This article belongs to the Section Cardiovascular Diseases)
Show Figures

Figure 1

25 pages, 8241 KB  
Article
Heart Sound Classification Using Wavelet Analysis Approaches and Ensemble of Deep Learning Models
by Jin-A Lee and Keun-Chang Kwak
Appl. Sci. 2023, 13(21), 11942; https://doi.org/10.3390/app132111942 - 31 Oct 2023
Cited by 12 | Viewed by 5003
Abstract
Analyzing the condition and function of the heart is very important because cardiovascular diseases (CVDs) are responsible for high mortality rates worldwide and can lead to strokes and heart attacks; thus, early diagnosis and treatment are important. Phonocardiogram (PCG) signals can be used [...] Read more.
Analyzing the condition and function of the heart is very important because cardiovascular diseases (CVDs) are responsible for high mortality rates worldwide and can lead to strokes and heart attacks; thus, early diagnosis and treatment are important. Phonocardiogram (PCG) signals can be used to analyze heart rate characteristics to detect heart health and detect heart-related diseases. In this paper, we propose a method for designing using wavelet analysis techniques and an ensemble of deep learning models from phonocardiogram (PCG) for heart sound classification. For this purpose, we use wavelet scattering transform (WST) and continuous wavelet transform (CWT) as the wavelet analysis approaches for 1D-convolutional neural network (CNN) and 2D-CNN modeling, respectively. These features are insensitive to translations of the input on an invariance scale and are continuous with respect to deformations. Furthermore, the ensemble model is combined with 1D-CNN and 2D-CNN. The proposed method consists of four stages: a preprocessing stage for dividing signals at regular intervals, a feature extraction stage through wavelet scattering transform (WST) and continuous wavelet transform (CWT), a design stage of individual 1D-CNN and 2D-CNN, and a classification stage of heart sound by the ensemble model. The datasets used for the experiment were the PhysioNet/CinC 2016 challenge dataset and the PASCAL classifying heart sounds challenge dataset. The performance evaluation is performed by precision, recall, F1-score, sensitivity, and specificity. The experimental results revealed that the proposed method showed good performance on two datasets in comparison to the previous methods. The ensemble method of the proposed deep learning model surpasses the performance of recent studies and is suitable for predicting and diagnosing heart-related diseases by classifying heart sounds through phonocardiogram (PCG) signals. Full article
(This article belongs to the Special Issue AI-Based Biomedical Signal Processing)
Show Figures

Figure 1

15 pages, 3612 KB  
Article
Heart Murmur Classification Using a Capsule Neural Network
by Yu-Ting Tsai, Yu-Hsuan Liu, Zi-Wei Zheng, Chih-Cheng Chen and Ming-Chih Lin
Bioengineering 2023, 10(11), 1237; https://doi.org/10.3390/bioengineering10111237 - 24 Oct 2023
Cited by 4 | Viewed by 3443
Abstract
The healthcare industry has made significant progress in the diagnosis of heart conditions due to the use of intelligent detection systems such as electrocardiograms, cardiac ultrasounds, and abnormal sound diagnostics that use artificial intelligence (AI) technology, such as convolutional neural networks (CNNs). Over [...] Read more.
The healthcare industry has made significant progress in the diagnosis of heart conditions due to the use of intelligent detection systems such as electrocardiograms, cardiac ultrasounds, and abnormal sound diagnostics that use artificial intelligence (AI) technology, such as convolutional neural networks (CNNs). Over the past few decades, methods for automated segmentation and classification of heart sounds have been widely studied. In many cases, both experimental and clinical data require electrocardiography (ECG)-labeled phonocardiograms (PCGs) or several feature extraction techniques from the mel-scale frequency cepstral coefficient (MFCC) spectrum of heart sounds to achieve better identification results with AI methods. Without good feature extraction techniques, the CNN may face challenges in classifying the MFCC spectrum of heart sounds. To overcome these limitations, we propose a capsule neural network (CapsNet), which can utilize iterative dynamic routing methods to obtain good combinations for layers in the translational equivariance of MFCC spectrum features, thereby improving the prediction accuracy of heart murmur classification. The 2016 PhysioNet heart sound database was used for training and validating the prediction performance of CapsNet and other CNNs. Then, we collected our own dataset of clinical auscultation scenarios for fine-tuning hyperparameters and testing results. CapsNet demonstrated its feasibility by achieving validation accuracies of 90.29% and 91.67% on the test dataset. Full article
Show Figures

Figure 1

Back to TopTop