Next Article in Journal
Definition of an Inflammatory Biomarker Signature in Plasma-Derived Extracellular Vesicles of Glioblastoma Patients
Previous Article in Journal
The Current Role of Parathyroid Fine-Needle Biopsy (P-FNAB) with iPTH-Washout Concentration (iPTH-WC) in Primary Hyperparathyroidism: A Single Center Experience and Literature Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Local Structure Awareness-Based Retinal Microaneurysm Detection with Multi-Feature Combination

1
School of Information and Communication Engineering, University of Electronic Science and Technology of China, Chengdu 611731, China
2
School of Optoelectronic Science and Engineering, University of Electronic Science and Technology of China, Chengdu 611731, China
3
Laboratory of Imaging Detection and Intelligent Perception, University of Electronic Science and Technology of China, Chengdu 611731, China
4
Department of Ophthalmology, Sichuan Academy of Medical Sciences and Sichuan Provincial People’s Hospital, Chengdu 610072, China
*
Authors to whom correspondence should be addressed.
Biomedicines 2022, 10(1), 124; https://doi.org/10.3390/biomedicines10010124
Submission received: 10 December 2021 / Revised: 31 December 2021 / Accepted: 3 January 2022 / Published: 7 January 2022
(This article belongs to the Topic Medical Image Analysis)

Abstract

:
Retinal microaneurysm (MA) is the initial symptom of diabetic retinopathy (DR). The automatic detection of MA is helpful to assist doctors in diagnosis and treatment. Previous algorithms focused on the features of the target itself; however, the local structural features of the target and background are also worth exploring. To achieve MA detection, an efficient local structure awareness-based retinal MA detection with the multi-feature combination (LSAMFC) is proposed in this paper. We propose a novel local structure feature called a ring gradient descriptor (RGD) to describe the structural differences between an object and its surrounding area. Then, a combination of RGD with the salience and texture features is used by a Gradient Boosting Decision Tree (GBDT) for candidate classification. We evaluate our algorithm on two public datasets, i.e., the e-ophtha MA dataset and retinopathy online challenge (ROC) dataset. The experimental results show that the performance of the trained model significantly improved after combining traditional features with RGD, and the area under the receiver operating characteristic curve (AUC) values in the test results of the datasets e-ophtha MA and ROC increased from 0.9615 to 0.9751 and from 0.9066 to 0.9409, respectively.

1. Introduction

The number of diabetes patients worldwide is gradually increasing and, with the progression of diabetes, patients may develop DR, which may eventually cause vision loss or even blindness [1]. MA is the initial symptom of DR, and the early identification and timely treatment of retinal MA can prevent further progression of DR. Therefore, it is of great medical significance to realize the automatic detection of MA and assist doctors in the diagnosis of retinal lesions through computer technology.
Color fundus images are the primary way ophthalmologists assess retinal lesions, they judge whether the retina is normal and the grade of DR by visually observing whether there are microaneurysms, hard exudations, soft exudations, hemorrhages, and neovessels in the color fundus images [2]. MA occupies only a few pixels in fundus images and has low local contrast, as shown in Figure 1. At the same time, due to factors, such as the environment and equipment, color fundus images often have different brightness, contrast, and color. Artificial detection of MA is time-consuming, with low accuracy, and easily leads to ophthalmologist fatigue. Therefore, many researchers have studied the automatic detection of MA.
The current mainstream MA detection algorithm [3,4,5,6,7] can be briefly summarized as the following three steps: preprocessing, candidate extraction, and candidate classification. The preprocessing methods mainly include color correction, contrast enhancement, reflective elimination, and other image enhancement operations. The main purpose of pretreatment is to better observe the lesions in the retina and prepare for subsequent algorithms. Candidate extraction is primarily to extract image blocks that may contain MA from color fundus images, mainly using morphology, filtering, and other methods. In the step of candidate classification, the accurate detection of MA can be realized by extracting hand-crafted features of each candidate and classifying them with a machine learning classifier.
In this work, we first performed color correction on fundus images, and then used morphological methods to extract MA candidates. We found that MAs and blood vessels were mainly present in the candidate region, and traditional features based on the candidate area are not sufficiently capable and interpretable of distinguishing between them. To make up for the lack of traditional features, a novel local structure feature called ring gradient descriptor (RGD) is proposed, which scans the background around the target in an annular way to find the region most similar to the target and calculates the similarity between the region and the target. Then, a combination of RGD and the salience and texture features of candidate objects is used by Gradient Boosting Decision Tree (GBDT) for the final candidate classification.
The major contributions of this paper can be summarized as follows.
  • A novel method is proposed for the accurate and reliable detection of microaneurysms with the possibility of applying this method in large screening setups.
  • A simple candidate extraction algorithm based on morphology is proposed to extract the potential MA in fundus images.
  • A new local structure feature RGD is proposed that can describe the local structure of object and its surrounding background and improve the classification performance.

2. Related Works

The detection algorithm of MA can be divided into the physical model-based method, classifier-based method, and deep learning-based detection methods.
Physical model-based MA detection methods are mainly based on the physical characteristics of retinal MA. Joshi et al. [8] employed morphological methods to enhance fundus images and remove blood vessels and then extracted MAs. Zhang et al. [9] proposed a feature-transfer network and local background suppression for MA detection by using the similarity matrix of feature distances to measure the difference between background noise and retinal objects to suppress the local background. Quellec et al. [10] did not perform MA detection in color fundus images but used a lesion template for MA matching in its wavelet-transformed images.
The classifier-based method is the most prevalent method at present, the main process of this method includes candidate extraction, feature extraction, and candidate classification. Orlando et al. [3] employed morphological reconstruction to extract MA candidates. In the process of feature extraction, they constructed a Convolutional Neural Network (CNN) model to extract depth features and combined it with color features, textural features, and geometrical features for candidates classification. Dashtbozorg et al. [4] used a gradient-weighting technique and an iterative thresholding approach to extract MA candidates and used the response of local convergence index filters and the salience of the candidate area for classification.
Melo et al. [5] used a sliding band filter for MA enhancement and they also used the filter response and the salience of the candidate area for classification. Antal and Hajdu [6] proposed an ensemble-based framework for MA detection; they selected the optimal results under different preprocessing and candidate extraction methods. Shah et al. [7] removed blood vessels from the green channel and extracted MA candidates using a local thresholding technique. They classified MAs and non-MAs based on statistical features.
The deep learning-based detection method mainly regards MA detection as a segmentation task. Xu et al. [11] improved the U-Net model and achieved pixel-level segmentation of MA. Liao et al. [12] proposed a novel deep convolutional encoder–decoder network for MA detection. Budak et al. [13] used a CNN trained with preprocessed RGB patches to classify MA patches and non-MA patches.
Due to the complex structure of retina and the uneven color and brightness of fundus images, physical model-based methods often have unstable detection effects and low detection accuracy. Classification-based methods tend to have high accuracy because a large number of features are extracted from candidate region images. Deep neural networks have been widely used in the field of computer vision; however, MA detection methods based on deep learning may lead to the existence of over-fitting due to the small amount of data. In addition, deep convolutional neural networks have a large number of parameters and are therefore not easy to use clinically.
In this paper, the classifier-based method was used for MA detection. Previous researchers focused on the target salience of candidates. In addition, we also paid attention to the local structure of the target and background, and proposed the novel local structure feature RGD to overcome the shortcomings of the salience features.

3. Materials

We conducted experiments using two publicly available datasets: e-ophtha-MA [14] and ROC [15]. The main specifications of the two datasets are summarized in Table 1.
ROC: The Retinopathy Online Challenge (ROC) contains 50 training images and 50 test images, and all MAs were annotated by four experts. The images have different resolutions, ranging from 768 × 576 to 1394 × 1392 pixels with a 45° field of view (FOV). Since the test images do not have MA annotations, only 50 training images were used to verify our proposed algorithm.
E-ophtha-MA: The e-ophtha-MA is a public dataset of color fundus images designed for scientific research in red lesion (MA and small hemorrhage) detection. It contains 233 healthy images and 148 DR images with four resolutions, ranging from 1440 × 960 to 2544 × 1696 pixels with 45° FOV. All images are used to verify our proposed algorithm.
To assess the ability of the new structural features proposed in this paper, the results of the classification between MAs and non-MAs can be evaluated by the receiver operating characteristic (ROC) curve [16] by plotting the true positive rate (TPR) against the false positive rate (FPR) and the area under the ROC curve (AUC). Different TPR and FPR values can be obtained with different thresholds. They are defined as:
T P R = T P P
F P R = F P N
where P and N correspond to the number of MAs and non-MAs in candidates, respectively. T P is the number of MAs correctly detected, F P is the number of MAs incorrectly detected.
In addition, we evaluated the performance of the detection algorithm at the lesion level. The free-response operating characteristic (FROC) curve [17] was used to evaluate the lesion level of MA detection results of all color images. The abscissa of FROC curve is the average number of false positives per image (FPI), and the ordinate is the sensitivity. The sensitivity represents the proportion of MAs correctly detected by the algorithm. These are calculated as follows:
S e n s i t i v i t y = T P N M A
F P I = F P N i
where N M A is the number of MAs in all fundus images in the test dataset, N i is the number of images in the test dataset, T P is the number of MAs correctly detected, F P is the number of MAs incorrectly detected. By setting the threshold to classify MA and non-MA, we can obtain pairs < F P I , S e n s i t i v i t y > to draw the FROC curve.
Under the same FPI, higher sensitivity means better detection performance. Similarly, under the same sensitivity, lower FPI means fewer misdetected MA in one image. In order to compare with different methods, we obtained the sensitivity values from FROC curve as the FPI values are 1/8, 1/4, 1/2, 1, 2, 4, and 8. In addition, the average of sensitivity at these seven predefined FPIs ( F s c o r e ) and the partial area under FROC curves between 1/8 and 8 FPI normalized by dividing with the maximum FPI ( F A U C ) were obtained as the comprehensive evaluation indexes.
Under different detection tasks, F s c o r e and F A U C often have different optimal ranges. Even under the same detection task, the number of images also has a great influence on them. In the ROC dataset, the optimal value of these two evaluation indexes should be greater than 0.4, while in the e-ophtha-MA dataset, it should be greater than 0.5

4. Methods

A schematic diagram of our method is illustrated in Figure 2. It includes three parts: First, the original image was preprocessed to eliminate the interference of uneven color (see Section 4.1). Second, we extracted the MA candidates (see Section 4.2). Finally, a patch centered on each region was collected to extract target salience and local structural features.The target salience include mean, standard deviation (SD), third moment (TM), energy, entropy, and contrast. The local structural features include the texture feature based on Gray level co-occurrence matrix (GLCM) and RGD. Then, we used the combined features to classify candidates to MAs and non-MAs (see Section 4.3).

4.1. Image Preprocessing

Due to the interference of environment and equipment in the process of retinal image acquisition and the differences of people’s age and ethnicities [18], the captured fundus images often generally have nonuniform illumination and different colors.
In order to reduce the subsequent computation, we first resized the input image with a scaling factor χ / 1400 , where χ is related to the width in pixels of the input image. Subsequently, a novel approach proposed by Grinsven et al. [19] was utilized on each channel to enhance the original fundus image with the following equation:
I i , j ; σ = α · I i , j + τ · G a u s s i a n i . j ; σ I i , j + γ
where ∗ is a convolution operator, σ is the standard deviation of the Gaussian filter, α , τ , and γ are constants. These parameters were set following Grinsven et al. [19], i.e., α = 4 , τ = 4 , γ = 128 and σ = χ / 30 .
After the image enhancement, there is a great deal of noise in the retina edge. In order to eliminate this noise, we performed region of interest (ROI) detection on it; As shown in Figure 3c, the binary mask of the ROI was obtained by threshold segmentation. Finally, the preprocessed image I c e was measured by means of morphology operations as given in Equation (6):
I c e i , j = I u n i i , j · ε B G m a s k
where I u n i and G m a s k correspond to the enhanced image and ROI mask, respectively. ε B denotes the erosion of an image performed by structural element B with disc type. Figure 3 illustrates the entire preprocessing procedure.

4.2. Candidate Extraction

After color correction, the gray value of fundus image is more uniform, which is more conducive to the extraction of the lesion area by threshold segmentation. An effective candidate extraction method should capture MAs as much as possible and capture fewer non-MAs. To accomplish this, we propose a novel candidate extraction algorithm based on dual-gray threshold segmentation and morphological processing. Figure 4 illustrates the entire candidate extraction procedure.
Due to the main information of microaneurysm being in the green channel [3], we first extracted the green channel G c e from I c e . As the main non-MAs in the retinal image come from blood vessels, we first performed vessel segmentation.
We obtained the first binary image I h t of the low-gray area through threshold segmentation with a higher gray threshold T h with vessels with more connectivity. The value of T h ranges from 100 to 115. Then, we reserve the connected domain with an area greater than S through connected domain analysis. The retinal blood vessels mask I v e can be obtained by Equation (7):
I v e ( i , j ) = I h t ( x , y ) if s ( x , y ) > S 0 if s ( x , y ) S
where I v e corresponds to the vessel mask, and I h t corresponds to the first binary image. s ( x , y ) correspond to the area of the connected domain in which pixel ( x , y ) is located. As the retinal vessels occupy a large area, the value of S is set to 400.
After the vessel mask I v e is obtained, we expanded I v e to ensure that the vessel edge can also be eliminated. In addition, we obtained the second binary image I l t with a lower threshold T l through threshold segmentation, so that fewer non-MAs can be captured. The value of T l ranged from 90 to 100. Then, the binary mask of candidate area I b w was obtained according to the following equation:
I b w = I l t · ( 1 ε B ( I v e ) )
where I b w and I l t correspond to the binary mask of candidate area and the vessel mask, respectively. ε B ( ) denotes erosion of an image performed by rectangular element B with size k × k . In this article, k = 5 . Then, we conducted the connected domain analysis for I b w and deleted the pixels whose connected domain area was equal to 1.
We considered each of the connected domains in the binary image I B W as a candidate region for the possible existence of MA. Consequently, we use the center coordinates of each connected domain to extract a certain size image from the preprocessed image I c e for feature extraction and target recognition.

4.3. Feature Extraction and Classification

The salience and local structural features were extracted from each candidate. The local structure features include local texture based on Gray level co-occurrence matrix (GLCM) and our proposed RGD. Then, we combined all the features and used GBDT to classify MA candidates.

4.3.1. Object Salience

Saliency features of objects have been widely used in classification and detection tasks [20,21,22,23,24]. Since different candidate regions have different sizes and shapes, we extracted image patches of a certain size, which is sufficient to contain one MA from the center of each candidate region as the source of salience features. The size is set to 11 × 11 . In this study, six salience features were extracted from the green channel, including the mean, standard deviation(SD), third moment (TM), energy, entropy, and contrast [25,26,27].

4.3.2. Local Structures

Local structures include the local texture and RGD. The local texture indicate the homogeneity information of objects [28,29,30,31], which is calculated depending on the pixels and their surroundings [32]. We implemented texture feature extraction based on GLCM [33] and six features based on Haralick features [34] were obtained on one offset GLCM matrix. In our work, we selected four different offsets (0 , 45 , 90 , 135 ) resulting in 24-dimensional Haralick features.
The local texture features used in this article are shown in Table 2. p ( i , j ) correspond the ( i , j ) th entry in normalized GLCM, p x ( i ) correspond ith entry in the marginal-probability matrix obtained by summing the rows of p ( i , j ) , p y ( j ) correspond jth entry in the marginal-probability matrix obtained by summing the columns of p ( i , j ) , and μ x , μ y , σ x , and σ y are the means and standard deviations of p x and p y . p x y = i = 1 N g j = 1 N g p ( i , j ) , i j = 0 , 1 , , N g 1 , and N g is the number of gray levels.
The object salience and local texture ignore the relationship between the target and the surrounding background, so their description ability is not enough. By introducing the surrounding background, we can observe a large degree of structural difference between MA and non-MA in this local region.
Vessels are the major component of non-MAs. We defined the l × l area in the center of the candidate area as the central area A c e , and the adjacent area with width b as the surrounding area A s r . As shown in Figure 5, if the target contained in the candidate region image is MA, we can see that the surrounding region is the retinal background. If the target of the candidate region image is a blood vessel, we can always find a region with the lowest gray mean similar to the central region in the surrounding region. Therefore, we propose a novel local structure feature called a ring gradient descriptor (RGD) to calculate the minimum gradient between the candidate image block and its surroundings to distinguish MAs from vessels.
First, we find the region with the minimum gray mean A s m in the surrounding region in the area around the target through annular scanning, and the size of the scan box is A s m is b × b . Since the blood vessels are multi-directional, we set the moving step of the scan box to 1 pixel each time. Then, the number of scanning K can be calculated as 4 ( b + l ) and the minimum gradient between A c e and A s m can be calculated by the following equation:
R G D = 1 b 2 i = 1 b j = 1 b A s m ( i , j ) A c e ( i , j ) G ( i , j ; σ )
where G ( i , j ; σ ) is the Gaussian kernel with standard deviation σ , which has the same size as the central region. ∗ is the convolution operator. Since the gray level of MA is lower in the center and higher in the surrounding area, Gaussian convolution is used instead of calculating the gray level mean, mainly to control the weight of each pixel.
The two most important parameters in the RGD algorithm are l and b, which are determined by the size of MA and the distance from the surrounding blood vessels. The value of σ transforms with l. As shown in Figure 6, different candidates have different optimal parameters. In order to make the model more robust, we calculated 10 RGDs under different parameters as new local structural features as shown in Table 3.

4.3.3. Classify

To distinguish between MAs and non-MAs, the GBDT classifier [35], which is an ensemble classifier that has been used in general applications, was employed in our work. The increasing popularity of this classifier is mainly attributed to its faster training speed and its robustness. As introduced before, a training set S G B = x i , y i , i = 1 , 2 , , N is constructed by combining features x and corresponding label.
The establishment process of the GBDT is shown in Figure 7, It consists of M base classifiers. The base learner of GBDT is the classification and regression tree (CART). The complete algorithm process of GBDT binary classification algorithm is as follows:
(1) Initialize the first weak CART:
F 0 x = l o g P ( Y = 1 | x ) 1 P ( Y = 1 | x )
where P ( Y = 1 | x ) is the proportion of MAs in the training sample.
(2) Perform m ( m = 1 , 2 , 3 M ) iterations on the base learner:
For i = 1 , 2 , , N , calculate the response value corresponding to the mth tree (negative gradient of the loss function):
r m , i = y i 1 1 + e F m 1 ( x i )
For i = 1 , 2 , , N , use CART regression tree fitting data ( x i , r m , i ) to get the mth regression tree, whose corresponding leaf node area is R m , j , where j = 1 , 2 , , J m and J m is the number of leaf nodes of the mth regression tree.
For J m leaf node region j = 1 , 2 , J m , the best fitting value c m , j was calculated:
c m , j = x i R m , j r m , i x i R m . j ( y i r m , i ) ( 1 y i + r m , i )
Update the strong classifier F m x :
F m x = F m 1 x + j = 1 J m c m , j I ( x R m , j )
(3) Then, the final strong classifier F M x can be expressed as:
F M x = F 0 x + m = 1 J m j = 1 J m c m , j I ( x R m , j )
(4) Finally, the classification value of sample x can be expressed as:
P ( Y = 1 | x ) = 1 1 + e F M ( x )
The learning rate of model ( ε ) and the number of levels of trees (C) are the hyperparameters that have the most obvious impact on the accuracy of GBDT model. The prediction accuracy of GBDT is significantly affected by the value of hyperparameter [36]. After plenty of experimentation, the optimal values of ε and C were set to 0.03 and 80, respectively.

5. Results

5.1. Candidate Extraction Evaluation

The performance of candidate extraction proposed in this article was evaluated by sensitivity and compared with the previously published candidate extractor algorithms as demonstrated in Table 4. The proposed method achieved a sensitivity value of 0.51 for the ROC dataset. Although this value did not reach the maximum value, it had a relatively low FPI value, and the performance of the candidate extraction algorithm proposed is better than some algorithms. We obtained a sensitivity value of 0.72 in the ephtha-MA dataset, corresponding to an FPI value of 200.74.

5.2. Candidate Classification Evaluation

We adopted five-fold cross-validation for model training. We trained three models using target saliency and local texture (TSLT), RGD, and the combined features (CF). The performance of classification based on object salience features and local structure features was evaluated by ROC curve as shown in Figure 8. Their AUC values are shown in Table 5.
The combined features achieved much higher AUC (AUC = 0.9752 in e-optha-MA, AUC = 0.9409 in ROC) than using traditional TSLT features individually (AUC = 0.9615 in e-optha-MA, AUC = 0.9066 in ROC). The AUC value (AUC = 0.9566) obtained using RGD features were close to those obtained using TSLT features (AUC = 0.9615) in dataset e-optha-MA, and the AUC values (AUC = 0.9205) obtained by using RGD features in dataset ROC exceeded those (AUC = 0.9066) obtained by using TSLT features alone. Therefore, combining traditional features with RGD can improve the classification performance of the model.
Figure 9 presents the FROC curves obtained by not considering the missing MAs from the candidate extraction step and consider. Since many MAs were omitted in the extraction algorithm of candidates, the F s o c r e and F A U C values of the final model would be reduced. The values of F s c o r e and F A U C in dataset e-optha-MA without considering the omission of MA in the second step were 0.591 and 0.794, which are much higher than the values of 0.434 and 0.583 considering the omission of MA, and the values of 0.349 and 0.519 in dataset ROC are higher than the values of 0.188 and 0.280.
The final results after candidate extraction and candidate classification at the lesion level were compared with other MA detection methods in the ROC and e-ophtha MA datasets as shown in Table 6. After verification, our algorithm is ahead of some other algorithms in the ROC dataset ( F s c o r e = 0.264 , F A U C = 0.356 ), and the F A U C and F s c o r e values achieved a leading level in the e-ophtha MA dataset ( F s c o r e = 0.547 , F A U C = 0.630 ). Although we achieved a high detection performance in the e-ophtha MA datasets, the result in the ROC dataset was poor because of the simplicity of the candidate extraction algorithm.
The sensitivity values of 0.51 and 0.72 in the ROC and e-ophtha MA datasets were achieved in the candidate extraction step, respectively, which can be said to be the upper limit of the final detection algorithm. The proposed method achieved the sensitivity values of 0.468 and 0.696, respectively, at the two datasets where FPIs is 8, which were close to the upper limit. This shows that our classification algorithm has high performance and the features we extracted were very effective.
In order to more qualitatively display the performance of RGD on MA detection, we used six 224 × 224 color images containing MAs for pixel-level validation, as shown in Figure 10. The MA candidate region was obtained through dual-threshold segmentation proposed by us. RGD was performed for each pixel in the MA candidate region and the values of b, l, and σ are 5.7, and 0.6, respectively. Then, we normalized the results and segmented them with thresholds of 0.6 and 0.8, respectively. We found that the use of RGD alone was also effective to detect MAs.

6. Discussion

In this paper, we proposed an efficient local structure awareness-based retinal MA detection method with the multi-feature combination (LSAMM). First, the color correction was performed on the images, and simple morphology and threshold segmentation method was used to extract MA candidates. The sensitivity values of this stage in the ROC and e-ophtha MA dataset, respectively, were 0.51 and 0.72, and their corresponding FPI values were 243.38 and 200.74, respectively.
In the candidate classification step, a novel local structure feature RGD was proposed that can effectively distinguish MA and vessels and improve the performance of classification. After using this, the AUC value in the e-ophtha MA dataset increased from 0.96153 to 0.97515, and in the ROC dataset, the AUC value increased from 0.90658 to 0.94060. The whole MA detection algorithm proposed achieved a high detection performance in the e-ophtha MA dataset ( F s c o r e = 0.567 ); however, its performance in the ROC dataset was mediocre ( F s c o r e = 0.264 ) as the sensitivity value in candidate extraction step was low.
The candidate extraction algorithm performed worse in the ROC dataset than in the e-ophtha MA dataset. Nevertheless, the results of candidate classification showed high performance in both datasets. The detection performance of MA can be improved by changing the candidate extraction algorithm and combining RGD with other conventional features in the candidate classification step.

Author Contributions

All authors have contributed substantially to, and are in agreement with the content of, the manuscript. Conception/design, provision of study materials, and the collection and/or assembly of data: J.D. and P.T.; data analysis and interpretation: X.Z. and C.Q.; manuscript preparation: J.D., T.P. and Z.P.; final approval of the manuscript: J.D., P.T., X.Z., T.P., Z.P. and C.Q. The guarantor of the paper takes responsibility for the integrity of the work as a whole, from its inception to publication. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China (grant numbers 61775030, 61571096), and the Sichuan Science and Technology Program (grant numbers 22ZDYF2690).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

All data are available from the corresponding author upon request.

Acknowledgments

The authors acknowledge the staff at Laboratory of Imaging Detection and Intelligent Perception University of Electronic Science and Technology of China, Sichuan Academy of Medical Sciences and Sichuan Provincial People’s Hospital of China.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Long, H.; Chen, B.; Li, W.; Xian, Y.; Peng, Z. Blood glucose detection based on Teager-Kaiser main energy of photoacoustic signal. Comput. Biol. Med. 2021, 134, 104552. [Google Scholar] [CrossRef]
  2. Wang, H.; Yuan, G.; Zhao, X.; Peng, L.; Wang, Z.; He, Y.; Qu, C.; Peng, Z. Hard exudate detection based on deep model learned information and multi-feature joint representation for diabetic retinopathy screening. Comput. Methods Programs Biomed. 2020, 191, 105398. [Google Scholar] [CrossRef] [PubMed]
  3. Orlando, J.I.; Prokofyeva, E.; Del Fresno, M.; Blaschko, M.B. An ensemble deep learning based approach for red lesion detection in fundus images. Comput. Methods Programs Biomed. 2018, 153, 115–127. [Google Scholar] [CrossRef] [Green Version]
  4. Dashtbozorg, B.; Zhang, J.; Huang, F.; ter Haar Romeny, B.M. Retinal microaneurysms detection using local convergence index features. IEEE Trans. Image Process. 2018, 27, 3300–3315. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Melo, T.; Mendonça, A.M.; Campilho, A. Microaneurysm detection in color eye fundus images for diabetic retinopathy screening. Comput. Biol. Med. 2020, 126, 103995. [Google Scholar] [CrossRef]
  6. Antal, B.; Hajdu, A. An ensemble-based system for microaneurysm detection and diabetic retinopathy grading. IEEE Trans. Biomed. Eng. 2012, 59, 1720–1726. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Shah, S.A.A.; Laude, A.; Faye, I.; Tang, T.B. Automated microaneurysm detection in diabetic retinopathy using curvelet transform. J. Biomed. Opt. 2016, 21, 101404. [Google Scholar] [CrossRef] [Green Version]
  8. Joshi, S.; Karule, P.T. Mathematical morphology for microaneurysm detection in fundus images. Eur. J. Ophthalmol. 2020, 30, 1135–1142. [Google Scholar] [CrossRef]
  9. Zhang, X.; Wu, J.; Meng, M.; Sun, Y.; Sun, W. Feature-transfer network and local background suppression for microaneurysm detection. Mach. Vis. Appl. 2021, 32, 1–13. [Google Scholar] [CrossRef]
  10. Quellec, G.; Lamard, M.; Josselin, P.M.; Cazuguel, G.; Cochener, B.; Roux, C. Optimal wavelet transform for the detection of microaneurysms in retina photographs. IEEE Trans. Med. Imaging 2008, 27, 1230–1241. [Google Scholar] [CrossRef] [Green Version]
  11. Xu, Y.; Zhou, Z.; Li, X.; Zhang, N.; Zhang, M.; Wei, P. FFU-Net: Feature Fusion U-Net for Lesion Segmentation of Diabetic Retinopathy. BioMed Res. Int. 2021, 2021, 6644071. [Google Scholar] [CrossRef] [PubMed]
  12. Liao, Y.; Xia, H.; Song, S.; Li, H. Microaneurysm detection in fundus images based on a novel end-to-end convolutional neural network. Biocybern. Biomed. Eng. 2021, 41, 589–604. [Google Scholar] [CrossRef]
  13. Budak, U.; Şengür, A.; Guo, Y.; Akbulut, Y. A novel microaneurysms detection approach based on convolutional neural networks with reinforcement sample learning algorithm. Health Inf. Sci. Syst. 2017, 5, 1–10. [Google Scholar] [CrossRef] [PubMed]
  14. Decenciere, E.; Cazuguel, G.; Zhang, X.; Thibault, G.; Klein, J.C.; Meyer, F.; Marcotegui, B.; Quellec, G.; Lamard, M.; Danno, R.; et al. TeleOphta: Machine learning and image processing methods for teleophthalmology. Irbm 2013, 34, 196–203. [Google Scholar] [CrossRef]
  15. Niemeijer, M.; Van Ginneken, B.; Cree, M.J.; Mizutani, A.; Quellec, G.; Sánchez, C.I.; Zhang, B.; Hornero, R.; Lamard, M.; Muramatsu, C.; et al. Retinopathy online challenge: Automatic detection of microaneurysms in digital color fundus photographs. IEEE Trans. Med. Imaging 2009, 29, 185–195. [Google Scholar] [CrossRef]
  16. Mo, J.; Zhang, L.; Feng, Y. Exudate-based diabetic macular edema recognition in retinal images using cascaded deep residual networks. Neurocomputing 2018, 290, 161–171. [Google Scholar] [CrossRef]
  17. Bunch, P.C.; PC, B.; GH, S. A free-response approach to the measurement and characterization of radiographic-observer performance. Neurocomputing 1978, 127, 124–135. [Google Scholar]
  18. Xie, Y.; Nguyen, Q.D.; Hamzah, H.; Lim, G.; Bellemo, V.; Gunasekeran, D.V.; Yip, M.Y.; Lee, X.Q.; Hsu, W.; Lee, M.L.; et al. Artificial intelligence for teleophthalmology-based diabetic retinopathy screening in a national programme: An economic analysis modelling study. Lancet Digital Health 2020, 2, e240–e249. [Google Scholar] [CrossRef]
  19. Van Grinsven, M.J.; van Ginneken, B.; Hoyng, C.B.; Theelen, T.; Sánchez, C.I. Fast convolutional neural network training using selective data sampling: Application to hemorrhage detection in color fundus images. IEEE Trans. Med. Imaging 2016, 35, 1273–1284. [Google Scholar] [CrossRef]
  20. Peng, Z.; Zhang, Q.; Wang, J.; Zhang, Q.P. Dim target detection based on nonlinear multifeature fusion by Karhunen-Loeve transform. Opt. Eng. 2004, 43, 2954–2958. [Google Scholar]
  21. Huang, S.; Liu, Y.; He, Y.; Zhang, T.; Peng, Z. Structure-adaptive clutter suppression for infrared small target detection: Chain-growth filtering. Remote Sens. 2020, 12, 47. [Google Scholar] [CrossRef] [Green Version]
  22. Wang, G.; Tao, B.; Kong, X.; Peng, Z. Infrared Small Target Detection Using Non-Overlapping Patch Spatial-Temporal Tensor Factorization with Capped Nuclear Norm Regularization. IEEE Trans. Geosci. Remote Sens. 2021. [Google Scholar] [CrossRef]
  23. Guan, X.; Peng, Z.; Huang, S.; Chen, Y. Gaussian scale-space enhanced local contrast measure for small infrared target detection. IEEE Geosci. Remote Sens. Lett. 2019, 17, 327–331. [Google Scholar] [CrossRef]
  24. Liu, Y.; Zhang, P.; He, Y.; Peng, Z. River detection based on feature fusion from synthetic aperture radar images. Journal of Applied Remote Sensing. J. Appl. Remote Sens. 2020, 14, 016505. [Google Scholar] [CrossRef]
  25. Sun, L.; Wang, Z.; Pu, H.; Yuan, G.; Guo, L.; Pu, T.; Peng, Z. Attention-embedded complementary-stream CNN for false positive reduction in pulmonary nodule detection. Comput. Biol. Med. 2021, 133, 104357. [Google Scholar] [CrossRef] [PubMed]
  26. Cao, Z.; Kong, X.; Zhu, Q.; Cao, S.; Peng, Z. Infrared dim target detection via mode-k1k2 extension tensor tubal rank under complex ocean environment. ISPRS J. Photogramm. Remote Sens. 2021, 181, 167–190. [Google Scholar] [CrossRef]
  27. Zhang, T.; Peng, Z.; Wu, H.; He, Y.; Li, C.; Yang, C. Infrared small target detection via self-regularized weighted sparse model. Neurocomputing. Neurocomputing 2021, 420, 124–148. [Google Scholar] [CrossRef]
  28. Wang, Y.; Peng, Z.; Han, Y.; He, Y. Seismic attribute analysis with saliency detection in fractional fourier transform domain. J. Earth Sci. 2018, 29, 1372–1379. [Google Scholar] [CrossRef]
  29. Han, Y.; Yang, X.; Pu, T.; Peng, Z. Fine-Grained Recognition for Oriented Ship Against Complex Scenes in Optical Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2021. [Google Scholar] [CrossRef]
  30. Zhang, L.; Peng, L.; Zhang, T.; Cao, S.; Peng, Z. Infrared small target detection via non-convex rank approximation minimization joint l2, 1 norm. Remote Sens. 2018, 10, 1821. [Google Scholar] [CrossRef] [Green Version]
  31. Zhang, L.; Peng, Z. Infrared small target detection based on partial sum of the tensor nuclear norm. Remote Sens. 2019, 11, 382. [Google Scholar] [CrossRef] [Green Version]
  32. Du, J.; Zou, B.; Chen, C.; Xu, Z.; Liu, Q. Automatic microaneurysm detection in fundus image based on local cross-section transformation and multi-feature fusion. Comput. Methods Programs Biomed. 2020, 196, 105687. [Google Scholar] [CrossRef]
  33. Wu, P.; Manjunath, B.S.; Newsam, S.; Shin, H.D. A texture descriptor for browsing and similarity retrieval. Signal Process. Image Commun. 2000, 16, 33–43. [Google Scholar] [CrossRef]
  34. Haralick, R.M.; Shanmugam, K.; Dinstein, I.H.A. Textural features for image classification. IEEE Trans. Syst. Man Cybern. 1973, SMC-3, 610–621. [Google Scholar] [CrossRef] [Green Version]
  35. Bikmukhametov, T.; Jschke, J. Oil Production Monitoring using Gradient Boosting Machine Learning Algorithm. IFAC-PapersOnLine 2019, 52, 514–519. [Google Scholar] [CrossRef]
  36. Zhang, W.; Yu, J.; Zhao, A.; Zhou, X. Predictive model of cooling load for ice storage air-conditioning system by using GBDT. Energy Rep. 2021, 7, 1588–1597. [Google Scholar] [CrossRef]
  37. Dai, B.; Wu, X.; Bu, W. Retinal microaneurysms detection using gradient vector analysis and class imbalance classification. PLoS ONE 2016, 11, e0161556. [Google Scholar] [CrossRef] [Green Version]
  38. Adal, K.M.; Sidibé, D.; Ali, S.; Chaum, E.; Karnowski, T.P.; Mériaudeau, F. Automated detection of microaneurysms using scale-adapted blob analysis and semi-supervised learning. Comput. Methods Programs Biomed. 2014, 114, 1–10. [Google Scholar] [CrossRef] [Green Version]
  39. Walter, T.; Massin, P.; Erginay, A.; Ordonez, R.; Jeulin, C.; Klein, J.C. Automatic detection of microaneurysms in color fundus images. Med. Image Anal. 2007, 11, 555–566. [Google Scholar] [CrossRef] [PubMed]
  40. Zhang, B.; Wu, X.; You, J.; Li, Q.; Karray, F. Detection of microaneurysms using multi-scale correlation coefficients. Pattern Recognit. 2010, 43, 2237–2248. [Google Scholar] [CrossRef]
  41. Chudzik, P.; Majumdar, S.; Calivá, F.; Al-Diri, B.; Hunter, A. Microaneurysm detection using fully convolutional neural networks. Comput. Methods Programs Biomed. 2018, 158, 185–192. [Google Scholar] [CrossRef]
  42. Eftekhari, N.; Pourreza, H.R.; Masoudi, M.; Ghiasi-Shirazi, K.; Saeedi, E. Microaneurysm detection in fundus images using a two-step convolutional neural network. Biomed. Eng. Online 2019, 18, 1–16. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  43. Wu, B.; Zhu, W.; Shi, F.; Zhu, S.; Chen, X. Automatic detection of microaneurysms in retinal fundus images. Comput. Med. Imaging Graph. 2017, 55, 106–112. [Google Scholar] [CrossRef] [PubMed]
  44. Wang, S.; Tang, H.L.; Hu, Y.; Sanei, S.; Saleh, G.M.; Peto, T. Localizing microaneurysms in fundus images through singular spectrum analysis. Health information science and systems. IEEE Trans. Biomed. Eng. 2016, 64, 990–1002. [Google Scholar] [CrossRef] [PubMed]
Figure 1. An example of a color fundus image with MAs. The areas with MAs are identified with a green box and zoomed in on the right side of the image, the dark blobs that the white arrow points to are MAs.
Figure 1. An example of a color fundus image with MAs. The areas with MAs are identified with a green box and zoomed in on the right side of the image, the dark blobs that the white arrow points to are MAs.
Biomedicines 10 00124 g001
Figure 2. Schematic diagram of the automatic MA detection method, including preprocessing, candidate extraction, feature extraction, and classification.
Figure 2. Schematic diagram of the automatic MA detection method, including preprocessing, candidate extraction, feature extraction, and classification.
Biomedicines 10 00124 g002
Figure 3. Illustration of the entire procedure for image preprocessing. (a) Original resized image. (b) Binary mask of the ROI. (c) The preprocessed image.
Figure 3. Illustration of the entire procedure for image preprocessing. (a) Original resized image. (b) Binary mask of the ROI. (c) The preprocessed image.
Biomedicines 10 00124 g003
Figure 4. An example to illustrate the candidate extraction method. From top to bottom and left to right: original image I, the green channel of the preprocessed image G c e , the binary mask of candidate area I b w , details in the green channel of the preprocessed image I c e , the first binary image I h t under a higher threshold, the vessels mask I v e , the second binary image I l t under a lower threshold, and the binary mask of candidate area I b w . The areas with MAs are identified with white boxes.
Figure 4. An example to illustrate the candidate extraction method. From top to bottom and left to right: original image I, the green channel of the preprocessed image G c e , the binary mask of candidate area I b w , details in the green channel of the preprocessed image I c e , the first binary image I h t under a higher threshold, the vessels mask I v e , the second binary image I l t under a lower threshold, and the binary mask of candidate area I b w . The areas with MAs are identified with white boxes.
Biomedicines 10 00124 g004
Figure 5. Examplesof candidate images. (a) Candidate image containing an MA. (b) Candidate image containing a vessel.
Figure 5. Examplesof candidate images. (a) Candidate image containing an MA. (b) Candidate image containing a vessel.
Biomedicines 10 00124 g005
Figure 6. Illustrative example of the optimal parameter settings of different candidates. (ad) are MAs in different environments, with different pixel widths and distances from the adjacent blood vessel. These two values are marked by red and blue markers respectively.
Figure 6. Illustrative example of the optimal parameter settings of different candidates. (ad) are MAs in different environments, with different pixel widths and distances from the adjacent blood vessel. These two values are marked by red and blue markers respectively.
Biomedicines 10 00124 g006
Figure 7. The construction of GBDT [36].
Figure 7. The construction of GBDT [36].
Biomedicines 10 00124 g007
Figure 8. ROC curves of MAs and non-MAs classification on different features. (a) The results of the e-optha-MA dataset. (b) The results of the ROC dataset.
Figure 8. ROC curves of MAs and non-MAs classification on different features. (a) The results of the e-optha-MA dataset. (b) The results of the ROC dataset.
Biomedicines 10 00124 g008
Figure 9. FROC curves of MAs and non-MAs classification. (a) The results of the e-optha-MA dataset. (b) The results of the ROC dataset.
Figure 9. FROC curves of MAs and non-MAs classification. (a) The results of the e-optha-MA dataset. (b) The results of the ROC dataset.
Biomedicines 10 00124 g009
Figure 10. Illustration of pixel level validation. For better visualization, the correctly detected MAs and miss detection MAs are highlighted by red and blue squares, respectively. The FP candidates are highlighted by yellow circles.
Figure 10. Illustration of pixel level validation. For better visualization, the correctly detected MAs and miss detection MAs are highlighted by red and blue squares, respectively. The FP candidates are highlighted by yellow circles.
Biomedicines 10 00124 g010
Table 1. Dataset specifications.
Table 1. Dataset specifications.
Image Size (px)FOV (Degree)FOV Diameter (px)NENINNIPNMA
ROC768 × 576, 1394 × 139245°720–134541337336
e-ophtha-MA1440 × 960, 2544 × 196645°910–192512331481306
FOV: field of view. NE: number of experts. NIN: number of images without MAs. NIP: number of images with MAs. NMA: total number of MAs.
Table 2. The local texture-feature-based GLCM matrix [34].
Table 2. The local texture-feature-based GLCM matrix [34].
Feature NameDescription
correlation i j ( i j ) p ( i , j ) μ x μ y σ x σ y
inverse difference moment i j 1 1 + ( i j ) 2 p ( i , j )
difference variancevariance of p x y
entropy i j p ( i , j ) l o g ( p ( i , j ) )
angular moment i j ( p ( i , j ) ) 2
contrast i j ( i j ) 2 p ( i , j )
Table 3. Local structural features.
Table 3. Local structural features.
ParametersDescription
bThe width of the surrounding area3333355555
lThe width of the central area57911135791113
σ The standard deviation of Gaussian kernel0.90.70.50.50.50.90.70.50.50.5
Table 4. Candidate extraction performance using the ROC dataset.
Table 4. Candidate extraction performance using the ROC dataset.
MethodSensitivityFPI
Proposed method0.51243.38
Shah et al. [7]0.4865.00
Dai et al. [37]0.69569.39
Adal et al. [38]0.4535.2
Walter et al. [39]0.36154.42
Zhang et al. [40]0.33328.30
Dashtbozorg et al. [4]0.82755.50
Table 5. Validation of different features in MA candidate classification.
Table 5. Validation of different features in MA candidate classification.
DatabaseMethodAUC
e-optha-MATSLT0.9615
RGD0.9566
CF0.9752
ROCTSLT0.9066
RGD0.9205
CF0.9409
Table 6. Performance of different MA detection methods in the ROC and e-ophtha MA datasets.
Table 6. Performance of different MA detection methods in the ROC and e-ophtha MA datasets.
DatabaseWorkSensitivty against FPIs F score F AUC
ROC 1/81/41/21248
Proposed work0.0830.1040.2000.2570.3440.3940.4680.2640.356
Chudzik et al. [41]0.0390.0670.1410.1470.2430.3060.3850.193-
Dashtbozorg et al. [4]0.4350.4430.4540.4790.4810.4950.5060.4710.484
Eftekhari et al. [42]0.0470.1730.3510.5520.6130.7220.7690.4610.660
Wu et al. [43]0.0370.0560.1030.2060.2950.3390.3760.202-
Budak et al. [13]0.0390.0610.1210.2200.3380.3720.3940.221-
Wang et al. [44]0.2730.3790.3980.4810.5450.5760.5980.464-
Dai et al. [37]0.2190.2570.3380.4290.5280.5980.6620.4330.553
Antal and Hjdu [6]0.1730.2750.3800.4440.5260.5990.6430.4340.551
Melo et al. [5]0.0530.0660.0770.0980.1460.2080.2590.1300.185
e-ophtha-MAProposed work0.3350.4240.4960.5780.6340.6680.6960.5470.630
Wu et al. [43]0.0630.1170.1720.2450.3230.4170.5730.273-
Dashtbozorg et al. [4]0.3580.4170.4170.5220.5580.6050.6380.5100.575
Eftehari et al. [42]0.0910.2580.4010.5340.5790.6670.7710.4710.637
Chudzik et al. [41]0.1850.3130.4650.6040.7160.8010.8490.562-
Melo et al. [5]0.1780.2840.3830.5190.5870.5870.5870.4460.551
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Deng, J.; Tang, P.; Zhao, X.; Pu, T.; Qu, C.; Peng, Z. Local Structure Awareness-Based Retinal Microaneurysm Detection with Multi-Feature Combination. Biomedicines 2022, 10, 124. https://doi.org/10.3390/biomedicines10010124

AMA Style

Deng J, Tang P, Zhao X, Pu T, Qu C, Peng Z. Local Structure Awareness-Based Retinal Microaneurysm Detection with Multi-Feature Combination. Biomedicines. 2022; 10(1):124. https://doi.org/10.3390/biomedicines10010124

Chicago/Turabian Style

Deng, Jiakun, Puying Tang, Xuegong Zhao, Tian Pu, Chao Qu, and Zhenming Peng. 2022. "Local Structure Awareness-Based Retinal Microaneurysm Detection with Multi-Feature Combination" Biomedicines 10, no. 1: 124. https://doi.org/10.3390/biomedicines10010124

APA Style

Deng, J., Tang, P., Zhao, X., Pu, T., Qu, C., & Peng, Z. (2022). Local Structure Awareness-Based Retinal Microaneurysm Detection with Multi-Feature Combination. Biomedicines, 10(1), 124. https://doi.org/10.3390/biomedicines10010124

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop