Next Article in Journal
Graphene-Based Sensor for the Detection of Cortisol for Stress Level Monitoring and Diagnostics
Previous Article in Journal
The Relationship between Visceral Adiposity and Nonalcoholic Fatty Liver Disease Diagnosed by Controlled Attenuation Parameter in People with HIV: A Pilot Study
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Improved Detection Algorithm for Ischemic Stroke NCCT Based on YOLOv5

1
School of Information and Communication Engineering, Beijing University of Posts and Telecommunications, Beijing 100876, China
2
State Key Laboratory of Networking & Switching Technology, Beijing University of the Posts and Telecommunications, Beijing 100876, China
3
Department of Rehabilitation Medicine, Beijing Tiantan Hospital, Capital Medical University, Beijing 100070, China
4
Sloan School of Management, Massachusetts Institute of Technology, Cambridge, MA 02139, USA
5
Center for Statistics and Data Science, Massachusetts Institute of Technology, Cambridge, MA 02139, USA
*
Author to whom correspondence should be addressed.
Diagnostics 2022, 12(11), 2591; https://doi.org/10.3390/diagnostics12112591
Submission received: 2 September 2022 / Revised: 11 October 2022 / Accepted: 19 October 2022 / Published: 26 October 2022
(This article belongs to the Section Machine Learning and Artificial Intelligence in Diagnostics)

Abstract

:
Cerebral stroke (CS) is a heterogeneous syndrome caused by multiple disease mechanisms. Ischemic stroke (IS) is a subtype of CS that causes a disruption of cerebral blood flow with subsequent tissue damage. Noncontrast computer tomography (NCCT) is one of the most important IS detection methods. It is difficult to select the features of IS CT within computational image analysis. In this paper, we propose AC-YOLOv5, which is an improved detection algorithm for IS. The algorithm amplifies the features of IS via an NCCT image based on adaptive local region contrast enhancement, which then detects the region of interest via YOLOv5, which is one of the best detection algorithms at present. The proposed algorithm was tested on two datasets, and seven control group experiments were added, including popular detection algorithms at present and other detection algorithms based on image enhancement. The experimental results show that the proposed algorithm has a high accuracy (94.1% and 91.7%) and recall (85.3% and 88.6%) rate; the recall result is especially notable. This proves the excellent performance of the accuracy, robustness, and generalizability of the algorithm.

1. Introduction

Cerebral Stroke (CS) is a heterogeneous syndrome caused by multiple disease mechanisms and causes a disruption of cerebral blood flow with subsequent tissue damage [1,2]. CS is an acute cerebrovascular disease with a high incidence rate, high mortality rate, and high disability rate [3]. CS occurs when the blood supply in the brain is damaged. It is a result that blocks cerebral vessels and stops blood from flowing to an area of the brain (called an ischemic stroke) or bursts cerebral vessels and causes intracerebral hemorrhage (called a hemorrhagic stroke). Every year, 15 million people suffer from stroke all over the world, including 5 million deaths and 5 million permanent disabilities. Survivors can experience loss of vision and/or speech, paralysis, and confusion [4]. Investigations have shown that CS is the primary reason for adult disability in some countries [2].
The incidence rate of ischemic stroke (IS) is higher at 60~70% than CS. The most common symptoms of IS are a sudden weakness of the face, arms, or legs, sudden dizziness and unconsciousness, sudden mouth/eye deviation, hemiplegia, confusion, difficulty in speaking or understanding, monocular or binocular vision difficulty, difficulty in walking, dizziness, loss of balance or coordination, severe headache without cause, fainting, etc. [5,6,7]. The severity and symptom duration of neurological dysfunction after cerebral artery stenosis and occlusion can be divided into three types: transient ischemic attack, reversible ischemic neurologic deficit, and complete stroke, among which the symptoms of a complete stroke are the most serious [8]. After the onset of IS, it is painful for patients, their families, and society. IS is a cerebrovascular disease with a high incidence in middle age, mostly in people over 40 years old.
At present, prevention is considered to be the best measure for IS because of the lack of effective treatment. For the prevention of stroke, effective prevention strategies include reducing the intake of salt in a diet, eating more fresh vegetables and fruits, strengthening an appropriate amount of physical exercise, avoiding excessive drinking, and banning smoking, etc., in daily life, which has been proven to effectively reduce the incidence of cardiovascular disease. In addition, the regular screening of healthy people and regular monitoring and intervention of high-risk populations with disease risk have also proved to be effective. Necessary intervention measures are proposed for potential patients too. Timely and accurate diagnosis plays an important role in preventing and developing a treatment plan. The clinical diagnosis includes detection in blood and urine, electrocardiogram and blood pressure, cerebral angiography, and brain imaging technology, in which brain imaging technology plays an important role in the cause, location, and area of a stroke [9]. Brain imaging technology for imaging brain tissue or structures utilizes rays (X-ray, γ, etc.) or radio waves, such as noncontrast computer tomography (NCCT), CT angiography (CTA), CT perfusion (CTP), magnetic resonance imaging (MRI). The imaging duration of NCCT is less than one hour at the fastest, and the cost is half or even lower than other brain imaging technologies [10]. Therefore, NCCT is actually a common detection method for stroke, especially for acute IS [11,12].
Computer-aided diagnosis (CAD) is increasingly becoming a reliable and essential means against the background of artificial intelligence (AI) [13,14,15,16,17,18]. For the region of interest (ROI) detection, the current CAD models detect the ROI by mining valuable features in the images. Consequently, the more obvious the features of the ROI, the better the detection. However, the characteristics of NCCT for acute IS are relatively simple, whether it is the composition elements or the color in the image area, especially in the ROI, the hypodensity in the infarct area, the gray matter, types of gray matters, etc. [19], which have differences compared to some diseases, such as tumors with obvious tissue deformation; this leads to a lower feature detection sensitivity of NCCT for IS [20]. Chalela et al. evaluated NCCT to screen acute ischemic stroke. The result was only 56 acute IS patients were screened from 217 acute IS patients, with a sensitivity of 26% [21]. Lansberg et al. also found that the screen rate of acute IS from NCCT was 42–63% [22]. Marbun et al. proposed a method that combined contrast-limited adaptive histogram equalization (CLAHE) and CNN to detect stroke NCCTs and achieved 90% accuracy and 76% sensitivity using a small test set [23]. Kuang et al. studied a model based on a random forest classifier to detect 100 testing acute IS patient NCCTs with a specificity of 91.8% and a sensitivity of 66.2% [24]. Luo et al. proposed the UCATR for the detection of acute IS lesions. UCATR uses a transformer to learn the global features of NCCT, and CNN is used to detect the ROI. The result showed good performance, with a Dice similarity coefficient of 73.58% and a sensitivity of 73.12% [25]. Therefore, we studied a detection algorithm that can realize high accuracy and sensitivity detection for IS NCCT.

2. Materials and Methods

2.1. Image Enhancement

The enhanced image is a kind of data transformation for image data in order to highlight some feature information, creating a better visual effect that can easily fit the subsequent image processing. Its purpose is to make some features that are not obvious in the original image easier to interpret and recognize. These features are often selective. At the same time, the way to enhance them is relative, which can enhance the feature information of interest or weaken the feature information of interest.
The common methods of enhancement can be divided into the spatial domain and frequency domain, according to the processing spaces of the target object [26,27]. The objective of the spatial method is the enhancement of the image itself. The operation is for pixels and components composed of pixels in the image, including the grayscale transformation, histogram equalization, and the smoothing of image areas; the frequency method refers to the enhancement or suppression of some features through processing for the frequency domain. For example, reducing noise by setting a low-pass filter.
As one of the most classic spatial enhancement methods, histogram equalization (HE) is an algorithm that uses the pixel conversion of an image to increase image contrast [28]. Take the grayscale image as an example; firstly, create a histogram of the grayscale values in the selected area; secondly, convert the grayscale histogram into an approximate uniform distribution and, finally, transform the grayscale histogram into an image based on the approximate uniform distribution (according to mapping function) to achieve an image contrast adjustment. With the deepening of research into different application scenarios, some new derivative algorithms of histogram equalization have been proposed, such as adaptive histogram equalization (AHE) [29], contrast-limited adaptive histogram equalization (CLAHE) [30], and adaptive contrast enhancement using the local region stretching [31]. The adaptive contrast enhancement using local region stretching algorithm (ACELES) firstly divides the image according to its brightness and then performs HE on the three regions; then, after the HE of the three regions, the equalization result and the input are weighted averages to realize the output of the result. In addition, grayscale stretching is an image enhancement method that uses grayscale transformation via a linear transformation function. Its main idea is to improve the dynamic range of the grayscale levels in image processing [32]. Zhang et al. studied an image enhancement algorithm based on K-means, which was used to deal with imaging in low-brightness environments [33].

2.2. The Model of Detection Based on YOLOv5

Conventional convolutional neural networks (CNN), such as region CNN (R-CNN) and fast region CNN (fast R-CNN), are two-stage target detection algorithms [34,35]. In the two-stage detection, the model will extract the features of the detection region and generate the region proposal (RP) during the first stage, and then use the CNN to classify the samples in the second stage. The YOLO (you only look once) series as deep learning with one stage has been proposed since 2016 and has developed into YOLOv5 (the fifth vision YOLO algorithm). YOLOv5 is the one of state-of-the-art algorithms in the YOLO family, which is one of the most popular target detection algorithms at present and is applied in target detection and recognition [36,37]. The YOLO algorithm is a one-stage target detection algorithm, which the model can directly use for the network structure to extract features and realize the detection and classification of samples, leaving out the process of RP. Therefore, the YOLO model is more lightweight and efficient, relying on the detection accuracy and generalization ability [38,39,40].
YOLOv5 is a kind of CNN with the inherent network composition structure of a traditional CNN, including the input layer, convolution layer, and pooling layer. In addition, it also adds new network composition modules [41,42]. The structure of YOLOv5 is shown in Figure 1. It mainly consists of four parts: Input, Backbone, Neck, and Head. The image is input into YOLOv5 through the Input Layer and is completed by the image preprocessing, including normalization, data enhancement, and the setting of the bounding box. The function of the Backbone network is to realize the learning and classification of the image features. It includes one convolution component, the focus component, with one convolution module and several residual network modules. The Neck network is used to improve the diversity of the features and the robustness of the network while acquiring the features of different stages. The design of the head output is used to realize the output of the target detection results. The structure of YOLOv5 is shown in Figure 1.
Specifically, the image inputs the YOLOv5 network from the Input Layer. The Input Layer mainly completes the basic preprocessing operation, including image compression and normalization. The process of normalization is completed with the Batchnorm2d module, as shown in Equation (1); in addition, YOLOv5 uses the Mosaic data enhancement operation in the Input Layer to improve the training speed of the model and the accuracy of the network. In addition, the YOLOv5 model also adds an adaptive anchor frame module and an adaptive picture scaling module. The function of the Backbone network is to realize the learning and classification of the image features. It includes a csparknet53 structure and a focused structure, including one convolution module and 23 residual network modules; the Neck network can improve the variety of features and the robustness of the network by acquiring features at different stages. It includes an important feature fusion module: FPN-PAN. The FPN structure includes three maximum pooling layers, which can enhance the fusion of different features. The PAN can further enhance the fusion of the FPN features. In the FPN-PAN structure, the deep semantic features can be transmitted to the shallow layer through FPN, and the shallow positioning information can be transmitted to the deep layer through the PAN, thus enhancing the multi-scale and multi-level feature map fusion. In the process of learning and acquiring features, the Backbone and Neck networks need to activate functions to fit the features. The activate function is called Leaky_ReLU, as shown in Equation (2). The Head output is located in the last layer of the network. It is the output of the target detection result. It includes three detections using the convolution structure and detects the target on the feature map from the previous network, finally outputting the detection result. At this time, the Sigmoid as an activation function will also be used, as shown in Equation (3).
n = m E m V a r m + ο α + β
where, m is the original data to be normalized, E m and V a r m , respectively, represent the mean and variance of m, n is the normalized data, α and β are the parameters for linear transformation; the parameter ο can avoid meaningless equations when the V a r m is equal to zero.
y i = x i a i , x i < 0 x i , x i 0
where i represents the coefficient of the ith hidden layer, and xi and yi, respectively, represent the input and output of the ith hidden layer, after being processed by the activation function a i 1 , + .
f x = 1 1 + e x
where x and y, respectively, represent the input and output in output layer.

2.3. AC-YOLOv5

In this paper, we propose an improved detection algorithm for IS NCCT based on YOLOv5 (AC-YOLOv5). The original IS NCCT is enhanced by ACELES; then, the enhanced image input to the model based on YOLOv5 is for target detection. The Algorithm 1 is as follows.
Algorithm 1. AC-YOLOv5
Input: The original image;
Output: The Network structure, location, and accuracy of the target(s);
Step 1: Adaptively acquire the gray values of pixel points in the original image, and divide all pixel points into three classes based on the gray values, i.e., light, medium and dark according to the gray values;
Step 2: Perform HE on the three classes respectively, and then weight the results after HE and output the image;
Step 3: Input the image into the YOLOv5. Generate the original feature maps and feature matrix A about the object(s) based on the image;
Step 4: Training to predict and get feature maps and feature matrices A ^ ;
Step 5: The loss function is calculated and optimized according to the matrix A and the matrix A ^ ;
Step 6: Perform iterations from step 3 to step 5 until the desired accuracy is achieved.
The structure of AC-YOLOv5 is shown in Figure 2. Firstly, the model performs ACELES to achieve image enhancement. This step leads to the features of ROI of NCCT becoming more prominent in the global region. The NCCT image is a grayscale image; thus, the grayscale value of a pixel actually represents the brightness of the pixel; Secondly, Input the enhanced image into YOLOv5. The image is divided into several grids of equal size and generates the bounding box(es) of the object(s). The original feature maps for the objects are generated too, which are used to store three feature vectors for the object in the image in the grid, including object position information (center point, height, and the width of the box), the class is inside or outside the grid and the class information. Thirdly, training to obtain the anchor point and box of the object(s); then, predict and get feature maps, including the center point and height and width of the box. The loss function is calculated and optimized according to the original feature maps and the predicted feature maps; finally, repeat the second and third steps until the network convergence condition is reached. Output the position information of the target objects and the width and height of the bounding box(es).
In terms of convergence conditions, we use the loss function as the preset. The loss function consists of three parts: the bounding box deviation, the confidence deviation, and the prediction accuracy deviation, which are shown in Equation (4). There are five terms of polynomial accumulation in the formula, wherein the sum of the first term and the second item describes the position information of the bounding box, the third term and the fourth term describe the confidence information, and the fifth term describes the deviation of the prediction probability.
L o s s = λ coord i = 0 s 2 j = 0 B 1 i j o b j x i x ^ i 2 + y i y ^ i 2 + λ coord i = 0 s 2 j = 0 B 1 i j o b j w i w ^ i 2 + h i h ^ i 2 + i = 0 s 2 j = 0 B 1 i j o b j C i C ^ i 2 + λ n o o b j i = 0 s 2 j = 0 B 1 i j n o o b j C i C ^ i 2 + i = 0 S 2 1 i o b j c c l a s s e s p i c p ^ i c 2
where xi, yi, respectively, represent the abscissa and ordinate of the center point of the anchor box, hi is the width of the anchor box, wi is the height of the anchor box, Ci is the confidence score for the object inside or outside the anchor box, pi(c) is probability of classification. 1 i j o b j represents the jth predictor in the ith box. If there is an object in the area and the confidence of jth predictor is the highest, then the 1 i j o b j is 1. On the contrary, if there is no object in the ith cell, then the 1 i j n o o b j is 1. λ c o o r d and λ n o o b j are constants, and the weight of related terms can be adjusted.

3. Results

3.1. Dataset and Construction of Detection Model

The experimental data for the IS NCCT images we selected consisted of two kinds. One is from GitHub, which is one of the most popular software open-source websites at present. We selected the Hypodense-Segmentation-Using-CNN and AISD datasets. The first dataset contains 250 CT images. The second dataset, AISD, contains 397 CT images of IS, all of which are from IS patients and have been labeled for the ROI regions. The second kind of datat came from open-source website resources [43,44,45,46,47], and the ROI region labeling was also already carried out. The multisource experimental data can better verify the performance of the model, including accuracy and generalization ability.
In terms of the construction of the detection model, the dataset of Hypodense-Segmentation-Using-CNN was selected to complete this process. The dataset is divided into three parts, including the training dataset, validation dataset, and testing dataset. The ratio was 3:1:1, that is, 150 training samples, T, 50 verification samples, V, and 50 test samples, T1. Otherwise, we selected 64 samples from AISD and open website resources to compose the test sample, T2, to verify the accuracy and generalization ability of the model.

3.2. Results and Evaluation

As described in Figure 3, there were training results in the YOLOv5 model’s own results window. There were some items in the results window, including three feature vectors (box (the position information of object, i.e., center point, height, and width of box), obj (inside or outside the grid), cls (class information)), metrics (precision, recall, and mAP (mean Average Precision)). After hundreds of epochs, the three feature vectors and the metrics reach the convergence conditions set by the loss function. This indicates that the AC-YOLOv5 model shows excellent performance, including the accuracy and robustness of the model.
In order to better evaluate the performance of the AC-YOLOv5 model, we added seven control group experiments: Fast R-CNN (FR-CNN), YOLOv5(without image enhancement), the image enhancement based on Kmeans + YOLOv5 (K-YOLOv5), the image enhancement of Grayscale stretching + YOLOv5 (G-YOLOv5), the image enhancement of HE + YOLOv5 (H-YOLOv5), the image enhancement of AHE + YOLOv5 (AH-YOLOv5), and the image enhancement of CLAHE + YOLOv5 (CL-YOLOv5). In addition, we used Accuracy and Recall to evaluate the accuracy of the algorithm. The generalization was verified by T1 and T2.
The experimental results for T1 and T2 are shown in Table 1. The experimental results show that image enhancement based on HE can significantly improve the accuracy of detection and the generalization ability of the model. Moreover, K-YOLOv5 had the worst results, whether that was for T1 or T2. AC-YOLOv5 had the best comprehensive performance. Specifically, the accuracy of the proposed AC-YOLOv5 algorithm on the T1 and T2 sets was 94.1% and 91.7%, and the recall rate was 85.3% and 88.6%, respectively, especially in generalization ability.

4. Discussion

Research shows that timely and rapid detection of IS and targeted interventions can minimize the damage and even save the patient’s life. As a fast-imaging technology, NCCT is one of the most important methods to detect IS [48]. In this paper, we presented a detection algorithm for IS NCCT. The algorithm enhances the features of NCCT to make the ROI easier to detect, and the results show high accuracy and recall performance.
Our proposed algorithm includes an image enhancement module and a detection module. The image enhancement module introduces the adaptive mechanism in the enhancement of NCCT, which makes feature enhancement processing become more adaptive and generalizable. Meanwhile, the image enhancement module can make the local features more prominent rather than global, potentially useless features, which will undoubtedly improve the subsequent detection accuracy. In the detection module, ROI detection is completed based on the YOLOv5 target detection model. YOLOv5, as one of the most advanced target detection algorithms, has a lighter network structure and stronger generalization capability than traditional CNN networks. It has been proven to be an efficient target detection model in many fields [49,50].
Our algorithm has achieved excellent accuracy and recall performance, especially for recall. Other studies have also focused on solving the problem of the insufficient sensitivity of IS NCCT features. From their research results, recall is often an urgent problem, and the verification of the model is mainly completed on a single dataset [51]. Our algorithm has achieved good results and has been verified over several datasets. We propose that AC-YOLOv5 achieves 94.1% accuracy and 85.3% recall and 91.7% accuracy and 88.6% recall on the two different datasets, respectively; it has the best comprehensive performance when compared to the seven control groups in the experiment. In addition, from the experimental results, we also found that the YOLOv5 algorithm has better overall performance than traditional CNN algorithms; the model training of CNN training needs a lot of data as a basis [12], and this will also add additional time costs. Both the traditional CNN algorithm and the YOLOv5 algorithm have the problem of insufficient sensitivity toward feature detection in NCCT, but image enhancement based on HE can effectively solve this problem.

5. Conclusions

CS is a serious disease that threatens human health and normal life, especially IS, which accounts for the majority of CS cases. NCCT is one of the most important methods for IS detection. The AC-YOLOv5 model can effectively solve the problem of feature insensitivity in the CAD of IS NCCT and achieve efficient diagnosis effects with accuracy and generalizability. We also hope to provide a feasible and effective exploration for the CAD of NCCT in other diseases to save time and human resources.

Author Contributions

Conceptualization, L.Z. and A.H.; methodology, L.Z. and H.C.; software, J.L.; validation, L.Z., J.L. and Y.T.; formal analysis, A.H.; investigation, L.Z. and H.C.; resources, L.Z., J.L. and Y.T.; data curation, L.Z., J.L. and Y.T.; writing—original draft preparation, L.Z. and; writing—review and editing, L.Z. and R.E.W.; visualization, L.Z., J.L.; supervision, H.C.; project administration, H.C.; funding acquisition, H.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by [National Natural Science Foundation of China] grant number [62171049] and [On demand service intelligent network management and control prototype system] grant number [2020YFB1807805]. And the APC was funded by [62171049] and [2020YFB1807805].

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Markus, H.S. Cerebral perfusion and stroke. J. Neurol. Neurosurg. Psychiatry 2004, 75, 353–361. [Google Scholar] [CrossRef] [PubMed]
  2. Gibson, C.L. Cerebral Ischemic Stroke: Is Gender Important? J. Cereb. Blood Flow Metab. 2013, 33, 1355–1361. [Google Scholar] [CrossRef] [Green Version]
  3. Cipolla, M.J.; Liebeskind, D.S.; Chan, S.-L. The importance of comorbidities in ischemic stroke: Impact of hypertension on the cerebral circulation. J. Cereb. Blood Flow Metab. 2018, 38, 2129–2149. [Google Scholar] [CrossRef] [PubMed]
  4. Thomalla, G.; Christian, G. Acute imaging for evidence-based treatment of ischemic stroke. Curr. Opin. Neurol. 2019, 32, 521–529. [Google Scholar] [CrossRef] [PubMed]
  5. Thomalla, G.; Boutitie, F.; Fiebach, J.B.; Simonsen, C.Z.; Pedraza, S.; Lemmens, R.; Nighoghossian, N.; Roy, P.; Muir, K.W.; Ebinger, M.; et al. Clinical characteristics of unknown symptom onset stroke patients with and without diffusion-weighted imaging and fluid-attenuated inversion recovery mismatch. Int. J. Stroke 2018, 13, 66–73. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Anderson, D.N. Koro: The Genital Retraction Symptom after Stroke. Br. J. Psychiatry 1990, 157, 142–144. [Google Scholar] [CrossRef] [PubMed]
  7. Bousser, M.G.; Eschwege, E.; Haguenau, M.; Lefaucconnier, J.M.; Thibult, N.; Touboul, D.; Touboul, P.J. “AICLA” controlled trial of aspirin and dipyridamole in the secondary prevention of athero-thrombotic cerebral ischemia. Stroke 1983, 14, 5–14. [Google Scholar] [CrossRef] [Green Version]
  8. Jyotirekha, D.; Rajanikant, G.K.N. Post stroke depression: The sequelae of cerebral stroke. Neurosci. Biobehav. Rev. 2018, 90, 104–114. [Google Scholar]
  9. Diener, H.-C.; Hankey, G.J. Primary and Secondary Prevention of Ischemic Stroke and Cerebral Hemorrhage: JACC Focus Seminar. J. Am. Coll. Cardiol. 2020, 75, 1804–1818. [Google Scholar] [CrossRef]
  10. CT Scan, vs. MRI—Uses, Cost, Risks & More. Available online: https://www.docpanel.com/blog/post/ct-scan-vs-mri-uses-cost-risks-more (accessed on 13 August 2019).
  11. Rudkin, S.; Cerejo, R.; Tayal, A.; Goldberg, M.F. Imaging of acute ischemic stroke. Emerg. Radiol. 2018, 25, 659–672. [Google Scholar] [CrossRef]
  12. Soun, J.E.; Chow, D.S.; Nagamine, M.; Takhtawala, R.S.; Filippi, C.G.; Yu, W.; Chang, P.D. Artificial Intelligence and Acute Stroke Imaging. Am. J. Neuroradiol. 2021, 42, 2–11. [Google Scholar] [CrossRef] [PubMed]
  13. Shen, D.; Wu, G.; Suk, H.-I. Deep-Learning-Based Thrombus Localization and Segmentation in Patients with Posterior Circulation Stroke. Annu. Rev. Biomed. Eng. 2017, 19, 221–248. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Yang, Y.; Zhou, Y.; Zhou, C.; Ma, X. Novel computer aided diagnostic models on multimodality medical images to differentiate well differentiated liposarcomas from lipomas approached by deep learning methods. Orphanet J. Rare Dis. 2022, 17, 158. [Google Scholar] [CrossRef] [PubMed]
  15. Xie, Y.; Zaccagna, F.; Rundo, L.; Testa, C.; Agati, R.; Lodi, R.; Manners, D.N.; Tonon, C. Convolutional Neural Network Techniques for Brain Tumor Classification (from 2015 to 2022): Review, Challenges, and Future Perspectives. Diagnostics 2022, 12, 1850. [Google Scholar] [CrossRef]
  16. Zoetmulder, R.; Bruggeman, A.A.E.; Išgum, I.; Gavves, E.; Majoie, C.B.L.M.; Beenen, L.F.M.; Dippel, D.W.J.; Boodt, N.; Hartog Sanne, J.; Doormaal, P.J.; et al. Deep-Learning-Based Thrombus Localization and Segmentation in Patients with Posterior Circulation Stroke. Diagnostics 2022, 12, 1400. [Google Scholar] [CrossRef]
  17. Doi, K. Computer-aided diagnosis in medical imaging: Historical review, current status and future potential. Comput. Med. Imaging Graph. 2007, 31, 198–211. [Google Scholar] [CrossRef] [Green Version]
  18. Wang, J.; Tan, D.; Liu, J.; Wu, J.; Huang, F.; Xiong, H.; Luo, T.; Chen, S.; Li, Y. Merging Multiphase CTA Images and Training Them Simultaneously with a Deep Learning Algorithm Could Improve the Efficacy of AI Models for Lateral Circulation Assessment in Ischemic Stroke. Diagnostics 2022, 12, 1562. [Google Scholar] [CrossRef]
  19. Nowinski, W.L.; Gupta, V.; Qian, G.; He, J.; Poh, L.E.; Ambrosius, W.; Chrzan, R.M.; Polonara, G.; Mazzoni, C.; Mol, M.; et al. Automatic Detection, Localization, and Volume Estimation of Ischemic Infarcts in Noncontrast Computed Tomographic Scans. Investig. Radiol. 2013, 48, 661–670. [Google Scholar] [CrossRef]
  20. Wardlaw, J.M.; Mielke, O. Early Signs of Brain Infarction at CT: Observer Reliability and Outcome after Thrombolytic Treatment—Systematic Review. Radiology 2005, 235, 444–453. [Google Scholar] [CrossRef]
  21. Chalela, J.A.; Kidwell, C.S.; Nentwich, L.M.; Luby, M.; Butman, J.A.; Demchuk, A.M.; Hill, M.D.; Patronas, N.; Latour, L.; Warach, S. Magnetic resonance imaging and computed tomography in emergency assessment of patients with suspected acute stroke: A prospective comparison. Lancet 2007, 369, 293–298. [Google Scholar] [CrossRef] [Green Version]
  22. Lansberg, M.G.; Albers, G.W.; Beaulieu, C.; Marks, M.P. Comparison of diffusion-weighted MRI and CT in acute stroke. Neurology 2000, 54, 1557–1561. [Google Scholar] [CrossRef] [PubMed]
  23. Marbun, J.T.; Seniman; Andayani, U.J. Classification of stroke disease using convolutional neural network. Phys. Conf. Ser. 2018, 978, 012092. [Google Scholar] [CrossRef]
  24. Kuang, H.; Menon, B.K.; Qiu, W. Semi-automated infarct segmentation from follow-up noncontrast CT scans in patients with acute ischemic stroke. Med. Phys. 2019, 46, 4037–4045. [Google Scholar] [CrossRef] [Green Version]
  25. Luo, C.; Zhang, J.; Chen, X.; Tang, Y.; Weng, X.; Xu, F. UCATR: Based on CNN and Transformer Encoding and Cross-Attention Decoding for Lesion Segmentation of Acute Ischemic Stroke in Non-contrast Computed Tomography Images. In Proceedings of the 43rd Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), Guadalajara, Mexico, 31 October–4 November 2021; pp. 3565–3568. [Google Scholar]
  26. Maini, R.; Aggarwal, H. A Comprehensive Review of Image Enhancement Techniques. J. Comput. 2010, 2, 8–13. [Google Scholar]
  27. Hong, L.; Wan, Y.; Jain, A. Fingerprint image enhancement: Algorithm and performance evaluation. IEEE Trans. Pattern Anal. Mach. Intell. 1998, 20, 777–789. [Google Scholar] [CrossRef] [Green Version]
  28. Dhal, K.G.; Das, A.; Ray, S.; Gálvez, J.; Das, S. Histogram Equalization Variants as Optimization Problems: A Review. Arch. Comput. Methods Eng. 2021, 28, 1471–1496. [Google Scholar] [CrossRef]
  29. Pizer, S.M.; Amburn, E.P.; Austin, J.D.; Cromartie, R.; Geselowitz, A.; Greer, T.; Romeny, B.H.; Zimmerman, J.B.; Zuiderveld, K. Adaptive histogram equalization and its variations. Comput. Vis. Graph. Image Process. 1987, 39, 355–368. [Google Scholar] [CrossRef]
  30. Reza, A.M. Realization of the Contrast Limited Adaptive Histogram Equalization (CLAHE) for Real-Time Image Enhancement. J. VLSI Signal Process. Syst. Signal Image Video Technol. 2004, 38, 5–44. [Google Scholar] [CrossRef]
  31. Srinivasan, S.; Balram, N. Adaptive Contrast Enhancement Using Local Region Stretching. In Proceedings of the 9th Asian Symposium on Information Display, New Delhi, India, 8–12 October 2006; pp. 152–155. [Google Scholar]
  32. Wang, W.; Yuan, X.; Chen, Z.; Wu, X.J.; Gao, Z. Weak-Light Image Enhancement Method Based on Adaptive Local Gamma Transform and Color Compensation. J. Sens. 2021, 2021, 5563698. [Google Scholar] [CrossRef]
  33. Zhang, L.; Wang, J.; Wu, W. Image enhancement algorithm based on improved K-means clustering. J. Appl. Opt. 2016, 37, 549–554. [Google Scholar]
  34. Girshick, R. Fast R-CNN. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
  35. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  36. Kim, J.-H.; Kim, N.; Park, Y.W.; Won, C.S. Object Detection and Classification Based on YOLO-V5 with Improved Maritime Dataset. J. Mar. Sci. Eng. 2022, 10, 377. [Google Scholar] [CrossRef]
  37. Mekhalfi, M.L.; Nicolò, C.; Bazi, Y.; Rahhal, M.M.A. Contrasting YOLOv5, Transformer, and EfficientDet Detectors for Crop Circle Detection in Desert. IEEE Geosci. Remote Sens. Lett. 2021, 19, 3003205. [Google Scholar] [CrossRef]
  38. Du, J. Understanding of Object Detection Based on CNN Family and YOLO. J. Phys. Conf. Ser. 2018, 1004, 012029. [Google Scholar] [CrossRef]
  39. Nguyen, D.T.; Nguyen, T.N.; Kim, H.; Lee, H.-J. A High-Throughput and Power-Efficient FPGA Implementation of YOLO CNN for Object Detection. IEEE Tran.s Very Large Scale Integr. VLSI Syst. 2019, 27, 1861–1873. [Google Scholar] [CrossRef]
  40. Huang, Z.; Wang, J.; Fu, X.; Yu, T.; Guo, Y.; Wang, R. DC-SPP-YOLO: Dense connection and spatial pyramid pooling based YOLO for object detection. Inf. Sci. 2020, 522, 241–258. [Google Scholar] [CrossRef] [Green Version]
  41. Chiriboga, M.; Green, C.M.; Hastman, D.A.; Mathur, D.; Wei, Q.; Díaz, S.A.; Medintz, I.L.; Veneziano, R. Rapid DNA origami nanostructure detection and classification using the YOLOv5 deep convolutional neural network. Sci. Rep. 2022, 12, 3871. [Google Scholar] [CrossRef]
  42. Yao, J.; Qi, J.; Zhang, J.; Shao, H.; Yang, J.; Li, X. A Real-Time Detection Algorithm for Kiwifruit Defects Based on YOLOv5. Electronics 2021, 10, 1711. [Google Scholar] [CrossRef]
  43. Velayudhan, V.; Pawha, P.; Tanenbaum, L.N.; Patel, A.B. Radiology. Stroke Imaging. 2018. Available online: https://emedicine.medscape.com/article/338385-overview (accessed on 30 November 2018).
  44. Tehreem. Ischemic Stroke. 2019. Available online: https://medizzy.com/feed/14746660 (accessed on 30 November 2019).
  45. Warren, D.J.; Musson, R.; Connolly, D.J.A.; Griffiths, P.D.; Hoggard, N. Imaging in acute ischaemic stroke: Essential for modern stroke care. Postgrad. Med. J. 2010, 86, 409–418. [Google Scholar] [CrossRef] [Green Version]
  46. Alsaddi, Z.S. Ischemic Stroke. 2021. Available online: https://neuropedia.net/ischemic-stroke/articles/neurology/cerebrovascular/ (accessed on 8 February 2021).
  47. Caprio Fan, Z.; Sorond, F.A. Cerebrovascular disease: Primary and secondary stroke prevention. Med. Clin. 2019, 103, 295–308. [Google Scholar] [CrossRef] [Green Version]
  48. Fonarow, G.C.; Smith, E.E.; Saver, J.L.; Reeves, M.J.; Bhatt, D.L.; Grau-Sepulveda, M.V.; Olson, D.M.; Hernandez, A.F.; Peterson, E.D.; Schwamm, L.H. Timeliness of tissue-type plasminogen activator therapy in acute ischemic stroke: Patient characteristics, hospital factors, and outcomes associated with door-to-needle times within 60 minutes. Circulation 2011, 123, 750–758. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  49. Fang, Y.; Guo, X.; Chen, K.; Zhou, Z.; Ye, Q. Accurate and Automated Detection of Surface Knots on Sawn Timbers Using YOLO-V5 Model. BioResources 2021, 16, 5390–5406. [Google Scholar] [CrossRef]
  50. Wu, W.; Liu, H.; Li, L.; Long, Y.; Wang, X.; Wang, Z.; Li, J.; Chang, Y. Application of local fully Convolutional Neural Network combined with YOLO v5 algorithm in small target detection of remote sensing image. PLoS ONE 2021, 16, e0259283. [Google Scholar] [CrossRef] [PubMed]
  51. Paul, M.; Michael, G.D.L.; Grant, M. Computational Image Analysis of Nonenhanced Computed Tomography for Acute Ischaemic Stroke: A Systematic Review. J. Stroke Cerebrovasc. Dis. 2020, 29, 104715. [Google Scholar]
Figure 1. The architecture of YOLOv5.
Figure 1. The architecture of YOLOv5.
Diagnostics 12 02591 g001
Figure 2. The structure of AC-YOLOv5.
Figure 2. The structure of AC-YOLOv5.
Diagnostics 12 02591 g002
Figure 3. The results of AC-YOLOv5 model training.
Figure 3. The results of AC-YOLOv5 model training.
Diagnostics 12 02591 g003
Table 1. The experimental results on T1 and T2.
Table 1. The experimental results on T1 and T2.
T1T2
AccuracyRecallAccuracyRecall
F-RCNN67.80%54.30%2.00%4.10%
YOLOv583.10%67.50%55.10%40.40%
K-YOLOv534.30%22.00%7.60%7.61%
G-YOLOv586.30%69.00%57.10%42.20%
H-YOLOv587.10%65.00%77.60%65.10%
AH-YOLOv598.70%76.10%74.90%69.60%
CL-YOLOv587.00%71.40%57.70%65.20%
AC-YOLOv594.10%85.30%91.70%88.60%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhang, L.; Cui, H.; Hu, A.; Li, J.; Tang, Y.; Welsch, R.E. An Improved Detection Algorithm for Ischemic Stroke NCCT Based on YOLOv5. Diagnostics 2022, 12, 2591. https://doi.org/10.3390/diagnostics12112591

AMA Style

Zhang L, Cui H, Hu A, Li J, Tang Y, Welsch RE. An Improved Detection Algorithm for Ischemic Stroke NCCT Based on YOLOv5. Diagnostics. 2022; 12(11):2591. https://doi.org/10.3390/diagnostics12112591

Chicago/Turabian Style

Zhang, Lifeng, Hongyan Cui, Anming Hu, Jiadong Li, Yidi Tang, and Roy Elmer Welsch. 2022. "An Improved Detection Algorithm for Ischemic Stroke NCCT Based on YOLOv5" Diagnostics 12, no. 11: 2591. https://doi.org/10.3390/diagnostics12112591

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop