Next Article in Journal
A Comprehensive Review of Computer-Aided Models for Breast Cancer Diagnosis Using Histopathology Images
Next Article in Special Issue
Enhancing the Super-Resolution of Medical Images: Introducing the Deep Residual Feature Distillation Channel Attention Network for Optimized Performance and Efficiency
Previous Article in Journal
Utilizing 3D Printing Technology to Create Prosthetic Irises: Proof of Concept and Workflow
Previous Article in Special Issue
CLRD: Collaborative Learning for Retinopathy Detection Using Fundus Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Detection of Image Artifacts Using Improved Cascade Region-Based CNN for Quality Assessment of Endoscopic Images

1
School of Biomedical Engineering (Suzhou), Division of Life Sciences and Medicine, University of Science and Technology of China, Hefei 230026, China
2
Suzhou Institute of Biomedical Engineering and Technology, Chinese Academy of Sciences, Suzhou 215163, China
3
Department of Gastroenterology, Zhongda Hospital, School of Medicine, Southeast University, 87 Dingjiaqiao Road, Nanjing 210009, China
*
Author to whom correspondence should be addressed.
Bioengineering 2023, 10(11), 1288; https://doi.org/10.3390/bioengineering10111288
Submission received: 28 August 2023 / Revised: 29 October 2023 / Accepted: 1 November 2023 / Published: 6 November 2023
(This article belongs to the Special Issue Artificial Intelligence (AI) for Medical Image Processing)

Abstract

:
Endoscopy is a commonly used clinical method for gastrointestinal disorders. However, the complexity of the gastrointestinal environment can lead to artifacts. Consequently, the artifacts affect the visual perception of images captured during endoscopic examinations. Existing methods to assess image quality with no reference display limitations: some are artifact-specific, while others are poorly interpretable. This study presents an improved cascade region-based convolutional neural network (CNN) for detecting gastrointestinal artifacts to quantitatively assess the quality of endoscopic images. This method detects eight artifacts in endoscopic images and provides their localization, classification, and confidence scores; these scores represent image quality assessment results. The artifact detection component of this method enhances the feature pyramid structure, incorporates the channel attention mechanism into the feature extraction process, and combines shallow and deep features to improve the utilization of spatial information. The detection results are further used for image quality assessment. Experimental results using white light imaging, narrow-band imaging, and iodine-stained images demonstrate that the proposed artifact detection method achieved the highest average precision (62.4% at a 50% IOU threshold). Compared to the typical networks, the accuracy of this algorithm is improved. Furthermore, three clinicians validated that the proposed image quality assessment method based on the object detection of endoscopy artifacts achieves a correlation coefficient of 60.71%.

1. Introduction

An endoscope—the most direct examination device for gastrointestinal diseases—is introduced into the human body through natural cavities. This has significant advantages in both diagnosis and treatment, and it serves as a primary means for subsequent minimally invasive surgery and noninvasive treatment [1]. The collected endoscopic images are used to determine the patient’s condition and formulate subsequent treatment plans. However, endoscopic images contain numerous interfering factors, such as mirror reflection, motion blur, bubbles, etc., which can affect the visual interpretation of endoscopic examinations, affect the observation and diagnosis of clinicians regarding the lesion area, and negatively affect computer-aided diagnosis (CAD) processes [2]. In addition, the presence of interference is an important reference criterion for the assessment of image quality in gastrointestinal endoscopy images, contributing to the quality evaluation of clinical endoscopic examination procedures [3].
Image quality assessment aims to simulate human perception and is typically performed by human observers who evaluate algorithm values against subjective ratings. A typical approach involves comparing the distortion metric with an ideal imaging model or a perfect reference image [4]. Depending on the amount of information provided by the original reference image, it can be classified as full, reduced, or no-reference (also known as blind image quality assessment, BIQA) [5]. No-reference image quality assessment (NR-IQA) is one of the most challenging problems in image quality assessment because it solely relies on distorted images, and thus has garnered significant research interest in recent years.
NR-IQA has a significant practical value and is widely used in numerous practical applications where reference images are unavailable. Mittal et al. proposed a blind image spatial quality assessment method called the blind/referenceless image spatial quality evaluator (BRISQUE) [6]. This method uses locally normalized luminance coefficients to quantify the “naturalness” loss caused by distortion; it exhibits a low computational complexity, making it suitable for real-time applications. Liu et al. introduced RankIQA [7], which generates ordered degraded images to train a Siamese network for relative quality ranking. The trained network is then transferred to a traditional CNN, enabling it to estimate the absolute image quality from a single image. Junjie et al. proposed a multiscale image quality transformer called MUSIQ [8], which can handle full-resolution image inputs with different resolutions, sizes, and aspect ratios; capture image quality at various granularities; and perform well on multiple large-scale IQA datasets. Talebi et al. presented a neural network based on deep object recognition called NIMA [9], which can predict the distribution of human evaluations of images in terms of both perceptual quality (from a technical standpoint) and attractiveness (from an aesthetic standpoint). The proposed neural network exhibited scores that closely resembled the human subjective ratings, making it suitable for natural image quality assessment tasks. With the profound influence of machine learning, particularly deep learning, in various fields, image quality assessment in endoscopy is undergoing continuous innovation. Alexander et al. developed a new fidelity score for quantitative image quality assessment based on the structural similarity maps adopted in the human visual system (HVS), where the measure indicated the extent to which the structural information of relevant structures was preserved in the panorama [10]. Aubreville et al. [11] proposed an improved version of the Inception V3 network for detecting motion artifacts in endoscopic images. Kamen et al. [12] selected high-quality and information-rich images by calculating the image entropy. Outtas et al. [13] studied the usability of two general algorithms based on natural image quality assessments (NIQE [14] and BRISQUE [6]) in medical image environments. Prerna et al. [15] proposed the use of gray-level co-occurrence matrices for image quality assessment. Zhang et al. [16] performed a no-reference quality assessment of capsule endoscopy images by calculating the gradient field using the Sobel operator.
Considering the limitations of current no-reference endoscopic image quality assessment methods, which primarily focus on evaluating individual artifacts and often yield less interpretable results, we propose a framework for detecting endoscopic image artifacts and assessing image quality based on an improved cascade R-CNN designed to overcome these challenges. The architecture of the framework is shown in Figure 1. First, the original endoscopic image is obtained. After undergoing image preprocessing, the image is fed into the improved cascade region-based CNN for artifact detection. Subsequently, after extracting the detection result data, both location and area information are acquired and transmitted to the endoscopic image NR-IQA method. This process culminates in the generation of interpretable image quality assessment results.
The artifact detection method retains the multistage structure of the cascade R-CNN, with ResNeXt101 serving as the backbone network. We improved the original feature pyramid network (FPN) structure and introduced Generalized Intersection over Union (GIoU) loss as a new evaluation metric loss function, replacing the original IoU metric. Using the improved model, we successfully located and identified artifacts in the gastrointestinal tract. Additionally, we propose a method that combines multiple weights to calculate the image quality score based on artifact detection results.
The main contributions of this study are as follows:
(1) Nonvideo sequences of gastrointestinal endoscopy images from multiple clinical hospitals were used. The images have high content and modality diversity, including images from white light imaging (WLI), narrow-band imaging (NBI), and iodine staining, making them robust and relevant for general applications in gastrointestinal endoscopy.
(2) An improved feature pyramid network that incorporates channel attention mechanisms into the feature extraction process is proposed. Shallow and deep features were fused by introducing an additional channel from the shallow to deep layers, thereby enhancing the utilization of spatial information in the shallow layers. This not only strengthened the feature extraction of semantic and positional information through path aggregation but also established global spatial feature attention for the mapping and representation of artifact features across multiple branches in the image.
(3) The proposed endoscopic image quality assessment method successfully detects anomalies in small targets and handles multiple interferences. Moreover, it uses the detection results of high-quality artifact targets to provide numerical scores that are close to expert ratings, thereby demonstrating strong interpretability.
Overall, these contributions enhance the detection and assessment of endoscopic image artifacts and provide valuable insights for clinical multicenter endoscopic image quality assessment.

2. Materials

2.1. Data Acquisition

The endoscopic image data used in this study were obtained from 12 medical institutions in Jiangsu Province, with Zhongda Hospital Southeast University serving as the central hospital. In total, 2303 images containing artifacts were collected through data collection and clinicians’ annotations. The dataset encompassed images of various tissues and organs, including the esophagus and stomach, as well as multimodal endoscopic images (including WLI, NBI, and iodine-stained images), as shown in Table 1. The color space of the images was distributed as RGB, and eight types of interferences were present: blur, bubbles, specularity, saturation, contrast, instrument, blood, and artifacts, as listed in Table 2.
Specularity is caused by the reflection of the tube owing to the smooth surface of the electron endoscope; saturation and contrast are derived from light exposure and changes in videoscope topology; blur results from erratic hand movements and small movements of the digestive tract lining; and bubbles and blood result from changes in the digestive tract environment leading to the presence of fluids. This instrument, however, prevents the occlusion of the inner wall of the digestive tract. These artifacts originate from unpredictable imaging conditions, and they are present in more than 60% of endoscopic video frames [5].
For the image quality assessment, 100 images were randomly selected from the 2303 images containing artifacts and subjectively scored based on the evaluation criteria shown in Table 3.

2.2. Data Preprocessing

Significant differences exist between the sizes and types of endoscopic images captured by various models of digestive endoscopic instruments from various manufacturers. These images also contain nonendoscopic information such as instrument details and acquisition time. To mitigate the impact of these factors on the image quality assessment, this study employed dilation and erosion morphological processing methods to generate effective area masks. These masks were used to extract the relevant image regions, thereby enhancing the effectiveness of the endoscopic image quality assessment. The image preprocessing operation is illustrated in Figure 2.
CLOSE ( X ) = E ( D ( X ) )

2.3. Data Analysis

Upon analyzing the surface characteristics of the samples, substantial within-class variations and interclass correlations were observed among the artifact samples. For example, specularity and bubbles exhibit similarities in certain textural representations; however, specularity, bubbles, and artifacts follow distinct internal morphological patterns. The annotations were subjected to statistical analysis and the results are shown in Figure 3. Among the eight defined defect categories, the maximum difference in quantity approaches 64.5:1. Owing to the sample characteristics, the detection task encompassed highly imbalanced categories, extreme proportions, and small sample sizes.

2.4. Data Augmentation

The dataset used in this study was obtained from various devices across multiple hospitals, resulting in significant image resolution variations, such as 512 × 512 , 768 × 576 , 1081 × 808 , 1349 × 1079 , and 1920 × 1080 . To facilitate the algorithmic processing, this section standardizes the image size to 800 × 800 .
In addition to the dataset size and labels, certain characteristics of images and labels render the noise detection model less effective. Insufficiently qualified images with accurate annotations and imbalanced class distributions are common factors that lead to overfitting and limited generalization in neural network training. If the sample size for any of these eight types of defects is inadequate or significantly different, the trained model may exhibit a significant bias.
To address this issue, data augmentation techniques were employed in this study to alleviate the sample imbalance. Data augmentation involves the generation of variations in the original images, thereby increasing the training data volume without introducing additional images. Four types of geometric transformations are applied during the data augmentation process: mosaics [17], rescaling, inversion, and rotation. The results of the mosaic data enhancement are shown in Figure 4.

3. Methods

The overall approach of this method is divided into two main parts: artifact target detection using the improved cascade R-CNN model and endoscopic image quality assessment based on the detection results.

3.1. Cascade R-CNN

The cascade R-CNN [18] is an extension of the Faster R-CNN that aims to incorporate more semantic information into the detection task [19]. Unlike the traditional Faster R-CNN, the cascaded R-CNN employs a cascaded modular structure, enabling additional context and features to be extracted through advanced feature extraction steps. These features were then utilized in the ROI pooling layer to enhance the network performance by providing richer representations. In addition, the cascaded structure provides more supervised signals during training, leading to more accurate models.
The cascade R-CNN consists of a feature extraction network, an FPN, a region proposal network layer, and cascade detectors. The feature extraction network ResNeXt101 was used to extract image features. The original map is obtained by convolution operations, Conv1, Conv2, Conv3, Conv4, and Conv5, and feature fusion at different levels to obtain feature maps P2, P3, P4, and P5 at different scales. These feature maps were then input into the region proposal network to obtain the candidate target areas. Subsequently, ROI alignment operation was performed on the resulting candidate target areas to obtain an ROI feature map.
ResNeXt101 is a deep CNN built on the ResNet architecture [20]. This introduces the concept of ”group convolution,” where multiple parallel convolutional branches with the same structure are employed. Each branch processes different input characteristics, enabling an increased network width without overfitting. The ResNeXt101 network incorporates techniques, such as batch normalization and residual connections, to enhance its performance. Residual structures have been widely employed in the field of medical imaging to eliminate dependencies between the network’s weak and high levels [21]. The residual structure is shown in Figure 5.
In this study, an improved cascade R-CNN detection model is proposed, utilizing ResNeXt101 as the backbone network to locate and identify artifacts in the digestive tract. The original evaluation index, namely Intersection over Union (IoU), in the cascade R-CNN was replaced with GIoU loss as a new evaluation index loss function. IoU measures the degree of overlap between a predicted bounding box and a ground truth bounding box. GIoU is an improved version of IoU. It not only considers the intersection and union areas but also takes into account the spatial relationship between the two bounding boxes, which makes it more suitable for complex object shapes and rotations. This enhancement is aimed at making it more robust to variations in object rotation and shape. Figure 6 illustrates the enhanced FPN with added attention mechanisms that enable better learning capabilities.
The advantages of this model are as follows:
(1) It provides an effective feature extraction structure that maximizes the utilization of shallow feature information and enhances the detection of small targets.
(2) The incorporation of the channel attention mechanism captures the feature dependencies between different channel maps in the feature extraction network. This reduces the missed detection rate and leads to more reliable results.
(3) GIoU was employed as the new evaluation index loss function, replacing IoU, the original evaluation index in cascade R-CNN. This ensures scale invariance in the loss function target detection frame regression and maintains consistency between the optimization objective and loss function of the detection frame.

3.2. Improved Feature Pyramid Structure

Each layer of the feature pyramid employs a distinct convolution kernel size to extract the features. The lower layers capture large-scale features, such as edges, whereas the final layer captures fine details. These feature maps are then concatenated to create a comprehensive feature map for classification. The key advantage of a feature pyramid is its ability to capture features at various scales in an image, resulting in enhanced classification performance. This is particularly effective for handling images of different sizes by encompassing the features of diverse scales.
Although this approach exhibits a high accuracy in classifying and localizing larger objects, it faces challenges in accurately detecting smaller objects. This is due to the nature of the deep CNN, where extensive convolution and pooling operations result in expanded receptive fields and decreased resolutions in the network feature layers. Consequently, there is the risk of overlooking small targets.
In contrast, low-level features obtained from shallow neural networks have a higher resolution and contain more information, making them valuable for detecting small objects. By fusing features at different scales, the recognition accuracy for small targets can be improved while maintaining the accuracy for larger targets.
The feature enhancement method was employed in this study to fully use shallow feature information and increase the resolution of the feature maps for small targets. Deep features, which contain rich semantic information, were added elementwise to shallow features through bilinear interpolation. However, an accurate localization of shallow to deep features has become increasingly challenging. To address this issue, an additional pathway, called the pathway enhancement channel, was introduced. The number of convolution layers traversed by the information flow from deep to shallow layers is reduced by this pathway, enabling the propagation of shallow information to deep layers and enhancing the localization of deep positional information. The improved structure of the FPN is shown in Figure 7.
The original convolutional output layers in the bottom-up pathway are C 2 , C 3 , C 4 , and C 5 . First, convolution was applied to the input image, followed by dimensionality reduction of C 3 , C 4 , and C 5 . After applying the attention mechanism to C 5 , it underwent a two-fold upsampling to match the size of C 4 . Subsequently, the corresponding elements were added, and the obtained result was input into C 4 . The same process was used to obtain C 3 and C 2 . C 3 , C 4 , and C 5 were upsampled by 2x, 4x, and 8x, respectively, and added to the shallow feature maps using bilinear interpolation. Corresponding elements were added to increase the utilization of deep features for shallow features. The formula for bilinear interpolation f ( x , y ) is as follows:
f ( x , y ) = f x 1 , y 1 x 2 x 1 y 2 y 1 x 2 x y 2 y + f x 2 , y 1 x 2 x 1 y 2 y 1 x x 1 y 2 y + f x 1 , y 2 x 2 x 1 y 2 y 1 x 2 x y y 1 + f x 2 , y 2 x 2 x 1 y 2 y 1 x x 1 y y 1 ,
where ( x 1 , y 1 ) , ( x 1 , y 2 ) , ( x 2 , y 1 ) , ( x 2 , y 2 ) are four known points, and the characteristics of the P 0 -layer were obtained after bilinear interpolation as follows:
P 0 = C 2 + C 3 f ( x , y ) + C 4 f ( x , y ) + C 5 f ( x , y )
The final generated feature layers were P 2 , P 3 , P 4 , and P 5 , which fully utilized the deep and shallow features by fusing these different layers, resulting in improved prediction performance.
Digestive endoscopy images are complex, and the presence of artifacts makes it difficult to distinguish between true targets. To enhance the expressive power of the features of images, a channel attention mechanism was employed in the FPN object detection model. For feature-mapping layers of different scales, the channel attention mechanism was used to capture the feature dependencies between different channel maps. This involves calculating the weighted feature vectors that represent the explicit correlation between the feature channels, as shown in Figure 8.

3.3. Loss Function

We introduced GIoU [22] as a new evaluation index loss function to replace the original evaluation index, IoU, in the cascade R-CNN. The GIoU formula ensures that the loss function for the target detection box regression is scale-invariant and maintains consistency between the optimization objective and the loss function. In the context of artifact identification in digestive endoscopy, there is a significant imbalance between positive and negative samples, which makes the training of the bounding box scores more challenging. Metrics based on L1 and L2 norms may yield significantly different IoU values for the two bounding boxes at the same distance. As shown in Figure 9, the example with the bounding boxes is represented by two corners. In all three cases, the distance between the representation of the two rectangles is the same for L2 norm; however, the IoU and GIoU values are very different. Furthermore, when IoU is the same, it only indicates that the Intersection over Union of the target box and the detection box is identical, but the actual size of the predicted box may be completely different. This highlights that neither IoU nor the L2 norm can adequately reflect the detection performance. Therefore, such loss functions are not ideal for predicting bounding boxes. Unlike IoU, which focuses only on the overlapping area, GIoU considers both overlapping and nonoverlapping regions, providing a better reflection of their intersection.
The introduction of the GIoU loss function has significant implications in the proposed method; it overcomes the limitations associated with traditional loss functions in the context of artifact detection. This loss function promotes a more precise bounding box regression and enhances the model’s ability to detect artifacts of varying sizes and shapes. This improvement augments the depth and accuracy of our research, thereby contributing to a more reliable analysis of endoscopic images, as shown in Figure 10.
The GIoU formula is:
G I o U = I o U A c U A c ,
where U is the union of the prediction box and GT box, and A c is the minimum closure of the prediction box and real box.

3.4. Image Quality Score

Reading poor-quality medical images can significantly affect the efficiency of the diagnostic work. Therefore, based on the artifact detection results discussed in the previous section, we proposed a method to assess the quality of endoscopic images. This method aims to provide a visual indication of the effectiveness of an image. The image quality score (QS) is based on the (a) type, (b) region, (c) and location of the detected artifacts, and (d) the confidence of the detected artifacts. We improved the method based on Ali [3], where weights are assigned to each category, and the average weight is computed as the quality score.
Class weights ( w C ) : artifact (0.50), specularity (0.10), saturation (0.10), blur (0.20), contrast (0.10), bubbles (0.02), instrument (0.02), and blood (0.02).
Area weight ( w A ) : percentage of the total image area occupied by all detected artifact and normal areas.
Location weight ( w L ) : center (0.5), left (0.25), right (0.25), top (0.25), bottom (0.25), top left (0.125), top right (0.125), bottom left (0.125), and bottom right (0.125).
The confidence weight ( w C o ) represents the confidence level of the detection results.
Q S = 1 B λ A W C W C o W A + λ L W C W C o W L 0
To align with the numerical range of subjective ratings by experts and present the image assessment results in a more intuitive manner, this study mapped the QS values to a range of 0–5. The mapped results are presented in Figure 11.

4. Results and Discussion

4.1. Evaluation of Object Detection Results

The experimental setup in this study was as follows: Ubuntu 18.04 operating system, GeForce RTX 2060 graphics card, and Intel(R) Core(TM) i9-10940X processor. The improved cascade R-CNN network model proposed in this study was implemented within the mm detection framework version 2.25.0. The selected optimizer is a stochastic gradient descent with an initial momentum of 0.9. The total number of iterations was 5800. The training time for the first complete epoch of the model was 453 s, which gradually decreased over time. The total training time for 50 epochs was 18,537 s, 5.15 h.
This study employed commonly accepted evaluation metrics for object detection methods. Using a specific IoU threshold, true positives (TP), true negatives (TN), false positives (FP), and false negatives (FN) were defined as the classification outcomes. Furthermore, “all detections” refers to the total number of predicted bounding boxes, while “all ground truths” represents the total number of GT annotations.
Precision = T P T P + F P = T P all detections
Recall = T P T P + F N = T P all GTs
Average precision (AP) is a commonly used metric in object detection that evaluates model performance by calculating the area under the precision–recall (PR) curve. This reflected the AP at each unique recall level. AP is defined based on PASCAL VOC 2010 [23]:
A P = r = i n 1 r i + 1 r i p interp r i + 1
Through comparative experiments, we compared the performance of the proposed method with that of other networks in our dataset. From the AP values shown in Table 4 and Table 5, it can be seen that the proposed method achieved the best performance in all typical networks.
Through the implementation of ablative experiments, our primary objective was to investigate the performance characteristics of the network components. While keeping the other experimental conditions constant, we conducted separate experiments on the improved FPN and GIoU components using the proposed method. Based on the AP in Table 6, we observed that these components yielded improvements in the original model, thereby demonstrating the assisting role of the network components in learning pseudo-shadow features.
The confusion matrix is an N × N matrix used to evaluate the performance of a classification model, where N represents the number of target categories. The matrix compares the actual and predicted class labels and assesses the predictive performance of the model. By analyzing the entire dataset, one can determine the number of correctly and incorrectly predicted samples, thus measuring the model’s predictive effectiveness. The confusion matrix, as presented in Figure 12, illustrates the performance for each class, with all classes except “blood” exhibiting performance exceeding 50%. This finding substantiates the existence of significant differences in the learned features among various classes. However, the “blood” class displays a slightly lower performance owing to the diverse morphological characteristics of blood within the gastrointestinal tract lacking distinct typical features.
The test results using the proposed method are presented in Table 7, using images from WLI, NBI, and iodine-staining modalities as examples. The table shows that the accurate detection of most defects was achieved in these three types of images. The detected bounding boxes effectively encompassed the abnormal regions without redundancy.
Figure 13 depicts the region-based statistical distributions of the GT (left) and predicted results (right) in the test dataset, arranged based on the region size. We observed that the predicted results exhibited a similar trend to those of the GT, indicating stable and reliable detection. This observation was further supported by specific evaluation metrics. In addition, the figure shows magnified views of the detection results for small targets, demonstrating the network’s ability to accurately capture defects with lengths smaller than 30 pixels.

4.2. Evaluation of the Quality Score

This study used a set of 100 randomly selected images with different quality ratings to evaluate the proposed quality assessment scheme based on the quality rankings provided by the three experts. Endoscopic images are different from natural images, and clinicians will randomly select the images of interest to shoot when performing image acquisition. There is no excellent reference image without distortion; therefore, the image quality assessment method without reference is used for image comparison. Four no-reference image quality assessment methods, NIQE [13], BRISQUE [6], MUSIQ [8], and NIMA [9], were compared with the method proposed in this paper.
Because the numerical values of the image evaluation scores represent only the quality levels of the images and lack practical significance, we employed the Spearman correlation coefficient to assess the monotonic relationship between the two variables. A larger value indicated a stronger correlation. Correlations were calculated using the following formula:
ρ S = i = 1 N R i R ¯ S i S ¯ i = 1 N R i R ¯ 2 i = 1 N S i S ¯ 2 1 2
In Equation (9), N is the total number of observations, R i and S i are the grades of the observed value i. R ¯ and S ¯ are the average grades of the variables x and y, respectively.
A p-value test was used to determine the probability of the observed outcome occurring randomly.
Table 8 shows that the proposed quality score (QS) method has an average Spearman correlation coefficient of 60.71% with the rankings provided by the three experts. High correlation coefficients indicate a significant relationship between the proposed method and the quality scores suggested by the experts.
Figure 14 presents four example images displaying the scores given by the three experts, the average score, and QS computed using the proposed method. The QS scores of the images of different qualities are close to the expert scores. Figure 15 illustrates the correlation between the scores determined by the three experts and the proposed method.

4.3. Discussion

The evaluation scores of the proposed quality assessment method for endoscopic images had a strong correlation with the expert ratings. This correlation was based on the category, confidence, area, and position information extracted from the artifact detection results. These findings indicate that the quality scores have a practical significance. The poor performance of existing no-reference image quality assessment methods, such as NIQE, BRISQUE, MUSIQ, and MANI, in evaluating digestive endoscopic images may be because these methods are primarily designed for natural images. This highlights the requirement for more specific approaches to evaluate the quality of digestive endoscopy images, considering their unique characteristics and environment. Owing to the tight integration of the proposed image quality assessment method with the object detection results, a strong interpretability was achieved, compared to the four methods mentioned above. In clinical applications, clinicians can rely on the quality scores of digestive endoscopic images as a primary quality control measure, thereby reducing the time required for the manual inspection of patient images.
The proposed object detection method is more sensitive to contrast, blur, instruments, and specularity, owing to the availability of sufficient data and features. However, the detection of blood objects showed a slightly lower performance owing to the limited training data and diverse morphologies of blood presentations. Future studies should focus on improving the detection of blood-related issues. Gastrointestinal endoscopy images encompass various modalities, and the proposed object detection method was validated on WLI, NBI, and iodine-stained modalities based on the imaging modalities offered by Olympus and Fujifilm instruments. However, because of the lack of data for linked color imaging and blue laser imaging modalities, the detection performance of these modalities requires further verification.
Based on the research presented in this paper, future work should focus on image restoration for images containing artifacts, building upon the foundation of object detection, which has a clinical significance. In future work, we will consider conducting comparative experiments for other versions of the YOLO series models. To enhance the robustness and generalizability of the conclusions, it is proposed to replicate the study in diverse geographical regions around the world. Such multilocation replication endeavors are necessary to validate the applicability and reproducibility of the methods and results across different contexts and populations. Validating this research in other parts of the world will significantly enhance the credibility of the research and provide solid support for clinical practice and applications.

5. Conclusions

This study proposed an optimized method for detecting endoscopy artifacts that provides intuitive information regarding the location and category of eight types of artifacts and numerically indicates image quality. The average precision (AP) results of object detection demonstrate significant improvements, compared to traditional object detection methods. The results of the correlation analysis indicate that the proposed method achieved better performance in the evaluation of gastrointestinal endoscopic image quality.
The results also revealed that an improved FPN structure is employed in the cascade R-CNN model. The improved FPN structure integrated the features from different layers, by effectively using both deep and shallow features. A channel attention mechanism was used to capture the feature dependencies between different channel maps in the feature extraction network. In addition, the GIoU metric was introduced as a new evaluation index loss function in the cascade R-CNN. The experimental results in Section 4.1 demonstrate that the AP value of the proposed model surpassed those of existing classical object detection models. Moreover, the model accurately detected small targets within a range of less than 30 pixels. Building on the object detection results, this study introduced a novel NR-IQA method for digestive tract images that uses composite weights. The experimental results in Section 4.2 were highly correlated with the expert ratings, indicating the practical significance of the quality scores.

Author Contributions

W.S.: writing—original draft preparation, software, and methodology; P.L.: conceptualization and writing—review and editing; Y.L.: data curation, formal analysis, and validation; Y.F.: funding acquisition and resources; L.Z.: funding acquisition, project administration, and supervision. All authors read and agreed to the published version of the manuscript.

Funding

Jiangsu Provincial Key Research and Development Program: BE2019710; Suzhou Municipal Science and Technology Bureau: SYS2019008; Changzhou Municipal Science and Technology Program: CE20195001.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data that support the findings of this study are available from the corresponding author, upon reasonable request.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the study design; collection, analyses, or interpretation of data; writing of the manuscript; or decision to publish the results.

References

  1. Zhang, X.; Li, M.; Chen, S.; Hu, J.; Guo, Q.; Liu, R.; Zheng, H.; Jin, Z.; Yuan, Y.; Xi, Y.; et al. Endoscopic Screening in Asian Countries Is Associated With Reduced Gastric Cancer Mortality: A Meta-analysis and Systematic Review. Gastroenterology 2018, 155, 347–354.e9. [Google Scholar] [CrossRef] [PubMed]
  2. Zhang, Y.H.; Guo, L.J.; Yuan, X.L.; Hu, B. Artificial intelligence-assisted esophageal cancer management: Now and future. World J. Gastroenterol. 2020, 26, 5256–5271. [Google Scholar] [CrossRef] [PubMed]
  3. Ali, S.; Zhou, F.; Bailey, A.; Braden, B.; East, J.E.; Lu, X.; Rittscher, J. A deep learning framework for quality assessment and restoration in video endoscopy. Med. Image Anal. 2021, 68, 101900. [Google Scholar] [CrossRef] [PubMed]
  4. Wang, Z.; Simoncelli, E.P.; Bovik, A.C. Multiscale structural similarity for image quality assessment. In Proceedings of the Thrity-Seventh Asilomar Conference on Signals, Systems & Computers, Pacific Grove, CA, USA, 9–12 November 2003; Volume 2, pp. 1398–1402. [Google Scholar] [CrossRef]
  5. Wang, Z.; Bovik, A.C. Modern Image Quality Assessment. Synth. Lect. Image Video Multimed. Process. 2006, 2, 156. [Google Scholar] [CrossRef]
  6. Mittal, A.; Moorthy, A.K.; Bovik, A.C. No-Reference Image Quality Assessment in the Spatial Domain. IEEE Trans. Image Process. 2012, 21, 4695–4708. [Google Scholar] [CrossRef] [PubMed]
  7. Liu, X.; Weijer, J.V.D.; Bagdanov, A.D. RankIQA: Learning from Rankings for No-Reference Image Quality Assessment. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 1040–1049. [Google Scholar] [CrossRef]
  8. Ke, J.; Wang, Q.; Wang, Y.; Milanfar, P.; Yang, F. MUSIQ: Multi-scale Image Quality Transformer. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 5128–5137. [Google Scholar] [CrossRef]
  9. Talebi, H.; Milanfar, P. NIMA: Neural Image Assessment. IEEE Trans. Image Process. 2018, 27, 3998–4011. [Google Scholar] [CrossRef] [PubMed]
  10. Behrens, A.; Bommes, M.; Gross, S.; Aach, T. Image quality assessment of endoscopic panorama images. In Proceedings of the 2011 18th IEEE International Conference on Image Processing, Brussels, Belgium, 11–14 September 2011; pp. 3113–3116. [Google Scholar] [CrossRef]
  11. Aubreville, M.; Stoeve, M.; Oetter, N.; Goncalves, M.; Knipfer, C.; Neumann, H.; Bohr, C.; Stelzle, F.; Maier, A. Deep learning-based detection of motion artifacts in probe-based confocal laser endomicroscopy images. Int. J. Comput. Assist. Radiol. Surg. 2019, 14, 31–42. [Google Scholar] [CrossRef] [PubMed]
  12. Kamen, A.; Sun, S.; Wan, S.; Kluckner, S.; Chen, T.; Gigler, A.M.; Simon, E.; Fleischer, M.; Javed, M.; Daali, S.; et al. Automatic Tissue Differentiation Based on Confocal Endomicroscopic Images for Intraoperative Guidance in Neurosurgery. Biomed. Res. Int. 2016, 2016, 6183218. [Google Scholar] [CrossRef] [PubMed]
  13. Outtas, M.; Zhang, L.; Deforges, O.; Hamidouche, W.; Serir, A. Evaluation of No-reference quality metrics for Ultrasound liver images. In Proceedings of the 2018 Tenth International Conference on Quality of Multimedia Experience (QoMEX), Cagliari, Italy, 29 May–1 June 2018; pp. 1–3. [Google Scholar] [CrossRef]
  14. Mittal, A.; Soundararajan, R.; Bovik, A.C. Making a “Completely Blind” Image Quality Analyzer. IEEE Signal Process. Lett. 2013, 20, 209–212. [Google Scholar] [CrossRef]
  15. Singh, P.; Mukundan, R.; de Ryke, R. Quality analysis of synthetic ultrasound images using co-occurrence texture statistics. In Proceedings of the 2017 International Conference on Image and Vision Computing New Zealand (IVCNZ), Christchurch, New Zealand, 4–6 December 2017; pp. 1–6. [Google Scholar] [CrossRef]
  16. Zhang, T.; Wang, L.; Gu, J.; Deng, Y.J.; Li, X.D. Design and Implementation of A Real-time Capsule Endoscope Image Quality Assessment Method. In Proceedings of the 2009 3rd International Conference on Bioinformatics and Biomedical Engineering, Beijing, China, 11–13 June 2009; pp. 1–4. [Google Scholar] [CrossRef]
  17. Bochkovskiy, A.; Wang, C.; Liao, H.M. YOLOv4: Optimal Speed and Accuracy of Object Detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  18. Cai, Z.; Vasconcelos, N. Cascade R-CNN: High Quality Object Detection and Instance Segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 43, 1483–1498. [Google Scholar] [CrossRef] [PubMed]
  19. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed]
  20. Xie, S.; Girshick, R.; Dollár, P.; Tu, Z.; He, K. Aggregated Residual Transformations for Deep Neural Networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 5987–5995. [Google Scholar] [CrossRef]
  21. Rahman, H.; Bukht, T.F.N.; Imran, A.; Tariq, J.; Tu, S.; Alzahrani, A. A Deep Learning Approach for Liver and Tumor Segmentation in CT Images Using ResUNet. Bioengineering 2022, 9, 368. [Google Scholar] [CrossRef] [PubMed]
  22. Rezatofighi, H.; Tsoi, N.; Gwak, J.; Sadeghian, A.; Reid, I.; Savarese, S. Generalized Intersection Over Union: A Metric and a Loss for Bounding Box Regression. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 658–666. [Google Scholar] [CrossRef]
  23. Everingham, M.; Van Gool, L.; Williams, C.K.; Winn, J.; Zisserman, A. The Pascal Visual Object Classes (VOC) Challenge. Int. J. Comput. Vis. 2010, 88, 303–338. [Google Scholar] [CrossRef]
  24. Liu, S.; Qi, L.; Qin, H.; Shi, J.; Jia, J. Path Aggregation Network for Instance Segmentation. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 8759–8768. [Google Scholar] [CrossRef]
  25. Sun, P.; Zhang, R.; Jiang, Y.; Kong, T.; Xu, C.; Zhan, W.; Tomizuka, M.; Li, L.; Yuan, Z.; Wang, C.; et al. Sparse R-CNN: End-to-End Object Detection with Learnable Proposals. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 14449–14458. [Google Scholar] [CrossRef]
  26. Redmon, J.; Farhadi, A. YOLOv3: An Incremental Improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar] [CrossRef]
  27. Lin, T.-Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal Loss for Dense Object Detection. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2999–3007. [Google Scholar] [CrossRef]
  28. Li, Y.; Chen, Y.; Wang, N.; Zhang, Z.-X. Scale-Aware Trident Networks for Object Detection. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 6053–6062. [Google Scholar] [CrossRef]
  29. Jocher, G. YOLOv5 by Ultralytics, Version 7.0. Computer software. 2020. [CrossRef]
Figure 1. Architecture of the proposed framework for detection of image artifacts using improved cascade region-based CNN for no-reference endoscopic image quality assessment.
Figure 1. Architecture of the proposed framework for detection of image artifacts using improved cascade region-based CNN for no-reference endoscopic image quality assessment.
Bioengineering 10 01288 g001
Figure 2. Endoscopic image preprocessing operations.
Figure 2. Endoscopic image preprocessing operations.
Bioengineering 10 01288 g002
Figure 3. Analysis of statistical characteristics of sample data. The scale distribution of each defect category is shown, where the x-axis represents the height of the ground truth (GT) and the y-axis represents the width of the GT.
Figure 3. Analysis of statistical characteristics of sample data. The scale distribution of each defect category is shown, where the x-axis represents the height of the ground truth (GT) and the y-axis represents the width of the GT.
Bioengineering 10 01288 g003
Figure 4. Performance of mosaics in data augmentation.
Figure 4. Performance of mosaics in data augmentation.
Bioengineering 10 01288 g004
Figure 5. Residual structure.
Figure 5. Residual structure.
Bioengineering 10 01288 g005
Figure 6. Artifact detection model based on the improved cascade R-CNN.
Figure 6. Artifact detection model based on the improved cascade R-CNN.
Bioengineering 10 01288 g006
Figure 7. Improved feature pyramid structure.
Figure 7. Improved feature pyramid structure.
Bioengineering 10 01288 g007
Figure 8. Channel attention mechanism.
Figure 8. Channel attention mechanism.
Bioengineering 10 01288 g008
Figure 9. Three cases in L2 norm distance.
Figure 9. Three cases in L2 norm distance.
Bioengineering 10 01288 g009
Figure 10. Generalized Intersection over Union diagram.
Figure 10. Generalized Intersection over Union diagram.
Bioengineering 10 01288 g010
Figure 11. Quality assessment based on multiple parameters. The left side shows the artifact target detection results, and the right side shows the area ratio of different artifacts and the calculated mass score below.
Figure 11. Quality assessment based on multiple parameters. The left side shows the artifact target detection results, and the right side shows the area ratio of different artifacts and the calculated mass score below.
Bioengineering 10 01288 g011
Figure 12. Confusion matrix showing true positive detections and confused classes in all eight classes.
Figure 12. Confusion matrix showing true positive detections and confused classes in all eight classes.
Bioengineering 10 01288 g012
Figure 13. Square root of the area of the GT (left) and predicted box (right).
Figure 13. Square root of the area of the GT (left) and predicted box (right).
Bioengineering 10 01288 g013
Figure 14. Artifact target detection results, scores from various experts, scores from the proposed method, and the proportion of detected artifact regions.
Figure 14. Artifact target detection results, scores from various experts, scores from the proposed method, and the proportion of detected artifact regions.
Bioengineering 10 01288 g014
Figure 15. Heatmap between the mass scores of the three experts and the QS calculated by the proposed method.
Figure 15. Heatmap between the mass scores of the three experts and the QS calculated by the proposed method.
Bioengineering 10 01288 g015
Table 1. Image modalities and digestive tract site statistics.
Table 1. Image modalities and digestive tract site statistics.
ModeTrain (Esophagus/ Stomach/Unknown)Validation (Esophagus/ Stomach/Unknown)Test (Esophagus/ Stomach/Unknown)
WLI738/730/1991/90/293/89/4
NBI154/150/1328/6/427/5/5
iodine-stained17/21/04/5/04/4/0
Sum1842 (909/901/32)230 (123/101/6)231 (124/98/9)
Total2303 (1156/1100/47)
Table 2. Artifact types and label statistics.
Table 2. Artifact types and label statistics.
TypesNumber of Images that Contain the ArtifactNumber of LabelsLabel Ratio (%)
Blood861121.15
Blur4184334.44
Bubbles3037888.08
Instrument66710.73
Specularity776458246.98
Contrast8199109.33
Saturation4665825.97
Artifacts696227623.33
Total 9754100
Table 3. Endoscopic image quality assessment principles.
Table 3. Endoscopic image quality assessment principles.
Quality ScoreScoring Principles
5Endoscopic image is perfect, artifacts are negligible.
4Endoscopic image exhibits minimal and small-sized artifacts, which do not interfere with the image.
3Endoscopic image exhibits large-sized or numerous artifacts, which slightly interfere with the image.
2Endoscopic image exhibits large-sized and numerous artifacts, which interfere with the image.
1Endoscopic image exhibits large number of artifacts, which seriously interfere with the image.
Table 4. AP for typical networks.
Table 4. AP for typical networks.
NetworkAP AP IoU = 0.50 AP IoU = 0.75 AP small AP medium AP large
Cascade R-CNN [18]0.1490.3350.1020.0310.1090.186
Faster R-CNN [19]0.2030.3950.1780.0330.1440.266
PANet [24]0.2110.4160.1950.0300.1460.272
Sparse R-CNN [25]0.1360.2780.1200.0180.0830.175
YOLOV3 [26]0.1400.3210.1140.0180.0940.177
RetinaNet [27]0.1360.3100.0900.0590.1020.142
TridentNet [28]0.3240.5810.3010.1400.2160.338
YOLOV4 [17]0.3390.5970.3270.1600.2250.346
YOLOV5 [29]0.3460.6080.3300.1640.2390.358
Proposed0.3700.6240.3610.1710.2500.392
Table 5. AP for labels.
Table 5. AP for labels.
NetworkArtifactBloodBlurBubblesContrastInstrumentSaturationSpecularity
Cascade R-CNN0.1350.0100.1320.0880.3690.5230.2240.119
Faster R-CNN0.1460.0210.1870.0830.3670.5220.2210.129
PANet0.1500.0280.2170.0830.3490.5340.2260.130
Sparse R-CNN0.0860.0020.1150.0240.2840.3790.1480.051
YOLOV30.1160.0150.1570.0400.2500.3450.1430.057
RetinaNet0.1540.0010.3700.0550.2200.0020.1180.168
TridentNet0.2940.0110.5370.2000.3920.5240.3060.332
YOLOV40.2660.06180.5420.2080.3630.5970.2550.330
YOLOV50.2700.07490.5490.2210.3780.6670.2780.332
Proposed0.3180.1050.6080.2420.4100.5980.3330.342
Table 6. Ablation experiments.
Table 6. Ablation experiments.
ExperimentsImproved FPNGIoUAP
1--0.149
2-0.301
3-0.293
40.370
Table 7. Example of artifact detection using our network in different endoscopic modalities.
Table 7. Example of artifact detection using our network in different endoscopic modalities.
ModeArtifactBloodBlurContrastInstrumentSaturationSpecularityBubbles
WLIBioengineering 10 01288 i001Bioengineering 10 01288 i002Bioengineering 10 01288 i003Bioengineering 10 01288 i004Bioengineering 10 01288 i005Bioengineering 10 01288 i006Bioengineering 10 01288 i007Bioengineering 10 01288 i008
NBIBioengineering 10 01288 i009Bioengineering 10 01288 i010Bioengineering 10 01288 i011Bioengineering 10 01288 i012Bioengineering 10 01288 i013Bioengineering 10 01288 i014Bioengineering 10 01288 i015Bioengineering 10 01288 i016
Iodine-stainedBioengineering 10 01288 i017Bioengineering 10 01288 i018Bioengineering 10 01288 i019Bioengineering 10 01288 i020Bioengineering 10 01288 i021Bioengineering 10 01288 i022Bioengineering 10 01288 i023Bioengineering 10 01288 i024
Table 8. Measured relevance using the correlation measure of the Spearman scale and its corresponding p-value.
Table 8. Measured relevance using the correlation measure of the Spearman scale and its corresponding p-value.
IQA(NR)Spearman Rank Correlation   [ 1 , + 1 ]
Corr.Value p -Value
vs. Expert #1 vs. Expert #2 vs. Expert #3 vs. Expert #1 vs. Expert #2 vs. Expert #3
BRISQUE0.14310.04030.09570.15320.68930.3411
NIQE0.23890.31660.26740.01610.00130.007
MUSIQ0.22400.25650.28040.02430.00960.0045
NIMA0.07220.00250.05560.47270.98010.5811
Proposed0.58940.59860.63345.09 × 10 10 5.49 × 10 10 9.47 × 10 12
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Sun, W.; Li, P.; Liang, Y.; Feng, Y.; Zhao, L. Detection of Image Artifacts Using Improved Cascade Region-Based CNN for Quality Assessment of Endoscopic Images. Bioengineering 2023, 10, 1288. https://doi.org/10.3390/bioengineering10111288

AMA Style

Sun W, Li P, Liang Y, Feng Y, Zhao L. Detection of Image Artifacts Using Improved Cascade Region-Based CNN for Quality Assessment of Endoscopic Images. Bioengineering. 2023; 10(11):1288. https://doi.org/10.3390/bioengineering10111288

Chicago/Turabian Style

Sun, Wei, Peng Li, Yan Liang, Yadong Feng, and Lingxiao Zhao. 2023. "Detection of Image Artifacts Using Improved Cascade Region-Based CNN for Quality Assessment of Endoscopic Images" Bioengineering 10, no. 11: 1288. https://doi.org/10.3390/bioengineering10111288

APA Style

Sun, W., Li, P., Liang, Y., Feng, Y., & Zhao, L. (2023). Detection of Image Artifacts Using Improved Cascade Region-Based CNN for Quality Assessment of Endoscopic Images. Bioengineering, 10(11), 1288. https://doi.org/10.3390/bioengineering10111288

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop