Next Article in Journal
WAID: A Large-Scale Dataset for Wildlife Detection with Drones
Previous Article in Journal
Application of Hyperpure Chlorine Dioxide for Regenerative Endodontic Treatment of a Root-Canal-Treated Immature Tooth with External Cervical Resorption and Periapical Lesion: A Case Report
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Hybrid Intelligent Pattern Recognition Systems for Mass Segmentation and Classification: A Pilot Study on Full-Field Digital Mammograms

by
Anastasios Dounis
*,
Andreas-Nestor Avramopoulos
and
Maria Kallergi
Department of Biomedical Engineering, Egaleo Park Campus, University of West Attica, 12243 Athens, Greece
*
Author to whom correspondence should be addressed.
Appl. Sci. 2023, 13(18), 10401; https://doi.org/10.3390/app131810401
Submission received: 21 August 2023 / Revised: 15 September 2023 / Accepted: 16 September 2023 / Published: 17 September 2023

Abstract

:
Governments and health authorities emphasize the importance of early detection of breast cancer, usually through mammography, to improve prognosis, increase therapeutic options and achieve optimum outcomes. Despite technological advances and the advent of full-field digital mammography (FFDM), diagnosis of breast abnormalities on mammographic images remains a challenge due to qualitative variations in different tissue types and densities. Highly accurate computer-aided diagnosis (CADx) systems could assist in the differentiation between normal and abnormal tissue and the classification of abnormal tissue as benign or malignant. In this paper, classical, advanced fuzzy sets and fusion techniques for image enhancement were combined with three different thresholding methods (Global, Otsu and type-2 fuzzy sets threshold) and three different classifying techniques (K-means, FCM and ANFIS) for the classification of breast masses on FFDM. The aim of this paper is to identify the performance of the advanced fuzzy sets, fuzzy sets type-2 segmentation, decisions based on K-means and FCM, and the ANFIS classifier. Sixty-three combinations were evaluated on ninety-seven digital mammographic masses (sixty-five benign and thirty-two malignant). The performance of the sixty-three combinations was evaluated by estimating the accuracy, the F1 score, and the area under the curve (AUC). LH-XWW enhancement method with Otsu thresholding and FCM classifier outperformed all other combinations with an accuracy of 95.17%, F1 score of 89.42% and AUC of 0.91. This algorithm seems to offer a promising CADx system for breast cancer diagnosis on FFDM.

1. Introduction

Breast cancer is currently the most common cancer globally and the second leading cause of cancer death in women today [1]. Supporting physician’s diagnostic decisions has become an important application of artificial intelligence (AI), the role of which is expected to expand in the future [2]. Hybrid intelligent systems (HIS) in the field of medical pattern recognition have been widely applied for data classification. HIS combines intelligence tools and methods that make sense on a theoretical basis and can be applied practically. The unique characteristic of a HIS is that it can integrate two or more intelligent techniques that maintain their operating principles and attributes despite being fused or transformed. HIS performance depends on the selected intelligent systems and their architecture. In this paper, the HIS adaptive neuro-fuzzy inference system (ANFIS) is used to classify breast masses from digital mammograms that have also been processed with various enhancement and segmentation methods. A short review of related published work segmentation and classification follows below.
There are many CADx systems that involve several processing steps and different methodologies for enhancement, segmentation, feature selection and classification. Previous works have accomplished great results, but in real testing data, there is always the question of whether they will be able to carry on. Therefore, the methods that are used on systems that detect pathologies and help the physician perform their work more efficiently have to become more advanced to overcome the demands that new images may present.
This work focuses on the HIS ANFIS classification of digital mammographic masses that were previously enhanced by advanced fuzzy sets, as reported in our previous work [3]. K-means and Fuzzy C-means classifiers are also implemented for comparison purposes. The steps of the proposed classification algorithm are:
  • Advanced fuzzy sets for image enhancement.
  • Fused images based on OWA operators.
  • Global, Otsu and type-2 fuzzy sets for thresholding.
  • Feature extraction.
  • Feature selection.
  • Classification using K-means, Fuzzy C-means and ANFIS.
Image enhancement contains fuzzy methods in order to enhance the image contrast. Five types of fuzzy sets have been used: linguistic hedges; a fuzzy enhancement function; and advanced fuzzy sets that are intuitionistic, Pythagorean and Fermatean fuzzy sets [4]. Two types of OWA operators have been used, with the target being the fusion of the fuzzy techniques and the creation of an image that would provide better results.
Thresholding techniques are used so the enhanced image can undergo segmentation. Global and Otsu thresholding are classic methods for image segmentation [5]. Fuzzy sets type-2 is a thresholding method where a membership function is used, and the threshold index is extracted out of the ultrafuzziness of the image.
The features that are extracted are based on the geometry of the thresholded image, the texture of the initial image and the fuzzy features from the enhanced fuzzified image.
Classification is performed with the classic method of K-means and additionally more complicated methods of Fuzzy C-means (FCM) [6,7] and ANFIS technique that combines neural network and fuzzy inference engine methods.
The novelty and contribution of this work rely on the combination of various elements for the optimum classification of mammographic masses. The proposed methods include:
  • An adaptive neuro-fuzzy inference system (ANFIS) for classification.
  • A hybrid classification procedure of mammographic masses based on FCM/K-means clustering algorithms using a decision tree.
  • Advanced fuzzy sets for contrast enhancement.
In addition, the proposed methods are trained and tested on masses detected with FFDM, which is the current state-of-the-art in mammography. FFDM images differ significantly from digitized screen-film mammography (SFM) images and require either modified or new CADx approaches.
The study is structured as follows: Section 2 contains the mathematical formulas of fuzzy image enhancement, mammography image segmentation, feature extraction and classification. It also contains a description of the training and testing datasets. Section 3 presents the results of this study, followed by the discussion in Section 4 and conclusions in Section 5.

Related Works

In papers [8,9,10], the authors analyze, evaluate and review existing artificial intelligence-based classification and segmentation techniques proposed in the last years for early detection and diagnosis of breast cancer using medical images. Yazdanbakhsh et al. [11] created a deep neuro-fuzzy network based on definitions of Takagi–Sugeno–Kang and tested the classifier on three datasets. The highest accuracy that they achieved was 99.58%. Gao et al. [12] represented a review of deep and machine learning techniques for mammography, assuming that the most frequent supervised methods for classification are artificial neural networks (ANNs), support vector machine (SVM) and random forest (RF). The most frequent unsupervised methods are the clustering algorithms of K-means, principal component analysis (PCA) and singular value decomposition (SVD). In deep learning, convolutional neural networks (CNNs) are one of the most used algorithms. Currently, computers that try to simulate the human brain are far from the learning capabilities of it. Almutairi et al. [13] proposed the deep reinforcement learning-based Deep Q learning (DQL) method as a classifier and the gorilla troops optimization (GTO) algorithm for feature selection. The method GTO-DQL was tested on three datasets with an accuracy up to 99.02%. Malebary et al. [14] proposed an algorithm named breast mass classification system (BMC) that is based on K-means, long short-term memory network of recurrent neural network (RNN), CNN and random forest. The proposed BMC system was tested on DDSM and MIAS dataset. For DDSM, the accuracy and AUC were 96% and 94–97%, respectively. For MIAS, the accuracy and AUC were 95% and 94–98%, respectively. Li et al., in [15], introduced the Dual Core Net (DCN), which computes mass segmentation and classifies simultaneously. The proposed algorithm was tested on DDSM and INbreast datasets, achieving 92.27% DI coefficient and 85% AUC for DDSM and 93.69% DI coefficient and 93% AUC for INbreast dataset. Parah et al., in [16], proposed a procedure for breast cancer masses that includes the Watershed segmentation and a hybrid machine learning classifier that is based on multi-layer perceptron (MLP), J48 and K-means. The methods were tested on the database from the UCI Machine Learning Repository and achieved an accuracy of over 88%. A paper by Sarvestani et al. focused on the decision tree classification algorithm [17]. In [18], a novel mass detection process that includes enhancement, characterization and classification was described. The classification mechanism is based on Bayesian regularization back-propagation networks and ANFIS techniques. Hosseini et al. [19] present a review of the application of the adaptive neuro-fuzzy inference system as a classifier in medical image classification during the past 16 years.
The most common deep networks that are used for image processing are convolutional neural networks, recurrent networks, and generative adversarial models [20]. The authors [21] proposed a novel method of an ensemble classifier set and depthwise separable convolution to extract shallow features. Abdullah et al. [22] represented and developed a convolutional neural network model for brain tumor segmentation. Abdel Rahman et al. [23] introduced a modified version of InceptionV3 and ResNet50 classifiers that are similar to CNN and tested the procedure on the DDSM dataset. They managed to obtain an 85.7% accuracy. Salama et al. [24] tested segmentation and classification methods in order to identify masses as malignant or benign. Fully, CNNs were introduced. The technique where applying data augmentation with a modified U-Net model and classifying the data with the InceptionV3 method achieved 98.87% accuracy and 98.88% AUC. Nemade et al. [25] represented two deep learning-based ensemble models. They used the VGG16, InceptionV3 and VGG19 as base classifiers and the two ensemble models were trained. The first ensemble model used a linear meta-learner in form of logistic regression and the second ensemble model used a neural net as the meta-learner for classification. The second ensemble model achieved 98.1% accuracy.
There is significant published work on the segmentation and classification of mammographic images. The fuzzy sets are notably useful and often used to address uncertainty issues and quantify uncertainty. Lin et al. [26] proposed a fuzzy-based novel quality control algorithm for the segmentation of medical images. Salih et al. [27] presented an algorithm based on fuzzy set methods for segmenting the boundaries of breast masses. Chaira T. [28] refers to thresholding techniques in general and proposes a method where intuitionistic fuzzy sets can be used as a segmentation method for images that contain blood vessels and blood cells. Tahoun et al. [29] used the BGWO feature selection technique and analyzed the percentage of features that the algorithm could reduce in order to achieve a high ROC area. The researchers achieved an accuracy of 78% and AUC 0.871 for the classification as benign or malignant. Catro-Tapia et al. [30] analyzed 14 classifiers of deep learning methods that were categorized into five classes. Their study showed that the CNN GoogLeNet classifier achieved the best results with an accuracy of 91.92% and AUC of 99.29%. Mobark et al. [31] proposed a CNN model, namely CoroNet, to perform automated breast cancer detection. It attained an accuracy of 94.92% on a four-class classification (benign/malignant mass and calcification). Alshehri et al. [32] studied CNN with and without attention mechanisms (AMs). The implementation of a deep learning algorithm on infrared thermal breast images showed an accuracy of over 99% with AMs and 92.32% without. Du et al. [33] presented a procedure that classifies architectural distortion, a feature of breast cancer. Enhancing the images with top–bottom hat and exponential transformation, reducing the noise with NSCT and finding a threshold for the segmentation of the image using the improved PCNN achieved an accuracy of 93.16% and an AUC of 0.93.

2. Materials and Methods

2.1. Fuzzy Image Enhancement

Image enhancement methods were previously developed and evaluated based on fuzzy sets and the fusion of them on mammograms [3]. They provided the starting point for the development of a fully automated pattern recognition and classification algorithm. The enhancement methods were linguistic hedge concentration (LH-CON); a linguistic hedge that was proposed by Xie, Wang and Wu (LH-XWW); and intuitionistic, Pythagorean and formation fuzzy sets (IFS, PFS and FFS). The methods created images that were fused with ordered weighted average (OWA) operators, OWA with genetic algorithm (OWA-GA) and OWA where Hong and Kim (OWA-HK) presented a way of finding weights. Genetic algorithms were used in order to optimize parameters for LH-XWW, IFS, PFS, FFS, and OWA-GA.

2.2. Advanced Fuzzy Sets and OWA Fusion Images

LH-CON at fuzzy sets creates a linear correlation between the input and the enhanced image. LH-XWW also has a linear correlation, but the output is different for every image. The use of a genetic algorithm allowed for the adaptive parameter tuning of each image. Advanced fuzzy sets such as IFS, PFS and FFS gain attention because of their non-linear character. They also use genetic algorithms to find the optimal values for parameters that are not constant. The five new images that were created by the advanced fuzzy sets are fused by OWA operators. The weights of OWA operators are obtained with two methods. The first method uses a genetic algorithm (OWA-GA), and the second uses a function that Hong and Kim proposed (OWA-HK). The OWA aggregation method is applied at every pixel of the image.

2.3. Mammography Image Segmentation

The grey level thresholding plays a vital role in pattern recognition in medical images. It follows the process of enhancement, and its output improves feature extraction for the subsequent pattern recognition techniques. Three types of thresholding methods were used in this work: global, Otsu and type-2 fuzzy sets thresholding.

2.3.1. Segmentation via Global Thresholding

The technique of global thresholding is a common method that does not need many computational resources. The algorithm aims to find the optimal threshold value by searching in the spatial domain. An iterative process is followed that analyzes the parameters of the enhanced image and finds the global threshold. Parameter λ is a pre-defined constant [34]. The iterative process is:
  • Initializing the threshold value using the function T = 0.5 · I m a x + I m i n , where I m a x and I m i n are the maximum and minimum values of the image.
  • Thresholding the image with the value T . This way, the pixels are separated into two groups: I 1 , which contains the pixels with values T , and I 2 , which contains the pixels with values < T .
  • Finding the average m 1 and m 2 from the groups I 1 and I 2 that were calculated previously.
  • Finding a new threshold value using the function T = 0.5 m 1 + m 2 to compare it with the previous thresholding value.
  • If T T > λ , then the steps 2 to 4 must be repeated. Otherwise, the threshold value is T [34].

2.3.2. Otsu Thresholding

  • A classic method for thresholding an image is Otsu’s method. Otsu threshold is calculated in order to minimize the overlapping of class distributions. Otsu threshold is found when the minimum entropy of the sum of background and foreground is calculated [35,36]. For our work, we used the command of MATLAB “graythresh” to find the threshold value and “imbinarize” to create the thresholded image with the threshold value [37].

2.3.3. Interval Type-2 Fuzzy Sets for Thresholding

Another approach to image segmentation is using the type-2 fuzzy sets method. This algorithm also has an iterative procedure. Before the beginning of the iterative procedure, the Tizhoosh membership function must be applied to the enhanced image. The Tizhoosh function is:
μ ( g ) = 0 , g g m i n   o r   g g m a x L g = g g m i n T g m i n α , g m i n < g T R g = g m a x g g m i n T β , T < g < g m a x
The min and max indexes are used in the scanning procedure, which is described next [38].
The next step is to find two new membership functions for each pixel. There are two ways of creating these functions.
The first way is:
μ U x = μ x 1 α and μ L x = μ x α .
The second is:
μ U x = μ x 1 α and μ L x = μ x 1 + α .
That means:
μ U x = μ x 0.5 and μ L x = μ x 2
or
μ U x = μ x 0.75 and μ L x = μ x 1.25 .
In this paper, the first way was used, but with different exponents:
μ U x = μ x 0.25 and μ L x = μ x 4 [38].
The lower and upper functions can be seen in Figure 1.
The next action is to extract the ultrafuzziness of the system. The ultrafuzziness index acts as a means of comparison to find the optimal threshold value. The ultrafuzziness of the system is extracted for the spatial domain by the function [38]:
γ ~ A ~ = 1 M N i = 1 M 1 j = 1 N 1 μ U g i j μ L ( g i j )
and using the histogram of the image by the function:
γ ~ A ~ = 1 M N g = 0 L 1 h ( g ) × μ U g μ L ( g )
where M and N are the rows and columns of the image, and h is the histogram of the image.
The general structure of the algorithm of type-2 fuzzy sets and the calculation of the ultrafuzziness includes the following:
  • Selection of the membership function. In this paper, it was the Tizhoosh function. It can also be used as a sigmoid membership function.
  • Calculation of the histogram if the histogram ultrafuzziness (3) is used.
  • Initialize membership functions start.
  • Scanning the function across the entire grayscale of the image.
  • Finding the functions μ U x and μ L x for every step of the scanning procedure.
  • Finding maximum ultrafuzziness and the optimal T .
  • Thresholding the image using T of the maximum ultrafuzziness.
  • Finally, defuzzification of the result using the function x m n = μ m n · max v a l u e [38].
The thresholding methods that were used can change the optimal threshold value in every image. In Figure 2, we present the thresholded images that the thresholding methods created from an ROI of a mammogram.

2.3.4. Flood Fill Operation

A thresholded image may present several areas that are above the threshold value, as shown in Figure 3. MATLAB commands “regionprops”, “ismember” and “imfill” were used to remove unconnected regions and fill the voids to extract a solid region with the largest area [37].
In Figure 4, we can see a general overview of the segmentation process.

2.4. Feature Extraction

A classification system aims to achieve optimum classification by extracting the best statistical scores. The methods that were described create images that are used for feature extraction. The features of an image show different perspectives of it. In this paper, features were extracted that describe the geometry, the texture and the fuzziness of the masses. Nine features were extracted, two of which were geometric, four were texture and three were fuzzy features. The geometric features were the circularity and the minimum to maximum axial distance of the thresholded mass. The texture features were the contrast between the mass and the background, homogeneity, correlation and energy of the thresholded mass. The fuzzy features were the area, the perimeter and compactness.

2.4.1. Geometric Features

MATLAB command “regionprops” was used to extract the circularity of the minor and major axial distances [37]. Circularity indicates how round a mass is and how close it is to an ideal circle. If the index is close to one, the more circular it is. If it is close to zero, the more irregular it is. The minor and major axial distances were merged into one feature by dividing the minor by the major axial distance as a way to represent the elongation of a mass. The lower the value of this index, the more elongated it is, as the minimum axial distance is small, and the maximum axial distance is bigger.

2.4.2. Texture Features

Image Masks

The method of extracting masks from the images is helpful for the next feature extraction techniques. With the thresholding techniques and the creation of masks, matrices were created that contain exclusively the mass and exclusively the background. The matrix that contained the mass was created by inserting the value of the initial image if the pixel was above the threshold value, and the value of the initial image was ignored if it was below the threshold value. The opposite procedure was followed in order to create the background matrix.

GLCM Matrix

The process of extracting texture features is performed indirectly with the image. These features need a gray-level co-occurrence Matrix (GLCM) to be extracted. A GLCM matrix has lines and columns equal to the gray levels of the image. That means if an image has 256 gray levels, the dimensions of the GLCM matrix will be 256 × 256 . Each element of the table represents the combination of the row and the column of it. A GLCM matrix usually has large dimensions, and this is why much temporary data are stored, i.e., a 256 × 256 for each image. Such a matrix is very sensitive to extracting texture features. The gray levels of an image are usually decreased because of the large dimensions and sensitivity of a GLCM matrix. The composition of a GLCM matrix is described in Figure 5 for four gray levels. By scanning, from left to right, the image from which the GLCM matrix is generated, which means a degree of 0°, the upper left element of the GLCM matrix fills with the number of times that the combination 0,0 occurs. This indicates how many times a gray level 0 pixel (neighbor pixel) is to the right of another gray level 0 pixel (reference pixel). If the degree of creating the GLCM matrix is different, i.e., 45°, then the upper left element of the GLCM matrix will fill with the number of times that the combination 0,0 occurs; however, it would indicate how many times a pixel with gray level 0 (neighbor pixel) is diagonally at the upper right of another pixel with gray level 0 (reference pixel) [39].
The MATLAB command “graycomatrix” was used for the creation of the GLCM matrix with a degree of 0° [37].

Extraction of Texture Features

The features that were extracted with the help of the GLCM matrix were homogeneity, correlation and energy of the mass. The application of the GLCM matrix was performed on the mass mask of every image. For the contrast between the mass and the background, the GLCM matrix was not used [40]. The MATLAB command “graycoprops” may be applied for the extraction of homogeneity, correlation and energy [37]. Alternatively, the following functions may be used [41].
For homogeneity, the function is:
h o m o g e n e i t y = i , j = 1 n p i j 1 + ( i j ) 2
where p i j is the element of the GLCM matrix at row I and column j, and n is the maximum value of the gray levels.
For correlation:
c o r r e l a t i o n = i , j = 1 n p i j ( i μ i ) ( j μ j ) σ i 2
where μ k = i j k · p i j and σ k 2 = i j p i j i μ k 2 .
For energy:
e n e r g y = i = 1 n j = 1 n p i j 2
The extraction of the contrast between the mass and the background is performed with the help of the masks and without using the GLCM matrix. It was based on Weber’s law, where the ratio of the background and interest point difference intensity to the background intensity is constant [42]. The function is as follows [43]:
c o n t r a s t = I b I f I b
where I b is the intensity of the background, and I f is the intensity of the interest point.
For the intensity of the background and the interest point to be found, the background and interest point mask were used, respectively. The average value of the background mask was defined as the background intensity, and the average pixel value of the interest point mask was defined as the intensity of the interest point.

2.4.3. Fuzzy Features

The extraction of the fuzzy features was based on the fuzzified version of the enhanced images. The thresholding technique was useless for this kind of feature. The functions have a fuzzy matrix as an input and a value as an output. The features that were extracted during this method are the area, the perimeter and the compactness [44].
For the fuzzy area, the function is:
f u z z y   A r e a μ = m = 1 M n = 1 N μ m , n
where the matrix has dimensions M × N and μ m , n is membership participation of each pixel of the image.
For the fuzzy perimeter:
f u z z y   P e r i m e t e r μ = m = 1 M n = 1 N 1 μ m , n μ m , n + 1 + m = 1 M 1 n = 1 N μ m , n μ m + 1 , n
For the compactness of the image, the function was proposed by Rosenfeld:
c o m p a c t n e s s μ = f u z z y   A r e a μ f u z z y   P e r i m e t e r μ 2
Compactness obtains its maximum value when the interest area is a circle [45].

2.5. Data Normalization and Feature Selection

2.5.1. Normalization

The procedure of feature extraction is performed for all the mammograms, creating a matrix that contains all of them: the type of enhancement, the type of thresholding and an index for each image if it contains a malignant or a benign mass. If the mass is malignant, then the index is “1”, and if the mass is benign, then the index is “0”.
The normalization method is a necessary procedure and must be performed before the data are inserted into the classifier because the range of the features’ values may vary significantly. For example, one feature may have values ranging from 0 to 1, while another feature may have values from 0 to 5000. A normalization method maps all features to the same range, either from 0 to 1 or from −1 to 1. The value of 1 maps the maximum feature value, while 0 or −1 maps the minimum value. MATLAB’s command “normalize” was used to map all features’ values from 0 to 1 [37].
The normalization can also be performed with the following function [46]:
x i = x i x m i n x m a x x m i n
where x m i n and x m a x are the minimum and maximum value of a feature, respectively; x i is the i-th value from the total values that were extracted from the images; and x i is the new normalized value.

2.5.2. Feature Selection

The feature selection and ranking for the training of an automated decision system is a crucial and challenging process [39,40]. Feature selection methods are applied to different fields, including the pharmaceutical and oil industry, voice and pattern recognition, biotechnology, etc. [47]. The aim of these methods is to reduce the features that are inserted into the classifier by keeping out unnecessary ones and, by extension, decreasing noise in multidimensional data. A small number of features allows for better visualization and understanding of the data, fewer operations and storage requirements, and more efficient training and testing processes. Overtraining of the classifier can also be avoided while improving its performance, giving faster results and a better understanding of the training process. An exhaustive search was used in this study for feature selection. This method investigates all feature combinations and yields optimum results [48]. The relatively small dataset of our pilot study allowed for the application of the method that is usually computationally intensive and almost impractical to use with a large number of features and large datasets [41,42].

2.6. Classification

2.6.1. K-Means

Classification with K-means clustering is widely used because of its simplicity to understand and apply. We initialize how many clusters we want to have (k clusters), and the algorithm starts with random k clusters. Then, k centers are chosen from the clusters that are the centroids of them. New clusters are created according to the distance of the data from the centers. This procedure is repeated until the centers do not change [49,50]. Many researchers have applied clustering techniques to solve medical image segmentation problems. K-means is a well-known clustering algorithm that is widely applied for segmenting and classifying medical images. The training data create the centers that the testing data use to classify the masses [51,52,53]. MATLAB’s command “kmeans” was applied to create clusters and “pdist2” for the test data [37]. Two groups are created (“G1” and “G2” in Figure 6) with labels “1” and “2”, without knowing which is malignant or benign. Two truth tables are created, the first using the labeled data with “1” as malignant and the second truth table as “2” as malignant. Two accuracies are extracted (“Acc1” and “Acc2” in Figure 6). We use the truth table of the accuracy that is greater than the other.

2.6.2. Fuzzy C-Means

Since fuzzy sets were introduced, the concept of data having a membership grade to different membership functions has been used frequently. At the FCM algorithm, clusters are represented as the membership functions, and data points are assigned to several clusters. The membership degree to an input belonging to a particular cluster is evaluated. In the end, a data point consists of multiple membership functions, and the objective function is computed by the Euclidian distance [54,55]. Fuzzy C-means is a well-known clustering algorithm that is widely applied for segmenting and classifying medical images [56]. MATLAB commands were used to create the FCM classifier. Commands “fcm” and “genfis” were used to create the clusters [37]. The command “evalfis” was used for the test data [37]. Due to the nature of the method, the test data are labeled from a range (in Figure 7, the labels are extracted as “Y”). The ROC curve is extracted, and the optimum threshold is found (in Figure 7, T is the threshold). If a label of data is greater than the threshold, Y = 1; otherwise, Y = 0. As conducted at the K-means method, two groups are created (“G1” and “G2” in Figure 7) with labels “0” and “1”, without knowing which is the malignant or benign. Two truth tables are created, the first using the labeled data with “0” as malignant and the second truth table with “1” as malignant. Two accuracies are extracted (“Acc1” and “Acc2” in Figure 7). We use the truth table of the accuracy that is greater than the other.

2.6.3. ANFIS

ANFIS is a hybrid intelligent system that includes a TSK (Tagaki–Sugeno–Kang) model with a multi-layer neural network structure. The parameters of the TSK systems are determined by learning methodologies. Learning methodology (back-propagation method) automatically defines the parameters of fuzzy sets, such as membership functions and rules, without an operator’s intervention. The difference between ANFIS and neural networks is that the neural networks are created by training while ANFIS is formed by fuzzy rules, and then the learning method optimizes the membership functions and rule parameters [19,40,41]. Although ANFIS shows that it has a problem with the curse of dimensionality, it has a significant improvement in classification accuracy [57].
ANFIS is based on the Sugeno fuzzy model, and the structure can be represented by the following first-order rules:
R u l e   1 : I f   x   i s   A 1   a n d   y   i s   B 1 T h e n   f 1 = p 1 x + q 1 y + r 1 R u l e   2 : I f   x   i s   A 2   a n d   y   i s   B 2 T h e n   f 2 = p 2 x + q 2 y + r 2
where x and y are the inputs; A i and B i are the fuzzy sets; and f i are the outputs set by the fuzzy rules with parameters p i , q i and r i .
The parameters of fuzzy sets and the p i , q i and r i are those determined by the learning process. The learning method is hybrid as it uses the least squares technique and back propagation. The process is performed in two passes. The first pass is the forward direction; the fuzzy set parameters are kept constant while the parameters p i , q i and r i are determined using the least squares technique. The output y of ANFIS is extracted by using the new rule values. The desired output with the new output has an error, which is calculated by subtracting them from each other. The next pass is the back propagation, where the parameters of the fuzzy sets are updated. The parameters of the rules remain constant. Figure 8 shows an ANFIS system with three inputs, two fuzzy rules at each input and one output [41,58]. MATLAB’s command “anfis” was used for the training process and “evalfis” for the testing process [37].
The general procedure is shown in Figure 9.

2.7. Evaluation Metrics

For the evaluation of the classifiers, five different performance measures are applied. The measures are the accuracy, sensitivity, specificity, F1 score and AUC of the results. The measures are known by finding the number of true/false positive and true/false negative outputs:
  • True positive (TP) denotes the malignancy correctly predicted as malignancy.
  • True negative (TN) denotes the benignity correctly predicted as benignity.
  • False positive (FP) denotes the benignity mistakenly predicted as malignancy.
  • False negative (FN) denotes the malignancy mistakenly predicted as benignity.
The measures are defined as:
  • S e n s i t i v i t y = T P T P + F N × 100 %
  • S p e c i f i c i t y = T N T N + F P × 100 %
  • A c c u r a c y = T P + T N T P + T N + F P + F N × 100 %
  • F 1   s c o r e = 2 × T P T P + F P + F N × 100 %
Sensitivity is the ability of the classifier to detect that malignancy is present, given that malignancy is present.
Specificity is the ability of the classifier to detect that benignity exists, given that benignity exists.
Accuracy is the ability of the classifier to correctly detect benignity and malignancy.
The criterion to find the best combination of the features is to find the highest accuracy of the system. Once the highest accuracy is found, then the sensitivity and specificity are calculated [41,59].
For the classifiers FCM and ANFIS, ROC curves were extracted, helping to extract the area under the ROC curve (AUC) that is an objective index [33]. AUC has values between 0 and 1. The closer to 1, the better the classifier. ROC curve can enhance the optimum performance of a classifier only when the classifier has as an output a range of numbers. ROC curve tests different thresholds for classifying the range of the outputs and suggests the best threshold value. The K-means method has as an output only two values if the clusters are two, and the ROC curve cannot be applied to it.
The F1 score also provides an objective evaluation of a classifier [30,33].

2.8. Mammography Images (Dataset)

Direct digital mammograms with 97 masses, 65 benign and 32 malignant, were used for training and testing the classifier. This is an in-house FFDM database, the images of which were acquired with a Senographe 2000D or a Senographe Essential FFDM system (GE Healthcare, Chicago, IL, USA). The for-presentation images were used in this study that are either 1914 × 2294 or 2394 × 3062 pixels with 16 bits per pixel (12 bits were allocated). All but nine of the masses came from different mammograms. Nine of the masses were observed on both the mediolateral oblique and the craniocaudal view of the same mammogram and were considered 18 independent findings. Regions of interest (ROIs) were extracted from the full mammograms based on the ground truth files defined by expert mammographers. Figure 10 shows a representative example of ROI extraction. ROIs varied in size because each one was adapted to the mass size so as to encompass the mass and the local background (red circle in Figure 10).
The FFDM dataset was provided by Dr. Maria Kallergi and the Smart Hospital Research Laboratory of the Department of Biomedical Engineering of the University of West Attica. All cases were confirmed by biopsy or annual follow-up for at least three years. This is a secondary use of unidentified and non-coded data, and the work does not constitute research with human subjects because there was no interaction with any individual, and no identifiable private information was used. This work was IRB-exempt.

3. Results

To extract the results of the methods, we used computer resources with the following characteristics: Dell PE 350, CPU: Intel Xeon E-2314 2.8 GHz, 8 M Cache, DELL Memory 16 GB—DDR4 UDIMM 3200 MHz, Windows server 2019 standard edition 16 Cores 2 VMs. At the ROIs of the mammograms, median filtering and the methods that were proposed in our previous work were performed. The methods provided seven enhanced images, LH-CON, LH-XWW, IFS, PFS, FFS, OWA-GA and OWA-HK, with one initial ROI mammography as an origin [3].
Furthermore, in this paper, we performed thresholding, feature extraction, feature selection and classification of the mammograms. Each enhanced image is thresholded by three different methods of thresholding. Global, Otsu and type-2 fuzzy sets thresholding provide a total of twenty-one thresholded images. The images that we created with the initial images will provide the features that we need for the classifier. Geometrical, texture and fuzzy features are extracted according to the feature extraction section of this paper. This procedure happens to all the mammograms of the dataset. At the end of this step, there will be three three-dimension matrixes. The first matrix has rows of the ninety-seven mammograms; c columns of the nine features that are extracted from each mammography; and, as the depth of the matrix, the seven enhancement methods. The first three-dimensional matrix is about the features that are extracted with the method of global thresholding, the second three-dimensional matrix is about the method of Otsu thresholding and the third three-dimensional matrix is about the method of type-2 fuzzy sets thresholding. The data of the three matrixed are normalized at this stage. The normalization is performed on each feature of the mammograms. The feature selection is the exhaustive search where all the combinations of the features are tested with the K-means, FCM and ANFIS classifiers so the best combination can be found.
K-means, FCM and ANFIS classifiers have two up to nine inputs that depend on the combination that the exhaustive search gives. For the ANFIS classifier, we used two membership functions at the input, type “gbell”, and the rules are constant.
The dataset was divided into training and testing sets. The training set was 70% of the data, and the testing set was 30%. The classifiers were trained first with the training set, and the performance indicators were recorded from the testing set. The sets were randomly selected. The performance of the classifiers was tested based on the accuracy. The classification procedure was performed ten times so it can be statistically more accurate, and the results for each classifier are the average of the ten results. The following Table 1, Table 2 and Table 3 show the results. Table 1 shows the results of the global thresholding method, Table 2 shows the results of Otsu thresholding and Table 3 shows the type-2 fuzzy sets thresholding method. For each line, the first subline is the results of K-means clustering, the second subline of FCM and the third subline of ANFIS.
Based on Table 1, the enhancement method FFS with global thresholding and the FCM classifier scored the highest indicator values with an accuracy of 95.17%, F1 score of 86.98% and AUC of 0.91 (ROC curve in Figure 11a).
Based on Table 2, the enhancement method LH-XWW with Otsu thresholding and FCM classifier scored the highest indicator values with an accuracy of 95.17%, F1 score of 89.42% and AUC of 0.91 (ROC curve in Figure 11b).
Based on Table 3, the enhancement method LH-XWW with type-2 fuzzy sets thresholding and FCM classifier scored the highest indicator values with an accuracy of 94.14% and an AUC of 0.89 (ROC curve in Figure 11c). The best F1 score for type-2 fuzzy sets is achieved with the enhancement method OWA-HK and the classifier FCM, with a score of 86.22% (ROC curve in Figure 11d).
In general, the best scores were achieved by the FCM classifier. The highest accuracies and AUC were achieved by FFS enhancement, global thresholding, FCM classifier and LH-XWW enhancement, Otsu thresholding and FCM classifier, but the second combination reached a higher F1 score.
Advanced fuzzy sets (IFS, PFS and FFS) yielded satisfactory results, especially with global and Otsu thresholding methods.

4. Discussion

In this paper, we proposed the classification of breast masses as benign or malignant with an intelligent system that includes enhancement, thresholding, feature selection and classification with state-of-the-art methodologies [3]. The pilot evaluation was performed with 97 masses from FFDM for presentation images. Geometrical, texture and fuzzy features were computed for each mass ROI and used as inputs to the classifiers.
FCM and ANFIS classifiers achieved the best results compared to the K-means method. The best combination of methods overall was the LH-XWW enhancement with Otsu thresholding and FCM classifier, achieving an accuracy of 95.17%, F1 score of 89.42% and AUC of 0.91. The second best combination was the FFS enhancement with global thresholding and FCM classifier. The conventional K-means method was suggested as a simple method to classify the data but had the poorest results, leaving no room for future improvement. Otsu thresholding is usually considered a conventional method to threshold an image with average results. In the FFDM data of our study, however, it led to highly accurate classification when combined with the other proposed methodologies.

5. Conclusions

The benign/malignant classification of mammographic lesions depends on the radiologists’ experience and the limitations of mammography. FFDM brought significant improvements in breast imaging compared to SFM, particularly for dense breasts. However, it has not alleviated all issues. As a result, numerous biopsies are still performed annually for the diagnosis of breast cancer, and only about 20% of these biopsies yield a positive outcome [60]. The field of automatic detection of cancerous masses is a well-researched scientific topic for digitized SFM, but there is still room for improvement for FFDM.
This paper presented a pattern recognition methodology for FFDM data that included advanced fuzzy sets for contrast enhancement and a hybrid FCM/K-means clustering with a decision tree for classification. A pilot evaluation demonstrated promising performance, matching reports on digitized SFM at worst; classification accuracy was consistently above 90% independent of the enhancement or the thresholding approach.
Future work will include deep learning and the expansion of the digital image dataset for more efficient training and, furthermore, testing of the proposed intelligent system with additional features specific to the various mass types.

Author Contributions

Conceptualization, A.D. and A.-N.A.; methodology, A.D. and A.-N.A.; software, A.-N.A.; validation, A.D. and A.-N.A.; formal analysis, A.D. and A.-N.A.; investigation, A.D. and A.-N.A.; resources, A.D.; data curation, A.D., A.-N.A. and M.K.; writing—original draft preparation, A.D., A.-N.A. and M.K.; writing—review and editing, A.D., A.-N.A. and M.K.; visualization, A.D. and A.-N.A.; supervision, A.D.; project administration, A.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. American Cancer Society Breast Cancer Statistics|How Common Is Breast Cancer? Available online: https://www.cancer.org/cancer/types/breast-cancer/about/how-common-is-breast-cancer.html (accessed on 20 August 2023).
  2. Hall, J. Large Mammography Study Shows Significant Benefits with AI-Aided Screening. Available online: https://www.diagnosticimaging.com/view/large-mammography-study-significant-benefits-with-ai-aided-screening?utm_source=sfmc&utm_medium=email&utm_campaign=08062023_DI_eNL_unsupported&eKey=a2FsbGVyZ2lAdGVpYXRoLmdy (accessed on 20 August 2023).
  3. Dounis, A.; Avramopoulos, A.-N.; Kallergi, M. Advanced Fuzzy Sets and Genetic Algorithm Optimizer for Mammographic Image Enhancement. Electronics 2023, 12, 3269. [Google Scholar] [CrossRef]
  4. Premalatha, R.; Dhanalakshmi, P. Enhancement and Segmentation of Medical Images through Pythagorean Fuzzy Sets-An Innovative Approach. Neural Comput. Appl. 2022, 34, 11553–11569. [Google Scholar] [CrossRef] [PubMed]
  5. Otsu, N. A Threshold Selection Method from Gray-Level Histograms. IEEE Trans. Syst. Man Cybern. 1979, 9, 62–66. [Google Scholar] [CrossRef]
  6. Basir, O.; Zhu, H.; Karray, F. Fuzzy Based Image Segmentation. In Fuzzy Filters for Image Processing; Nachtegael, M., Van Der Weken, D., Kerre, E.E., Van De Ville, D., Eds.; Studies in Fuzziness and Soft Computing; Springer: Berlin/Heidelberg, Germany, 2003; Volume 122, pp. 101–128. ISBN 978-3-642-05591-1. [Google Scholar]
  7. Chowdhary, C.L.; Mittal, M.P.K.; Pattanaik, P.A.; Marszalek, Z. An Efficient Segmentation and Classification System in Medical Images Using Intuitionist Possibilistic Fuzzy C-Mean Clustering and Fuzzy SVM Algorithm. Sensors 2020, 20, 3903. [Google Scholar] [CrossRef] [PubMed]
  8. Tariq, M.; Iqbal, S.; Ayesha, H.; Abbas, I.; Ahmad, K.T.; Niazi, M.F.K. Medical Image Based Breast Cancer Diagnosis: State of the Art and Future Directions. Expert Syst. Appl. 2021, 167, 114095. [Google Scholar] [CrossRef]
  9. Hassanat, A.B.A.; Alkasassbeh, M.; Al-awadi, M.; Alhasanat, E.A.A. Color-Based Object Segmentation Method Using Artificial Neural Network. Simul. Model. Pract. Theory 2016, 64, 3–17. [Google Scholar] [CrossRef]
  10. Hassanat, A.B.A.; Alkasassbeh, M.; Al-awadi, M.; Alhasanat, E.A.A. Colour-Based Lips Segmentation Method Using Artificial Neural Networks. In Proceedings of the 2015 6th International Conference on Information and Communication Systems (ICICS), Amman, Jordan, 7–9 April 2015; pp. 188–193. [Google Scholar]
  11. Yazdanbakhsh, O.; Dick, S. A Deep Neuro-Fuzzy Network for Image Classification. arXiv 2020, arXiv:2001.01686. [Google Scholar] [CrossRef]
  12. Gao, Y.; Lin, J.; Zhou, Y.; Lin, R. The Application of Traditional Machine Learning and Deep Learning Techniques in Mammography: A Review. Front. Oncol. 2023, 13, 1213045. [Google Scholar] [CrossRef]
  13. Almutairi, S.; Manimurugan, S.; Kim, B.G.; Aborokbah, M.M.; Narmatha, C. Breast Cancer Classification Using Deep Q Learning (DQL) and Gorilla Troops Optimization (GTO). Appl. Soft Comput. 2023, 142, 110292. [Google Scholar] [CrossRef]
  14. Malebary, S.J.; Hashmi, A. Automated Breast Mass Classification System Using Deep Learning and Ensemble Learning in Digital Mammogram. IEEE Access 2021, 9, 55312–55328. [Google Scholar] [CrossRef]
  15. Li, H.; Chen, D.; Nailon, W.H.; Davies, M.E.; Laurenson, D.I. Dual Convolutional Neural Networks for Breast Mass Segmentation and Diagnosis in Mammography. IEEE Trans. Med. Imaging 2022, 41, 3–13. [Google Scholar] [CrossRef] [PubMed]
  16. Kaur, A.; Rashid, M.; Bashir, A.K.; Parah, S.A. Detection of Breast Cancer Masses in Mammogram Images with Watershed Segmentation and Machine Learning Approach. In Artificial Intelligence for Innovative Healthcare Informatics; Parah, S.A., Rashid, M., Varadarajan, V., Eds.; Springer International Publishing: Cham, Switzerland, 2022; pp. 35–60. ISBN 978-3-030-96568-6. [Google Scholar]
  17. Sarvestani, Z.M.; Jamali, J.; Taghizadeh, M.; Dindarloo, M.H.F. A Novel Machine Learning Approach on Texture Analysis for Automatic Breast Microcalcification Diagnosis Classification of Mammogram Images. J. Cancer Res. Clin. Oncol. 2023, 149, 6151–6170. [Google Scholar] [CrossRef] [PubMed]
  18. Mahersia, H.; Boulehmi, H.; Hamrouni, K. Development of Intelligent Systems Based on Bayesian Regularization Network and Neuro-Fuzzy Models for Mass Detection in Mammograms: A Comparative Analysis. Comput. Methods Programs Biomed. 2016, 126, 46–62. [Google Scholar] [CrossRef] [PubMed]
  19. Hosseini, M.S.; Zekri, M. Review of Medical Image Classification Using the Adaptive Neuro-Fuzzy Inference System. J. Med. Signals Sens. 2012, 2, 49–60. [Google Scholar] [CrossRef] [PubMed]
  20. Minaee, S.; Boykov, Y.Y.; Porikli, F.; Plaza, A.J.; Kehtarnavaz, N.; Terzopoulos, D. Image Segmentation Using Deep Learning: A Survey. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 3523–3542. [Google Scholar] [CrossRef] [PubMed]
  21. Alkassar, S.; Jebur, B.A.; Abdullah, M.A.M.; Al-Khalidy, J.H.; Chambers, J.A. Going Deeper: Magnification-invariant Approach for Breast Cancer Classification Using Histopathological Images. IET Comput. Vis. 2021, 15, 151–164. [Google Scholar] [CrossRef]
  22. Abdullah, M.A.M.; Alkassar, S.; Jebur, B.; Chambers, J. LBTS-Net: A Fast and Accurate CNN Model for Brain Tumour Segmentation. Healthc. Tech Lett. 2021, 8, 31–36. [Google Scholar] [CrossRef] [PubMed]
  23. Abdel Rahman, A.S.; Belhaouari, S.B.; Bouzerdoum, A.; Baali, H.; Alam, T.; Eldaraa, A.M. Breast Mass Tumor Classification Using Deep Learning. In Proceedings of the 2020 IEEE International Conference on Informatics, IoT, and Enabling Technologies (ICIoT), Doha, Qatar, 2–5 February 2020; pp. 271–276. [Google Scholar]
  24. Salama, W.M.; Aly, M.H. Deep Learning in Mammography Images Segmentation and Classification: Automated CNN Approach. Alex. Eng. J. 2021, 60, 4701–4709. [Google Scholar] [CrossRef]
  25. Nemade, V.; Pathak, S.; Dubey, A.K. Deep Learning-Based Ensemble Model for Classification of Breast Cancer. Microsyst. Technol. 2023. [Google Scholar] [CrossRef]
  26. Lin, Q.; Chen, X.; Chen, C.; Garibaldi, J.M. A Novel Quality Control Algorithm for Medical Image Segmentation Based on Fuzzy Uncertainty. IEEE Trans. Fuzzy Syst. 2023, 31, 2532–2544. [Google Scholar] [CrossRef]
  27. Salih, A.M.; Kamil, M.Y. Mammography Images Segmentation Based on Fuzzy Set and Thresholding. Malays. J. Sci. 2019, 29, 168–175. [Google Scholar] [CrossRef]
  28. Chaira, T. Intuitionistic Fuzzy Segmentation of Medical Images. IEEE Trans. Biomed. Eng. 2010, 57, 1430–1436. [Google Scholar] [CrossRef] [PubMed]
  29. Tahoun, M.; Almazroi, A.A.; Alqarni, M.A.; Gaber, T.; Mahmoud, E.E.; Eltoukhy, M.M. A Grey Wolf-Based Method for Mammographic Mass Classification. Appl. Sci. 2020, 10, 8422. [Google Scholar] [CrossRef]
  30. Castro-Tapia, S.; Castañeda-Miranda, C.L.; Olvera-Olvera, C.A.; Guerrero-Osuna, H.A.; Ortiz-Rodriguez, J.M.; Martínez-Blanco, M.D.R.; Díaz-Florez, G.; Mendiola-Santibañez, J.D.; Solís-Sánchez, L.O. Classification of Breast Cancer in Mammograms with Deep Learning Adding a Fifth Class. Appl. Sci. 2021, 11, 11398. [Google Scholar] [CrossRef]
  31. Mobark, N.; Hamad, S.; Rida, S.Z. CoroNet: Deep Neural Network-Based End-to-End Training for Breast Cancer Diagnosis. Appl. Sci. 2022, 12, 7080. [Google Scholar] [CrossRef]
  32. Alshehri, A.; AlSaeed, D. Breast Cancer Detection in Thermography Using Convolutional Neural Networks (CNNs) with Deep Attention Mechanisms. Appl. Sci. 2022, 12, 12922. [Google Scholar] [CrossRef]
  33. Du, G.; Dong, M.; Sun, Y.; Li, S.; Mu, X.; Wei, H.; Ma, L.; Liu, B. A New Method for Detecting Architectural Distortion in Mammograms by NonSubsampled Contourlet Transform and Improved PCNN. Appl. Sci. 2019, 9, 4916. [Google Scholar] [CrossRef]
  34. Deng, H.; Deng, W.; Sun, X.; Liu, M.; Ye, C.; Zhou, X. Mammogram Enhancement Using Intuitionistic Fuzzy Sets. IEEE Trans. Biomed. Eng. 2017, 64, 1803–1814. [Google Scholar] [CrossRef]
  35. Yousefi, J. Image Binarization Using Otsu Thresholding Algorithm; University of Guelph: Guelph, ON, Canada, 2011; p. 10. [Google Scholar]
  36. Xu, X.; Xu, S.; Jin, L.; Song, E. Characteristic Analysis of Otsu Threshold and Its Applications. Pattern Recognit. Lett. 2011, 32, 956–961. [Google Scholar] [CrossRef]
  37. MathWorks MATLAB for Artificial Intelligence. Available online: https://www.mathworks.com/ (accessed on 8 August 2023).
  38. Tizhoosh, H.R. Image Thresholding Using Type II Fuzzy Sets. Pattern Recognit. 2005, 38, 2363–2372. [Google Scholar] [CrossRef]
  39. Mohanaiah, P.; Sathyanarayana, P.; GuruKumar, L. Image Texture Feature Extraction Using GLCM Approach. Int. J. Sci. Res. Publ. 2013, 3, 290–294. [Google Scholar]
  40. Kalam, R.; Thomas, C.; Rahiman, M.A. Brain Tumor Detection in MRI Images Using Adaptive-ANFIS Classifier with Segmentation of Tumor and Edema. Soft Comput. 2023, 27, 2279–2297. [Google Scholar] [CrossRef]
  41. Hossain, E.; Rahaman, M.A. Bone Cancer Detection & Classification Using Fuzzy Clustering & Neuro Fuzzy Classifier. In Proceedings of the 2018 4th International Conference on Electrical Engineering and Information & Communication Technology (iCEEiCT), Dhaka, Bangladesh, 13–15 September 2018; pp. 541–546. [Google Scholar]
  42. Li, W.; Lu, C.; Zhang, J. A Lower Envelope Weber Contrast Detection Algorithm for Steel Bar Surface Pit Defects. Opt. Laser Technol. 2013, 45, 654–659. [Google Scholar] [CrossRef]
  43. Peli, E. Contrast in Complex Images. J. Opt. Soc. Am. A 1990, 7, 2032. [Google Scholar] [CrossRef] [PubMed]
  44. Pal, S.K.; Rosenfeld, A. Image Enhancement and Thresholding by Optimization of Fuzzy Compactness. Pattern Recognit. Lett. 1988, 7, 77–86. [Google Scholar] [CrossRef]
  45. Jähne, B.; Haussecker, H.; Geissler, P. Handbook of Computer Vision and Applications; Academic Press, Inc.: Orlando, FL, USA, 1999; Volume 2. [Google Scholar]
  46. Peshawa, J.M.A. Rezhna Hassan Faraj Data Normalization and Standardization: A Technical Report. Mach. Learn. Tech. Rep. 2014, 1, 1–6. [Google Scholar] [CrossRef]
  47. Pérez, N.P.; Guevara López, M.A.; Silva, A.; Ramos, I. Improving the Mann–Whitney Statistical Test for Feature Selection: An Approach in Breast Cancer Diagnosis on Mammography. Artif. Intell. Med. 2015, 63, 19–31. [Google Scholar] [CrossRef]
  48. Nievergelt, J. Exhaustive Search, Combinatorial Optimization and Enumeration: Exploring the Potential of Raw Computing Power. In SOFSEM 2000: Theory and Practice of Informatics; Hlaváč, V., Jeffery, K.G., Wiedermann, J., Eds.; Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2000; Volume 1963, pp. 18–35. ISBN 978-3-540-41348-6. [Google Scholar]
  49. Har-Peled, S.; Sadri, B. How Fast Is the K-Means Method? Algorithmica 2005, 41, 185–202. [Google Scholar] [CrossRef]
  50. Arthur, D.; Vassilvitskii, S. How Slow Is the K-Means Method? In Proceedings of the twenty-second annual symposium on Computational geometry, Sedona, AZ, USA, 5 June 2006; ACM: Sedona, AZ, USA; pp. 144–153. [Google Scholar]
  51. Khouj, Y.; Dawson, J.; Coad, J.; Vona-Davis, L. Hyperspectral Imaging and K-Means Classification for Histologic Evaluation of Ductal Carcinoma In Situ. Front. Oncol. 2018, 8, 17. [Google Scholar] [CrossRef]
  52. Kharrat, A.; Benamrane, N.; Messaoud, M.B.; Abid, M. Detection of Brain Tumor in Medical Images. In Proceedings of the 2009 3rd International Conference on Signals, Circuits and Systems (SCS), Medenine, Tunisia, 6–8 November 2009; pp. 1–6. [Google Scholar]
  53. Liu, Y.; Li, W.; Li, Y. Network Traffic Classification Using K-Means Clustering. In Proceedings of the Second International Multi-Symposiums on Computer and Computational Sciences (IMSCCS 2007), Iowa City, IA, USA, 13–15 August 2007; pp. 360–365. [Google Scholar]
  54. Nayak, J.; Naik, B.; Behera, H.S. Fuzzy C-Means (FCM) Clustering Algorithm: A Decade Review from 2000 to 2014. In Computational Intelligence in Data Mining—Volume 2; Jain, L.C., Behera, H.S., Mandal, J.K., Mohapatra, D.P., Eds.; Smart Innovation; Systems and Technologies; Springer: New Delhi, India, 2015; Volume 32, pp. 133–149. ISBN 978-81-322-2207-1. [Google Scholar]
  55. Budayan, C.; Dikmen, I.; Birgonul, M.T. Comparing the Performance of Traditional Cluster Analysis, Self-Organizing Maps and Fuzzy C-Means Method for Strategic Grouping. Expert Syst. Appl. 2009, 36, 11772–11781. [Google Scholar] [CrossRef]
  56. Aruna Kumar, S.V.; Harish, B.S. A Modified Intuitionistic Fuzzy Clustering Algorithm for Medical Image Segmentation. J. Intell. Syst. 2018, 27, 593–607. [Google Scholar] [CrossRef]
  57. Nguyen, T.-L.; Kavuri, S.; Park, S.-Y.; Lee, M. Attentive Hierarchical ANFIS with Interpretability for Cancer Diagnostic. Expert Syst. Appl. 2022, 201, 117099. [Google Scholar] [CrossRef]
  58. Cetisli, B. Development of an Adaptive Neuro-Fuzzy Classifier Using Linguistic Hedges: Part 1. Expert Syst. Appl. 2010, 37, 6093–6101. [Google Scholar] [CrossRef]
  59. Wong, H.B.; Lim, G.H. Measures of Diagnostic Accuracy: Sensitivity, Specificity, PPV and NPV. Proc. Singap. Healthc. 2011, 20, 316–318. [Google Scholar] [CrossRef]
  60. Effective Health Care Program, Agency for Healthcare Research and Quality Systematic Review: Core Needle and Open Surgical Biopsy for Diagnosis of Breast Lesions: An Update to the 2009 Report. Available online: https://effectivehealthcare.ahrq.gov/products/breast-biopsy-update/research (accessed on 20 August 2023).
Figure 1. Tizhoosh membership function and lower and upper function with L = 4 and U = 0.25 .
Figure 1. Tizhoosh membership function and lower and upper function with L = 4 and U = 0.25 .
Applsci 13 10401 g001
Figure 2. Images of the three threshold methods.
Figure 2. Images of the three threshold methods.
Applsci 13 10401 g002
Figure 3. Process of creating a single object without holes.
Figure 3. Process of creating a single object without holes.
Applsci 13 10401 g003
Figure 4. Mass segmentation.
Figure 4. Mass segmentation.
Applsci 13 10401 g004
Figure 5. GLCM calculation.
Figure 5. GLCM calculation.
Applsci 13 10401 g005
Figure 6. Classification using decision tree based on K-means.
Figure 6. Classification using decision tree based on K-means.
Applsci 13 10401 g006
Figure 7. Classification using decision tree based on FCM.
Figure 7. Classification using decision tree based on FCM.
Applsci 13 10401 g007
Figure 8. ANFIS system.
Figure 8. ANFIS system.
Applsci 13 10401 g008
Figure 9. Mass Classification.
Figure 9. Mass Classification.
Applsci 13 10401 g009
Figure 10. Extraction of extracted ROI [3].
Figure 10. Extraction of extracted ROI [3].
Applsci 13 10401 g010
Figure 11. Average of the 10 repetitions ROC curves of the combinations that performed the best: (a) FFS enhancement, global thresholding, FCM classifier with accuracy 95.17%, F1 score 86.98%, AUC 0.91; (b) LH-XWW enhancement, Otsu thresholding, FCM classifier with accuracy 95.17%, F1 score 89.42%, AUC 0.91; (c) LH-XWW enhancement, type-2 thresholding, FCM classifier with accuracy 94.14%, AUC 0.89; (d) OWA-HK enhancement, type-2 thresholding, FCM classifier with F1 score 86.22%.
Figure 11. Average of the 10 repetitions ROC curves of the combinations that performed the best: (a) FFS enhancement, global thresholding, FCM classifier with accuracy 95.17%, F1 score 86.98%, AUC 0.91; (b) LH-XWW enhancement, Otsu thresholding, FCM classifier with accuracy 95.17%, F1 score 89.42%, AUC 0.91; (c) LH-XWW enhancement, type-2 thresholding, FCM classifier with accuracy 94.14%, AUC 0.89; (d) OWA-HK enhancement, type-2 thresholding, FCM classifier with F1 score 86.22%.
Applsci 13 10401 g011aApplsci 13 10401 g011b
Table 1. Indicators for global thresholding method (The best scores are bold).
Table 1. Indicators for global thresholding method (The best scores are bold).
Contrast Methods Accuracy %Sensitivity %Specificity %F1 Score %AUC
LH-CONK-means83.796191.4664.61-
FCM93.7976.8498.3820.86
ANFIS92.0771.1298.6879.910.83
LH-XWWK-means87.5974.9492.7377.2-
FCM94.1479.4498.2185.270.84
ANFIS93.4584.1996.3785.470.88
IFSK-means81.3876.5483.170.58-
FCM92.4178.7796.8882.90.87
ANFIS92.0771.596.7878.460.81
PFSK-means80.3467.3485.4565.26-
FCM92.4172.996.9877.870.8
ANFIS93.4582.697.2286.330.86
FFSK-means86.9 77.6890.1177.39-
FCM95.1783.3598.286.980.91
ANFIS93.7980.6197.7583.750.85
OWA-GAK-means83.79 77.7285.0570.61-
FCM91.3874.1496.0477.40.82
ANFIS92.0764.3998.7774.760.75
OWA-HKK-means84.83 69.3690.7771.39-
FCM92.7673.989777.830.81
ANFIS92.4163.9599.0775.570.76
Table 2. Indicators for Otsu thresholding method (The best scores are bold).
Table 2. Indicators for Otsu thresholding method (The best scores are bold).
Contrast Methods Accuracy %Sensitivity %Specificity %F1 Score %AUC
LH-CONK-means86.5579.0388.775.78-
FCM92.7682.899683.890.9
ANFIS91.0361.5398.2172.170.73
LH-XWWK-means87.2466.9293.1169.24-
FCM95.1788.4197.3289.420.91
ANFIS90.6972.3496.0276.090.8
IFSK-means80.3465.3185.0163.76-
FCM92.7673.7997.8179.390.83
ANFIS91.0367.6396.4373.760.77
PFSK-means81.3870.0983.7464.25-
FCM92.7673.9697.0179.740.79
ANFIS93.180.2697.5985.980.86
FFSK-means80.6974.4183.5965.43-
FCM91.3874.0697.0781.120.8
ANFIS92.0769.2697.8276.770.75
OWA-GAK-means81.0382.4579.6473.22-
FCM92.0779.8395.982.290.87
ANFIS92.4164.7998.7175.390.77
OWA-HKK-means78.6243.3189.6645.84-
FCM92.7674.8297.381.640.84
ANFIS92.4170.2598.6779.830.83
Table 3. Indicators for type-2 fuzzy sets thresholding method (The best scores are bold).
Table 3. Indicators for type-2 fuzzy sets thresholding method (The best scores are bold).
Contrast Methods Accuracy %Sensitivity %Specificity %F1 Score %AUC
LH-CONK-means86.5563.6893.1564.81-
FCM92.4173.7397.2979.960.84
ANFIS90.3472.0795.2876.80.82
LH-XWWK-means89.3170.0596.3776.44-
FCM94.1483.5496.9983.670.89
ANFIS91.7279.3995.3581.340.86
IFSK-means82.4170.7187.0865.66-
FCM91.7269.8997.2977.460.81
ANFIS90.6967.8995.8873.190.75
PFSK-means82.4176.2785.5172.46-
FCM93.4573.5997.8779.840.82
ANFIS92.0782.5295.2683.710.86
FFSK-means85.8673.2691.0574.87-
FCM91.7281.1994.6982.360.84
ANFIS92.0775.1596.2176.490.83
OWA-GAK-means85.1783.9784.9877.77-
FCM92.7684.7194.5182.830.87
ANFIS90.6970.9695.7473.540.79
OWA-HKK-means85.1770.6391.0769.6-
FCM93.184.7996.386.220.88
ANFIS90.6966.2398.2676.480.76
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Dounis, A.; Avramopoulos, A.-N.; Kallergi, M. Hybrid Intelligent Pattern Recognition Systems for Mass Segmentation and Classification: A Pilot Study on Full-Field Digital Mammograms. Appl. Sci. 2023, 13, 10401. https://doi.org/10.3390/app131810401

AMA Style

Dounis A, Avramopoulos A-N, Kallergi M. Hybrid Intelligent Pattern Recognition Systems for Mass Segmentation and Classification: A Pilot Study on Full-Field Digital Mammograms. Applied Sciences. 2023; 13(18):10401. https://doi.org/10.3390/app131810401

Chicago/Turabian Style

Dounis, Anastasios, Andreas-Nestor Avramopoulos, and Maria Kallergi. 2023. "Hybrid Intelligent Pattern Recognition Systems for Mass Segmentation and Classification: A Pilot Study on Full-Field Digital Mammograms" Applied Sciences 13, no. 18: 10401. https://doi.org/10.3390/app131810401

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop