Next Article in Journal
Integrating Virtual, Mixed, and Augmented Reality to Human–Robot Interaction Applications Using Game Engines: A Brief Review of Accessible Software Tools and Frameworks
Previous Article in Journal
Water Quality Evaluation and Prediction Based on a Combined Model
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multiscale Analysis for Improving Texture Classification

by
Steve Tsham Mpinda Ataky
1,
Diego Saqui
2,
Jonathan de Matos
1,
Alceu de Souza Britto Junior
3 and
Alessandro Lameiras Koerich
1,*
1
École de Technologie Supérieure, Université du Québec, 1100, rue Notre-Dame Ouest, Montreal, QC H3C 1K3, Canada
2
Instituto Federal do Sul de Minas Gerais, Muzambinho 37890-000, Brazil
3
Programa de Pósgraduação em Informática, Pontifícia Universidade Católica do Paraná, Curitiba 80215-901, Brazil
*
Author to whom correspondence should be addressed.
Appl. Sci. 2023, 13(3), 1291; https://doi.org/10.3390/app13031291
Submission received: 17 November 2022 / Revised: 30 December 2022 / Accepted: 13 January 2023 / Published: 18 January 2023
(This article belongs to the Section Computing and Artificial Intelligence)

Abstract

:
Information from an image occurs over multiple and distinct spatial scales. Image pyramid multiresolution representations are a useful data structure for image analysis and manipulation over a spectrum of spatial scales. This paper employs the Gaussian–Laplacian pyramid to separately treat different spatial frequency bands of a texture. First, we generate three images corresponding to three levels of the Gaussian–Laplacian pyramid for an input image to capture intrinsic details. Then, we aggregate features extracted from gray and color texture images using bioinspired texture descriptors, information-theoretic measures, gray-level co-occurrence matrix feature descriptors, and Haralick statistical feature descriptors into a single feature vector. Such an aggregation aims at producing features that characterize textures to their maximum extent, unlike employing each descriptor separately, which may lose some relevant textural information and reduce the classification performance. The experimental results on texture and histopathologic image datasets have shown the advantages of the proposed method compared to state-of-the-art approaches. Such findings emphasize the importance of multiscale image analysis and corroborate that the descriptors mentioned above are complementary.

1. Introduction

The definition of texture has different flavors in the literature on computer vision. A common one considers texture as changes in the image intensity that form specific repetitive patterns [1]. These patterns may result from the physical properties of the object’s surface (roughness) that provide different types of light reflection. A smooth surface reflects the light at a defined angle (specular reflection), while a rough surface reflects it in all directions (diffuse reflection). Although texture recognition is easy for human perception, automatic procedures are different, and this task sometimes requires complex computational techniques. In computer vision, texture analysis is of a notable role, and its basis is extracting relevant information from an image to characterize its texture. This process involves a set of algorithms and techniques. Since humans’ perception of texture is not affected by rotation, translation, and scale changes, any numerical image texture characterization should be invariant to those aspects and any monotonic transformation in pixel intensity.
Several image texture analysis approaches have been developed over the past few decades, researching various properties to characterize texture information within an image [2,3]. For example, gray-level co-occurrence matrix (GLCM) [4], local binary patterns (LBP) [5], Haralick descriptors [6], Markov random fields [7], wavelet transform [8], Gabor texture discriminator among others, and more recently, the bioinspired texture descriptor (BiT) [9] are some of the classical and novel approaches developed for such a purpose.
In addition, convolutional neural networks (CNNs) have attracted the interest of academics due to their performance in tasks such as object detection. However, they do not lend themselves particularly well to texture classification [10,11]. Andrearczyk and Whelan [10], for instance, introduced a basic texture CNN (T-CNN) architecture that pooled an energy measure at its final convolution layer and discarded the shape information often acquired by conventional CNNs. Although the results were positive, the trade-off between accuracy and complexity was not beneficial. Similarly, various CNN architectures for texture classification with good performance were reported in [11,12,13].
Texture descriptors are also widely used in medical and biological image analysis, particularly in histopathologic images (HIs), which contain a variety of textures. For instance, regions with a high/low nuclei concentration and stroma [14]. Several scholars have investigated many textural descriptors for HI classification, including GLCM, LBP, Gabor, and Haralick descriptors [15].
Even though the descriptors mentioned above have demonstrated significant discriminative strength alone, combining them may be a potential technique for providing a representation based on many intrinsic textural properties. To increase the performance of texture classification, we investigate the combination of texture descriptors such as BiT, information-theoretic measures, GLCM, and Haralick descriptors. For such an aim, the feature vector representing an image is constructed by concatenating the features provided by distinct descriptors. The rationale is to compensate for the potential loss associated with utilizing a single technique to describe the texture. This technique uses Gaussian–Laplacian pyramids (GLP), a helpful data structure for image analysis and processing, to represent a spectrum of spatial scales [16].
The central idea behind GLCM is to estimate a joint probability distribution P ( x 1 , x 2 ) for the grayscale values in an image, where x 1 is the grayscale value at any randomly chosen pixel in the image and x 2 is the grayscale value at another pixel that is at a specific distance d from the first pixel. In most cases, only a subset of the grayscale values in an image are used to build the GLCM matrix. To requantize the gray levels to four bits, it is usual practice to construct a 16 × 16 GLCM matrix for eight-bit grayscale images with 256 gray levels (that is, the value of the intensity at each pixel is an integer between 0 and 255, both ends included) for texture characterization. As a statistical tool for analyzing texture, the GLCM considers the interaction between pixels in space. It generates a GLCM and then extracts statistical data based on the frequency with which pairs of pixels with particular values, as well as in a specific spatial relationship, occur in an image, both of which provide insight into the texture of the image.
Haralick descriptors, which are statistical measures, are computed on an entire image. An image’s texture is quantified using measures such as entropy and the sum of variance. The picture bit depth, or the number of gray levels, may be reduced by a quantization procedure when creating a GLCM from an image or an area of interest (ROI). Although determining a bit depth might be difficult, many approaches have explored suitable bit depths for various uses. Image noise, the size of the picture or region of interest (ROI), and the actual image content all play a role in determining the optimal bit depth. The values of the Haralick descriptors are also extremely sensitive to the bit depth and the maximum and minimum values employed in the quantization.
The BiT descriptor has characteristics of both the GLCM and the LBP descriptors. While the BiT descriptor’s taxonomic indices approximate second-order statistics at the level of the gray levels, it also characterizes textures based on second-order statistical properties, such as comparing pixels and determining how a pixel at a specific location relates statistically to pixels at different locations. These indices are grounded in group analysis and allow one to probe the neighborhood of regions displaced from a reference location. The BiT descriptor also has some features in common with Gabor filters. Gabor filters try to characterize a texture at its various periodicities by analyzing the image therefrom. All that can be explored at this point are the immediate areas surrounding each pixel. These periodicity properties within a neighborhood can be used to identify regional differences in texture. For this reason, biologists employ phylogenetic trees with diversity indices (part of BiT) to evaluate the similarities and differences in the neighborhood behaviors of various species across geographic and geographically close regions. In addition, diversity indices based on species richness play a fundamental role in determining a textural image’s comprehensive behavior (local and global characterization), constituting a nondeterministic complex system.
In addition to those mentioned earlier, multiscale analysis is essential for characterizing texture. As a mathematical approach, it permits the examination of a problem or system at several scales or resolution levels. Many occurrences or processes in the natural world, as well as in engineering and other fields, occur over a range of scales, from the microscopic (e.g., molecular or atomic scales) to the very vast (e.g., planetary scales) (e.g., global or cosmic scales). For instance, the macroscopic behavior of materials and structures (such as the strength of a bridge or the rigidity of a metal) is frequently impacted by the microscopic qualities and interactions of their constituent molecules or atoms. Similarly, the behavior of biological systems, such as cell and organ function, is frequently controlled by the behavior and interactions of their constituent molecules and proteins at the molecular scale. The multiscale analysis permits the investigation of these phenomena and the comprehension of how the behavior at one scale influences the behavior at other sizes. This has many applications, including material design, biological system analysis, metrology, topographies, agriculture, soil, material, and the comprehension of complex physical and chemical processes. More details about the application of multiscale onto those fields can be found in [17,18,19].
The multiscale analysis is a technique in pattern recognition and image processing that analyzes an image or pattern at various scales or levels of detail. This benefits multiple applications, such as object identification, image categorization, and feature extraction. In texture analysis, for instance, multiscale approaches can extract characteristics insensitive to changes in scale. This can be useful for tasks such as material classification, where the observed texture of the material may vary depending on the scale.
Thus, we state that an approach that computes features firmly based on information theory, second-order statistics, biodiversity, and taxonomic indices plays a fundamental role in defining an image texture’s global and local behaviors. Furthermore, such statistics should bring forth the all-inclusive behavior of a texture at the multiscale level. Moreover, they should take advantage of invariance in rotation, permutation, and scale, as well as having a significant effect in texture characterization, which is valid not only for images solely composed of textures but also for images that contain additional structures in addition to textures.
The contribution of this paper is threefold: (i) the combination of BiT descriptors, information theory, Haralick, and GLCM descriptors for texture characterization; (ii) a better discriminating ability while using color and gray-scale features on different categories of images; (iii) a method for texture classification that represents the state of the art on challenging datasets.
This paper is organized into five sections. First, the multiresolution concepts used in this work are presented in Section 2. Section 3 presents the proposed approach. The experimental design, the datasets used, our results, and related discussion are presented in Section 4. The last section presents the conclusion and future work prospects.

2. Multiscale and Multiresolution Analysis

Multiscale and multiresolution analysis are related techniques that involve analyzing a problem or system at multiple scales or levels of resolution. However, they differ in the specific approach used to perform the analysis. The multiscale analysis involves studying a problem or system simultaneously at multiple scales or levels of detail and can be done using various techniques, such as scale space representation, multiscale features, or multiscale modeling. The multiscale analysis aims to understand how the behavior at one scale is influenced by the behavior at other scales and to identify patterns or features present at multiple scales. On the other hand, multiresolution analysis involves decomposing a problem or system into multiple scales or levels of resolution using techniques such as wavelet decomposition or pyramid decomposition. This allows one to analyze the problem or system at different scales separately and capture patterns or features present at different scales. The goal of multiresolution analysis is often to represent a problem or system compactly or to extract robust features to changes in scale.
In short, multiscale analysis involves studying a problem or system at multiple scales simultaneously. In contrast, multiresolution analysis decomposes a problem or system into various scales and analyzes them separately. Both techniques can help understand phenomena or processes that occur over a range of scales and for extracting features that are robust to changes in scale.

2.1. Multiscale Analysis

Multiscale analysis is often used in pattern recognition and texture analysis when the patterns or textures being analyzed are expected to vary in size or scale. This can be due to variations in the distance, orientation, or perspective from which the patterns or textures are observed or inherent variations in the patterns or textures themselves. For example, consider the texture classification task, where the goal is to identify the type of material present in an image based on the texture of the surface. In this case, it may be helpful to use a multiscale analysis to extract features that are robust to changes in scale, as the texture of a material may vary depending on the size of the region being analyzed.
Multiscale analysis can also be helpful in texture analysis when the texture exhibits patterns at multiple scales. For instance, a texture may have a fine-scale pattern of wrinkles or scratches and a coarser-scale pattern of grain or veins. By analyzing the texture at multiple scales, one can more accurately capture the full range of patterns present in the texture, which can be helpful in tasks such as material identification or classification. Several methods exist for performing a multiscale analysis in pattern recognition and texture analysis. Some common approaches include:
  • Scale-space representation: this involves representing the image or pattern as a scale function and then analyzing the resulting scale space to extract features or detect patterns;
  • Multiresolution analysis involves decomposing the image or pattern into multiple scales or resolutions using techniques such as wavelet decomposition or pyramid decomposition and then analyzing the different scales separately;
  • Multiscale features: this involves extracting features from the image or pattern at multiple scales using multiscale histograms or edge detection techniques.
It is essential to choose an appropriate method for the multiscale analysis based on the characteristics of the patterns or textures being analyzed and the specific task at hand.

2.2. Multiresolution

Humans usually see regions of similar textures, colors, or gray levels that combine to form objects when looking at an image. If objects are small or have low contrast, it may be necessary to examine them in high resolution; a coarser view is satisfactory if they are large or have high contrast. If both types of objects appear in an image, it can be helpful to analyze them in multiple resolutions. Changing the resolution can also lead to creating, deleting, or merging image features. Moreover, there is evidence that the human visual system processes visual information in a multiresolution way [20]. Furthermore, sensors can provide data in various resolutions, and multiresolution algorithms for image processing offer advantages from a computational point of view and are generally robust.
When analyzing an image, it can be helpful to break it down into separate parts so there is no loss of information. The pyramid theory provides ways to decompose images at multiple levels of resolution. It considers a collection of representations of an image in different spatial resolutions, stacked on top of each other, with the highest resolution image at the bottom of the stack and subsequent images appearing over it in descending order of resolution. Such a procedure generates a pyramid-like structure, as shown in Figure 1a.
The traditional procedure for obtaining a lower-resolution image is to perform low-pass filtering followed by sampling [21]. In signal processing and computer vision, a pyramid representation is the primary type of multiscale representation for computing image features on different scales. The pyramid is obtained by repeated smoothing and subsampling of an image. Different smoothing kernels have been brought forward for generating the pyramid representation, and the binomial one strikingly shows up as valuable and theoretically well-founded [22].
Accordingly, for a bidimensional image, the normalized binomial filter may be applied ( 1 / 4 , 1 / 2 , 1 / 4 ) in most cases twice or even more, along all spatial dimensions. The subsampling of the image by a factor of two follows, which leads to an efficient and compact multilevel representation. There are low-pass and band-pass pyramid types [23].
To develop filter-based representations by decomposing images into information on multiple scales and extracting features/structures of interest from an image, the Gaussian pyramid (GP), the Laplacian pyramid (LP), and the wavelet pyramid are examples of the most frequently used pyramids. The GP (Figure 1) consists of a low-pass filter, a reduced density, where subsequent images of the preceding level of the pyramid are weighted down using a Gaussian average or Gaussian blur and scaled down. The original image is defined as the base level. Formally speaking, assuming that I ( x , y ) is a two-dimensional image, the GP is recursively defined as presented in (1).
G 0 ( x , y ) = I ( x , y ) , for level , l = 0 m = 2 2 n = 2 2 w ( m , n ) G l 1 ( 2 x + m , 2 y + n ) , otherwise ,
where w ( m , n ) is a weighting function (identical at all levels) termed the generating kernel, which adheres to the following properties: it is separable, symmetric, and each node at level n contributes the same total weight to nodes at level l + 1 . The pyramid name arose because the weighting function nearly approximates a Gaussian function. This pyramid holds local averages on different scales, leveraged for target localization and texture analysis.
Moreover, assuming the GP [ I 0 , I 1 , , I k ] , the LP is obtained by computing b k = I k E I k + 1 , where E I k + 1 represents an upsampled, smoothed version of I k + 1 of the same dimension. In the literature, LP is used for the analysis, compression, and enhancement of images and graphics applications [23].
The GP is used in this work for the multiresolution representation of the images right before the feature extraction.

3. Proposed Approach

This section describes how the proposed method integrates a multiresolution analysis and multiple texture descriptors for texture classification. To this end, we put forward an architecture structured in five main stages. Figure 2 shows an overview of the proposed scheme, and Algorithm 1 shows the steps of the first three stages.
Multiresolution representation: In this stage, for an input image, we generate three others corresponding to the levels of the Gaussian pyramid (L0: original image, L1: first level of the pyramid, L2: second level of the pyramid). The purpose is to represent an input image in different resolutions to capture intrinsic details.
Channel splitting: Each channel (R, G, B) is considered a separate input in this phase. The key reason behind the splitting of the channels is to exploit color information. Thus, we represent and characterize an input image in a given resolution by a set of local descriptors generated from the interaction of a pixel with its neighborhood inside a given channel (R, G, or B).
Feature extraction: After the channel-splitting step, the images undergo feature extraction, which looks for intrinsic properties and discriminative characteristics. For each GP level of an image, we extract BiT [9] (14-dimensional), Shannon entropy and multi-information, i.e., the total information (3-dimensional), Haralick [4] (13-dimensional), and GLCM [4] (6-dimensional) descriptors from each channel. Images are then represented by the concatenation of different measurements organized as feature vectors (Algorithm 1). For simplicity, we name the resulting descriptor three-in-one (TiO). It is worth mentioning that images are converted to a gray scale for GLCM and Haralick measurements to be extracted but remain in color for extracting the bioinspired indices. After feature extraction and before concatenation, the feature vectors have dimensions of 126, 117, 54, and 18 for the BiT, Haralick, GLCM, and Shannon entropy and multi-information descriptors, respectively. Thereby, after concatenation, we have a 315-dimensional feature vector.
Normalization: Because test points simulate real-world data, we split the data into training and test sets and perform a min-max normalization on the training data. Subsequently, we use the same min-max values to normalize the test set.
Training/classification: We have a normalized feature vector comprising the concatenation of features extracted from each resolution and color channel from the previous stage. This texture representation is used to train one monolithic and three ensemble-based models. The linear discriminant analysis (LDA) is used for the monolithic classification model, while ensemble models are created using the histogram-based algorithm for building gradient boosting ensembles of decision trees (HistoB), the light gradient boosting decision trees (LightB), and the CatBoost classifier (https://catboost.ai/, accessed on 12 January 2023), which is an efficient implementation of the gradient boosting algorithm.
Algorithm 1: Feature extraction procedure.
Result: Feature descriptor

1. Read an RGB image file;

2. Generate 3 multiresolution levels with Gaussian pyramids L0, L1, L2;

3. For each level (L0, L1, L2) of Gaussian pyramids;

   3.1. Separate the image in channels R, G, B;

   3.2. Convert R, G, and B channels to grayscale images (for GLCM and Haralick only);

   3.3. Compute bioinspired indices, Shannon entropy, multi-information, Haralick, and GLCM descriptors of R, G, B;

   3.4. Concatenate these values into a single vector (315-dimensional);

4. Repeat steps 1 to 4 for all images of the dataset
It is worth mentioning that the method proposed in this work to improve texture classification concentrates on a multiscale analysis of textural descriptors computed from the surface of 2D images.
Step 3.1 of the algorithm splits the image into different channels, to wit, R, G, and B. The rationale is that there is a difference between processing a 3-channel RGB image and splitting the R, G, and B planes separately in texture analysis. Processing a 3-channel RGB image involves analyzing the image as a single entity, usually a single grayscale image composed of proportions of the R, G, and B channels, i.e., without separating the R, G, and B planes. This is valuable when the patterns or features contained in the image are correlated across multiple planes. For example, if the texture exhibits consistent patterns across the R, G, and B planes, then processing a 3-channel RGB image may be sufficient to capture these patterns. On the other hand, splitting the R, G, and B planes separately allows one to analyze each plane independently and extract features or patterns that are specific to that plane. This is useful when the patterns at different planes are not necessarily correlated or when the patterns at one plane may be obscured by patterns at other planes. By analyzing the R, G, and B planes separately, one can capture the full range of patterns present in the image, even if the patterns are not correlated across the different planes.
Because in texture analysis, the decision between processing the entire 3-channel RGB image or separating the R, G, and B planes rely on the image’s particular properties and the objectives of the analysis, in the experimental sets, we tested both procedures and compared the outcomes to determine which was more suitable; in most cases, splitting R, G, and B planes outperformed the 3-channel RGB image.
As a further factor, in light of the fact that texture forms a nondeterministic system of patterns, we also included features obtained through bioinspired texture descriptors (BiT) in this work. Such descriptors postulate that textural patterns act in a manner that is analogous to ecological patterns, in which populations can self-organize into aggregations that form patterns from the nondeterministic and nonlinear processes that occur inside an ecosystem. In ecology, especially in restoration ecology, summing values to characterize site communities or geographic comparisons of biodiversity therefrom is strongly inefficient if changes in primary resources, such as hydrology, soil, climate, and biology, are ignored. However, considering diversity should unveil changes over time in preserving biodiversity. To this end, splitting the R, G, and B planes separately allows one to analyze each plane independently (as changes in an ecosystem) and extract features or patterns that are specific to that plane (ecosystem in distinct differentiation) to provide feature values that can reflect the background condition and summarize values that describe texture from an image to a great extent under distinct differentiations. This combines the all-inclusive behavior from the R, G, and B planes, instead of solely from the 3-channel RGB image.

4. Experimental Results

The experimental protocol considered five datasets, three of which are texture image collections, and two are composed of medical images. The KTH-TIPS dataset contains a collection of 810 color texture images of 200 × 200 pixel of resolution. The images were captured at nine scales, under three different illumination directions, and in three different poses, with 81 images per class. Seventy percent of images are used for training, while the remaining 30% are used for testing. The Outex TC 00011-r [24] has a total of 960 (24 × 20 × 2) images of the illuminant Inca. The training set consists of 480 (24 × 20) images and the test set of 480 (24 × 20) images. The Amsterdam Library of Textures (ALOT) is a color image collection of 250 (classes) rough textures. The authors systematically varied the viewing angle, illumination angle, and illumination color for each material to capture the sensory variation in object recordings. CRC is a dataset of colorectal cancer histopathology images of 5000 × 5000 pixels that were patched into 150 × 150 pixels images and labeled according to eight types of structures. The total number of images is 625 per structure type, resulting in 5000 images. Finally, the BreakHis dataset comprises 9109 microscopic images of breast tumor tissue collected from 82 patients using different magnification factors (40×, 100×, 200×, and 400×). Currently, it contains 2480 benign and 5429 malignant samples (700 × 460 pixels, three-channel RGB, eight-bit depth in each channel, PNG format). Figure 3 presents samples of images extracted from the KTH-TIPS, Outex, ALOT, CRC, and BreakHis datasets.
We employed two experimentation strategies: a train–test split where 70% of data were reserved for training and hyperparameter tuning, and 30% for testing; (ii) a 10-fold cross-validation. The use of two strategies allowed a better assessment of the performance of the proposed approach and a fair comparison with the state of the art. In the experiments, the full feature vector generated by concatenating all descriptors (TiO) was used for training a monolithic classifier using the LDA algorithm and ensembles using HistoB, LightB, and CatBoost algorithms.

4.1. Experiments with Texture Datasets

There have been numerous important advances in texture classification in the literature during the past few decades. Many of them, however, are new takes on old ideas. Choosing which examples best illustrate the state of the art is difficult. Therefore, we focused on the first ground-breaking initiatives in texture classification. The experimental protocol to assess the proposed approach’s validity included comparisons to important shallow and deep-based methods.
Table 1 and Table 2 present the results of the proposed method (TiO), BiT, GLCM, and Haralick descriptors on the Outex, KTH-TIPS, and ALOT datasets, using the train–test split and 10-fold cross-validation, respectively. For each dataset, we present the results obtained by combining all the descriptors (TiO) and the results obtained with each descriptor individually. The main purpose was to verify the effectiveness of TiO and the descriptors’ complementary. The proposed method achieved the best average accuracy of 100% on the Outex dataset with the LDA and CatBoost classifiers. On the ALOT dataset, the proposed method achieved an average accuracy of 98.48% with the LDA classifier. Again, TiO outperformed all individual descriptors. Finally, on the KTH-TIPS dataset, the proposed method achieved the best average accuracy of 100% with the LDA classifier. Moreover, TiO outperformed all individual descriptors. Despite the different experimental protocols (train–test split and 10-fold cross-validation), the results were similar for nearly all the classifiers on the respective datasets.
Different works have used the Outex dataset for texture classification. For instance, Mehta and Egiazarian [25] introduced an approach based on a rotation-invariant LBP, achieving an accuracy of 96.26% with a k-NN classifier. Du et al. [26] presented a rotation-invariant, impulse-noise-resistant, and illumination-invariant approach based on a local spiking pattern (LSP). This approach achieved an accuracy of 86.12% with a neural network but was not extended to color textures and required several input parameters. Ataky and Lameiras Koerich [9] introduced a bioinspired texture descriptor based on biodiversity (species richness and evenness) and taxonomic measures. The latter represented the image as an abstract model of an ecosystem where species’ diversity, richness, and taxonomic distinctiveness measures were extracted. Such a texture descriptor was invariant to rotation, translation, and scale and achieved an accuracy of 99.88% with an SVM. Table 3 compares our best results with a few works that also used the Outex dataset.
Table 4 presents a few works that also used the ALOT dataset. There was an improvement in the accuracy of nearly 1% and 0.1% compared with shallow and deep methods, respectively. Notwithstanding the difference is not significantly high, it is worth mentioning that the success of CNNs relies on the ability to leverage massive, labeled datasets to learn high-quality representations. Nonetheless, data availability for a few domains may be restricted, so CNNs become restrained from several fields. Moreover, some works evaluated small-scale CNN architectures, such as T-CNN and T-CNN Inception, with 11,900 and 1.5M parameters, respectively. Despite the reduced number of parameters and lower computational cost for training, both still required a large quantity of training data to perform satisfactorily. Even some small architectures of comparable performance, such as MobileNets, EfficientNets, and sparse architectures resulting from pruning, still have many training parameters. For instance, the number of parameters of a MobileNet CNN ranges between 3.5M and 4.2M, while the number of parameters of EfficientNet CNNs range between 5.3M and 66.6M. GoogleNet, ResNet, and VGG CNNs generally need extensive training, and the number of hyperparameters and the computational cost is high.
Likewise, the KTH-TIPS dataset has been used to evaluate texture characterization and classification approaches. Hazgui et al. [32] introduced an approach that integrated the genetic programming and the fusion of HOG and LBP features, which achieved an accuracy of 91.20% with a k-NN classifier. Such an approach did not use color information and global features. Nguyen et al. [33] put forth rotational and noise-invariant statistical binary patterns, which reached an accuracy of 97.73%, lower than the accuracy achieved by the proposed method by about 2.3%. This approach was resolution-sensitive and presented a high computational complexity. Qi et al. [34] proposed a rotation-invariant multiscale cross-channel LBP (CCLBP) that encoded the cross-channel texture correlation. The CCLBP computed the LBP descriptors in each channel and three scales and computed co-occurrence statistics before concatenating the extracted features. Such an approach achieved an accuracy of 99.01% for three scales with an SVM. Nevertheless, this method was not invariant to scale. Table 5 shows that the proposed approach outperformed other works that also used the KTH-TIPS.

4.2. Experiments with HI Datasets

Table 6 presents the accuracy achieved by monolithic classifiers and ensemble methods, both trained with the proposed method on the CRC dataset with TiO and BiT, GLCM, and Haralick descriptors, individually. The proposed approach provided its best accuracy of 94.71%, with the HistoB and LightB ensemble models with all feature descriptors. The accuracy difference between TiO and the first best-related work was nearly 2.00%, which corroborated the discriminating ability of our method. Additionally, compared to each descriptor that TiO is made up of, TiO outperformed all of them when employed individually.
Table 7 compares the results achieved by the proposed approach with some state-of-the-art works to assess its effectiveness. The results achieved by TiO on the CRC dataset showed that the proposed approach worked well on images with other structures apart from textures and with no need for data augmentation. Moreover, CNNs need to be trained with a large quantity of labeled data, which may be prohibitive in medical imaging and other related fields.
Table 8 shows the accuracy achieved by monolithic classifiers and ensemble methods trained with all feature descriptors (TiO). For the the BreakHis dataset, the HistoB ensemble model achieved the best accuracy for 40×, 100×, and 400× magnifications, and the LightB ensemble model for a 200× magnification.
Table 9 shows and compares the results achieved by the proposed approach using all feature descriptors with the state of the art for the BreakHis dataset. One can note that the proposed approach achieved a considerable accuracy of 98.64% with all feature descriptors for a 40× magnification, which slightly outperformed the accuracy of both shallow and deep methods. The difference in accuracy between the proposed approach and the second-best method (CNN) was nearly 1% for a 40× magnification. Likewise, the proposed method achieved a considerable accuracy of 97.85%, 98.76%, and 98.22% for 100×, 200×, and 400× magnifications, respectively, which slightly outperformed the second-best method with a difference of 0.9% for 100×, 1.05% for 200×, and 1.0% for 400× magnifications, respectively.
We also conducted additional experiments separately with the BiT, GLCM, and Haralick descriptors. For all the magnifications, TiO outperformed each descriptor with the maximum accuracy difference of 1.25%, 0.48%, 1.01%, 1.5% for 40×, 100×, 200×, and 400× magnifications, respectively. Thus, combining the descriptors mentioned above increased the accuracy by nearly 1%.

5. Discussion

For several reasons, combining multiple feature descriptors can effectively improve the performance and robustness of a texture classification or recognition system. One reason is that different feature descriptors can capture various aspects of the texture, such as frequency, orientation, contrast, and statistical dependencies between pixels. By using multiple feature descriptors, it is possible to capture a complete representation of an image’s texture, which can improve the accuracy of the classification or recognition system. For example, the GLCM descriptor captures statistical dependencies between pairs of pixels in an image, which can help capture texture patterns such as texture boundaries and texture regions. The Haralick descriptor, on the other hand, captures statistical properties of an image, such as contrast and energy, which can help capture texture patterns such as texture density and texture directionality. The BiT descriptor captures the all-inclusive local and global patterns of a textural image. Finally, the Gaussian pyramid captures the spatial frequency content of an image at multiple scales, which can help capture texture patterns such as texture periodicity and texture regularity. Combining these feature descriptors makes it possible to capture a more comprehensive representation of an image’s texture.
As an additional benefit, combining numerous feature descriptors can make a texture classification or recognition system more robust and generalizable. By using multiple feature descriptors, the system can be more resistant to data variations, such as lighting or noise changes. This can be especially important when working with real-world data, where the conditions of the data may be difficult to control. For example, consider a texture classification system designed to recognize the texture of different fabrics. If an automatic system only uses a single feature descriptor, it may be sensitive to lighting variation or background noise, which could affect its performance. However, suppose the system uses multiple feature descriptors. In that case, it may be more robust to these variations, as each feature descriptor may capture different aspects of the texture that are less sensitive to them.
Overall, combining multiple feature descriptors in this work is essential to effectively improve the performance and robustness of a texture classification or recognition system by capturing a more comprehensive representation of an image’s texture and being more resistant to variations in the data.
The proposed approach was assessed with three natural texture image datasets, the Outex, KTH-TIPS, and ALOT datasets, and two HI datasets, CRC, and BreakHis, with 24, 10, 250, 8, and 8 classes, respectively. The experiment protocol employed a train–test split (70/30) and 10-fold cross-validation. For the 70/30 experimental protocol on natural texture images, the results led to the following findings:
  • The accuracy performance of the BiT, GLCM, and Haralick descriptors on the Outex dataset was not too different, and they all presented an accuracy above 95%. Still, the BiT and Haralick descriptors outperformed the GLCM descriptor with nearly 4% in terms of accuracy using the LDA classifier, the lowest accuracy among all the classifiers. However, since TiO’s performance was roughly equivalent to that of the BiT and Haralick descriptors for the Outex dataset, such a combination was not necessary to this extent.
  • Along the same line, the accuracy performance of the BiT and Haralick descriptors on the KTH-TIPS dataset was similar, regardless of the classifier. The GLCM descriptor, nonetheless, showed a difference of about 11% (with its lowest accuracy) and 7% (with its highest accuracy) compared with the BiT and Haralick descriptors. However, the difference was insignificant when comparing the best descriptor, BiT, with TiO. Therefore, for the KTH-TIPS dataset, such a combination brought a slight improvement.
  • Unlike the Outex and KTH-TIPS datasets, the presented concatenation played a significant role in the classification performance of the ALOT dataset. First, the best performance of all the datasets was obtained with the LDA classifier. The highest accuracy performance with TiO was 98.48%, outperforming the BiT, GLCM, and Haralick descriptors by nearly 20%, 32%, and 14%, respectively. The lowest difference, 14%, was still a significant improvement. Such a classification performance is promising as it is above several state-of-the-art deep methods on the ALOT dataset. The ALOT dataset is challenging due to different variations in images regarding viewing angle, illumination angle, and color.
  • Furthermore, the combination of BiT, GLCM, and Haralick descriptors provided a state-of-the-art classification performance on HIs.
Notwithstanding, one states that an effective feature selection may be necessary to derive subsets of relevant features from each descriptor such that each dataset can be characterized to the maximum extent with as minimum a feature dimension as possible. Finally, it is worth mentioning that the experimental protocol with a train–test split and 10-fold cross-validation provided similar results.

6. Conclusions

This research provided a critical study regarding image analysis and manipulation over a spectrum of spatial scales and the complementarity of feature descriptors for texture classification. We stated that employing each descriptor separately may overlook relevant textural information, reducing the classification performance. Moreover, we exploited the pyramid’s multiresolution representation as a useful data structure for analyzing and capturing intrinsic details from texture over a spectrum of spatial scales. To produce a feature vector from an image, we combined several descriptors that were proven to be discriminating for the classification, namely, the BiT, information-theoretic measures, GLCM, and Haralick descriptors, to extract gray-level and color features at different resolutions. Such a combination aimed to bring features that characterized the texture to the maximum extent and with some advantages such as rotation-, permutation-, scale-invariance, reduced noise sensitivity, and a generic and high generalization ability, as it provided an effective performance for real-world datasets.
The proposed approach outperformed a few state-of-the-art shallow and deep methods. Moreover, the descriptors employed herein were proven complementary as their combination resulted in a better performance than using each separately. However, some features may be redundant after the concatenation into a single feature vector, given the different resolutions, channels, and descriptors. That may cause a downfall in the classification performance. Furthermore, such a concatenation may also lead to the Hughes phenomenon, which explains why we did not include other descriptors such as LBP, HOG, etc. However, we will consider including other descriptors in our future studies. Finally, to circumvent the possible feature redundancy and cope with the increase in dimensionality, we will investigate the impact of incorporating a decision-making multiobjective feature selection.

Author Contributions

Conceptualization, S.T.M.A., D.S., J.d.M. and A.L.K.; Methodology, S.T.M.A.; Writing—original draft preparation, S.T.M.A. and A.L.K.; Writing—review and editing, S.T.M.A., A.d.S.B.J. and A.L.K.; Supervision, A.L.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research was partially funded by the Regroupement Strategique REPARTI-Fonds de Recherche du Québec—Nature et Technologie (FRQNT) and by the Natural Sciences and Engineering Research Council of Canada (NSERC), Discovery grant number RGPIN-2016-04855.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The KTH-TIPS dataset is available at https://www.csc.kth.se/cvap/databases/kth-tips/index.html, accessed on 12 January 2023. The Outex dataset is available at https://color.univ-lille.fr/datasets/extended-outex, accessed on 12 January 2023. The ALOT dataset is available at https://color.univ-lille.fr/datasets/alot, accessed on 12 January 2023. The BreakHis dataset is available at https://web.inf.ufpr.br/vri/databases/breast-cancer-histopathological-database-breakhis/, accessed on 12 January 2023. The CRC dataset is available at https://zenodo.org/record/53169#.Y3MHVy_71qs, accessed on 12 January 2023. The source code will be available at https://github.com/stevetmat/BioInspiredFDesc, accessed on 12 January 2023.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Tuceryan, M.; Jain, A.K. Texture analysis. In Handbook of Pattern Recognition and Computer Vision; World Scientific: Singapore, 1993; pp. 235–276. [Google Scholar]
  2. Simon, P.; Uma, V. Review of texture descriptors for texture classification. In Data Engineering and Intelligent Computing; Springer: Berlin/Heidelberg, Germany, 2018; pp. 159–176. [Google Scholar]
  3. Liu, L.; Chen, J.; Fieguth, P.; Zhao, G.; Chellappa, R.; Pietikäinen, M. From BoW to CNN: Two decades of texture representation for texture classification. Int. J. Comput. Vis. 2019, 127, 74–109. [Google Scholar] [CrossRef] [Green Version]
  4. Haralick, R.M.; Shanmugam, K.; Dinstein, I.H. Textural features for image classification. IEEE Trans. Syst. Man Cybern. 1973, SMC-3, 610–621. [Google Scholar] [CrossRef] [Green Version]
  5. Pietikäinen, M.; Hadid, A.; Zhao, G.; Ahonen, T. Computer Vision Using Local Binary Patterns; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2011; Volume 40. [Google Scholar]
  6. Haralick, R.M. Statistical and structural approaches to texture. Proc. IEEE 1979, 67, 786–804. [Google Scholar] [CrossRef]
  7. Cross, G.R.; Jain, A.K. Markov random field texture models. IEEE Trans. Pattern Anal. Mach. Intell. 1983, PAMI-5, 25–39. [Google Scholar] [CrossRef] [PubMed]
  8. Arivazhagan, S.; Ganesan, L. Texture classification using wavelet transform. Pattern Recognit. Lett. 2003, 24, 1513–1521. [Google Scholar] [CrossRef]
  9. Ataky, S.T.M.; Lameiras Koerich, A. A novel bio-inspired texture descriptor based on biodiversity and taxonomic measures. Pattern Recognit. 2022, 123, 108382. [Google Scholar] [CrossRef]
  10. Andrearczyk, V.; Whelan, P.F. Using filter banks in Convolutional Neural Networks for texture classification. Pattern Recognit. Lett. 2016, 84, 63–69. [Google Scholar] [CrossRef] [Green Version]
  11. de Matos, J.; de Souza Britto Junior, A.; Soares de Oliveira, L.E.; Lameiras Koerich, A. Texture CNN for Histopathological Image Classification. In Proceedings of the 32nd IEEE Intl Symp on Computer-Based Medical Systems (CBMS), Cordoba, Spain, 5–7 June 2019; pp. 580–583. [Google Scholar] [CrossRef] [Green Version]
  12. Fujieda, S.; Takayama, K.; Hachisuka, T. Wavelet convolutional neural networks for texture classification. arXiv 2017, arXiv:1707.07394. [Google Scholar]
  13. Vriesman, D.; de Souza Britto Junior, A.; Zimmer, A.; Lameiras Koerich, A.; Paludo, R. Automatic visual inspection of thermoelectric metal pipes. Signal Image Video Process. 2019, 13, 975–983. [Google Scholar] [CrossRef]
  14. Ataky, S.T.M.; Lameiras Koerich, A. Texture Characterization of Histopathologic Images Using Ecological Diversity Measures and Discrete Wavelet Transform. arXiv 2022, arXiv:2202.13270. [Google Scholar]
  15. de Matos, J.; Ataky, S.T.M.; de Souza Britto Junior, A.; Soares de Oliveira, L.E.; Lameiras Koerich, A. Machine learning methods for histopathological image analysis: A review. Electronics 2021, 10, 562. [Google Scholar] [CrossRef]
  16. Ataky, S.T.M.; de Matos, J.; de Souza Britto Junior, A.; Soares de Oliveira, L.E.; Lameiras Koerich, A. Data Augmentation for Histopathological Images Based on Gaussian-Laplacian Pyramid Blending. In Proceedings of the 2020 International Joint Conference on Neural Networks (IJCNN), Glasgow, UK, 19–24 July 2020; pp. 1–8. [Google Scholar] [CrossRef]
  17. Scott, P.J. Pattern Analysis and Metrology: The Extraction of Stable Features from Observable Measurements. Proc. Math. Phys. Eng. Sci. 2004, 460, 2845–2864. [Google Scholar] [CrossRef]
  18. Brown, C.A.; Hansen, H.N.; Jiang, X.J.; Blateyron, F.; Berglund, J.; Senin, N.; Bartkowiak, T.; Dixon, B.; Le Goic, G.; Quinsat, Y.; et al. Multiscale analyses and characterizations of surface topographies. CIRP Ann. 2018, 67, 839–862. [Google Scholar] [CrossRef]
  19. Eseholi, T.; Coudoux, F.X.; Corlay, P.; Sadli, R.; Bigerelle, M. A multiscale topographical analysis based on morphological information: The HEVC multiscale decomposition. Materials 2020, 13, 5582. [Google Scholar] [CrossRef]
  20. Blakemore, C.; Campbell, F.W. On the existence of neurones in the human visual system selectively sensitive to the orientation and size of retinal images. J. Physiol. 1969, 203, 237–260. [Google Scholar] [CrossRef] [PubMed]
  21. Jolion, J.M.; Rosenfeld, A. A Pyramid Framework for Early Vision: Multiresolutional Computer Vision; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2012; Volume 251. [Google Scholar]
  22. Lindeberg, T. Scale-Space Theory in Computer Vision; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2013; Volume 256. [Google Scholar]
  23. Adelson, E.H.; Anderson, C.H.; Bergen, J.R.; Burt, P.J.; Ogden, J.M. Pyramid methods in image processing. RCA Eng. 1984, 29, 33–41. [Google Scholar]
  24. Ojala, T.; Maenpaa, T.; Pietikainen, M.; Viertola, J.; Kyllonen, J.; Huovinen, S. Outex-new framework for empirical evaluation of texture analysis algorithms. In Proceedings of the Object Recognition Supported by User Interaction for Service Robots, Quebec City, QC, Canada, 11–15 August 2002; Volume 1, pp. 701–706. [Google Scholar]
  25. Mehta, R.; Egiazarian, K. Dominant rotated local binary patterns (DRLBP) for texture classification. Pattern Recognit. Lett. 2016, 71, 16–22. [Google Scholar] [CrossRef]
  26. Du, S.; Yan, Y.; Ma, Y. Local spiking pattern and its application to rotation-and illumination-invariant texture classification. Optik 2016, 127, 6583–6589. [Google Scholar] [CrossRef]
  27. Armi, L.; Abbasi, E.; Zarepour-Ahmadabadi, J. Texture images classification using improved local quinary pattern and mixture of ELM-based experts. Neural Comput. Appl. 2021, 34, 21583–21606. [Google Scholar] [CrossRef]
  28. Dubey, S.R.; Singh, S.K.; Singh, R.K. Multichannel decoded local binary patterns for content-based image retrieval. IEEE Trans. Image Process. 2016, 25, 4018–4032. [Google Scholar] [CrossRef]
  29. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  30. Napoletano, P. Hand-crafted vs learned descriptors for color texture classification. In Proceedings of the International Workshop on Computational Color Imaging, Milan, Italy, 29–31 March 2017; Springer: Berlin/Heidelberg, Germany, 2017; pp. 259–271. [Google Scholar]
  31. Alpaslan, N.; Hanbay, K. Multi-resolution intrinsic texture geometry-based local binary pattern for texture classification. IEEE Access 2020, 8, 54415–54430. [Google Scholar] [CrossRef]
  32. Hazgui, M.; Ghazouani, H.; Barhoumi, W. Genetic programming-based fusion of HOG and LBP features for fully automated texture classification. Vis. Comput. 2021, 38, 457–476. [Google Scholar] [CrossRef]
  33. Nguyen, T.P.; Vu, N.S.; Manzanera, A. Statistical binary patterns for rotational invariant texture classification. Neurocomputing 2016, 173, 1565–1577. [Google Scholar] [CrossRef]
  34. Qi, X.; Qiao, Y.; Li, C.; Guo, J. Exploring Cross-Channel Texture Correlation for Color Texture Classification. In Proceedings of the British Machine Vision Conference, BMVC 2013, Bristol, UK, 9–13 September 2013; Burghardt, T., Damen, D., Mayol-Cuevas, W.W., Mirmehdi, M., Eds.; BMVA Press: Guildford, UK, 2013. [Google Scholar] [CrossRef] [Green Version]
  35. Kather, J.N.; Weis, C.A.; Bianconi, F.; Melchers, S.M.; Schad, L.R.; Gaiser, T.; Marx, A.; Zöllner, F.G. Multi-class texture analysis in colorectal cancer histology. Sci. Rep. 2016, 6, 27988. [Google Scholar] [CrossRef]
  36. Sarkar, R.; Acton, S.T. Sdl: Saliency-based dictionary learning framework for image similarity. IEEE Trans. Image Process. 2017, 27, 749–763. [Google Scholar] [CrossRef] [PubMed]
  37. Wang, C.; Shi, J.; Zhang, Q.; Ying, S. Histopathological image classification with bilinear convolutional neural networks. In Proceedings of the 39th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Jeju Island, Republic of Korea, 11–15 July 2017; pp. 4050–4053. [Google Scholar]
  38. Pham, T.D. Scaling of texture in training autoencoders for classification of histological images of colorectal cancer. In Proceedings of the International Symposium on Neural Networks, Athens, Greece, 25–27 August 2017; Springer: Berlin/Heidelberg, Germany, 2017; pp. 524–532. [Google Scholar]
  39. Raczkowski, Ł.; Możejko, M.; Zambonelli, J.; Szczurek, E. ARA: Accurate, reliable and active histopathological image classification framework with Bayesian deep learning. Sci. Rep. 2019, 9, 14347. [Google Scholar] [CrossRef] [Green Version]
  40. Spanhol, F.A.; Oliveira, L.S.; Petitjean, C.; Heutte, L. A Dataset for Breast Cancer Histopathological Image Classification. IEEE Trans. Biomed. Eng. 2016, 63, 1455–1462. [Google Scholar] [CrossRef]
  41. Erfankhah, H.; Yazdi, M.; Babaie, M.; Tizhoosh, H.R. Heterogeneity-aware local binary patterns for retrieval of histopathology images. IEEE Access 2019, 7, 18354–18367. [Google Scholar] [CrossRef]
  42. Alom, M.Z.; Yakopcic, C.; Nasrin, M.S.; Taha, T.M.; Asari, V.K. Breast cancer classification from histopathological images with inception recurrent residual convolutional neural network. J. Digit. Imaging 2019, 32, 605–617. [Google Scholar] [CrossRef] [Green Version]
  43. Han, Z.; Wei, B.; Zheng, Y.; Yin, Y.; Li, K.; Li, S. Breast cancer multi-classification from histopathological images with structured deep learning model. Sci. Rep. 2017, 7, 4172. [Google Scholar] [CrossRef]
  44. Bayramoglu, N.; Kannala, J.; Heikkilä, J. Deep learning for magnification independent breast cancer histopathology image classification. In Proceedings of the 23rd International Conference on Pattern Recognition (ICPR), Cancun, Mexico, 4–8 December 2016; pp. 2440–2445. [Google Scholar]
  45. Spanhol, F.A.; Oliveira, L.S.; Petitjean, C.; Heutte, L. Breast cancer histopathological image classification using convolutional neural networks. In Proceedings of the 2016 International Joint Conference on Neural Networks (IJCNN), Vancouver, BC, Canada, 24–29 July 2016; pp. 2560–2567. [Google Scholar]
Figure 1. An example of Gaussian pyramids from the same input image. (b) First three levels of the Gaussian pyramid.
Figure 1. An example of Gaussian pyramids from the same input image. (b) First three levels of the Gaussian pyramid.
Applsci 13 01291 g001
Figure 2. An overview of the proposed scheme.
Figure 2. An overview of the proposed scheme.
Applsci 13 01291 g002
Figure 3. Sample images from different datasets.
Figure 3. Sample images from different datasets.
Applsci 13 01291 g003
Table 1. Average accuracy (%) on the test set of the Outex, ALOT, and KTH-TIPS datasets. The best accuracy for each dataset is shown in boldface.
Table 1. Average accuracy (%) on the test set of the Outex, ALOT, and KTH-TIPS datasets. The best accuracy for each dataset is shown in boldface.
DatasetTexture DescriptorsClassification Algorithms
HistoB  1 LightB  2 LDACatBoost  3
OutexTiO100.0100.0100.0100.0
BiT99.9299.69100.099.84
GLCM98.6997.9995.9199.30
Haralick98.8499.5399.3899.84
ALOTTiO71.5073.0798.4893.47
BiT65.7166.5278.1474.90
GLCM51.2254.3266.6164.23
Haralick68.5263.4184.5783.23
KTH-TIPSTiO98.5397.20100.098.91
BiT96.2996.7099.5897.53
GLCM89.7190.1287.6592.01
Haralick95.8896.2998.3596.11
1 max_bins = 10, max_iter = 1500. 2 , 3 n_estimators = 1500.
Table 2. Average accuracy and standard deviation (%) on the Outex, ALOT, and KTH-TIPS datasets using a 10-fold cross-validation setup. The best accuracy for each dataset is shown in boldface.
Table 2. Average accuracy and standard deviation (%) on the Outex, ALOT, and KTH-TIPS datasets using a 10-fold cross-validation setup. The best accuracy for each dataset is shown in boldface.
DatasetTexture DescriptorsClassification Algorithms
HistoB  1 LightB  2 LDACatBoost  3
OutexTiO99.93 (0.001)99.88 (0.002)100.0 (0.0)99.49 (0.003)
BiT99.81 (0.001)99.53 (0.003)99.69 (0.003)99.88 (0.001)
GLCM97.16 (0.003)98.24 (0.004)96.52 (0.006)98.93 (0.004)
Haralick99.18 (0.001)99.19 (0.002)99.64 (0.003)99.29 (0.001)
ALOTTiO42.82 (0.009)53.31 (0.012)97.89 (0.015)95.89 (0.022)
BiT29.81 (0.001)32.53 (0.003)83.65 (0.002)82.12 (0.013)
GLCM13.14 (0.015)35.12 (0.212)69.72 (0.023)72.62 (0.010)
Haralick18.45 (0.020)29.57 (0.002)86.54 (0.024)81.33 (0.021)
KTH-TIPSTiO98.03 (0.013)97.41 (0.027)99.98 (0.006)98.66 (0.018)
BiT96.61 (0.023)96.80 (0.026)97.88 (0.006)97.79 (0.023)
GLCM90.86 (0.030)90.74 (0.035)84.44 (0.044)90.12 (0.034)
Haralick95.29 (0.017)96.04 (0.023)96.75 (0.004)96.40 (0.011)
1 max_bins = 10, max_iter = 1500. 2 , 3 n_estimators = 1500.
Table 3. Accuracy (%) of the proposed method and related works on the Outex dataset. The best result is shown in boldface.
Table 3. Accuracy (%) of the proposed method and related works on the Outex dataset. The best result is shown in boldface.
ReferenceTypeApproachAccuracy
Ataky and Lameiras Koerich [9]ShallowGLCM + LightB95.52
Haralick + k-NN96.92
BiT + SVM99.88
Mehta and Egiazarian [25]ShallowLBP-based + k-NN96.23
Du et al. [26]ShallowLSP + k-NN86.12
TiOShallowLDA or CatBoost100.0
Table 4. Accuracy (%) of shallow and deep approaches on the ALOT dataset. The best result is shown in boldface.
Table 4. Accuracy (%) of shallow and deep approaches on the ALOT dataset. The best result is shown in boldface.
ReferenceApproachAccuracy
Armi et al. [27]Shallow97.56
Dubey et al. [28]Shallow63.04
He et al. [29]ResNet5075.68
ResNet10175.60
Napoletano [30]ResNet10198.13
ResNet5098.35
VGG VeryDeep1994.93
VGG M12885.56
GoogleNet92.65
VGG M102492.58
VGG M204893.30
Alpaslan and Hanbay [31]Shallow97.20
TiO + LDAShallow98.48
Table 5. Accuracy (%) of the proposed method and related works on the KTH-TIPS dataset. The best result is shown in boldface.
Table 5. Accuracy (%) of the proposed method and related works on the KTH-TIPS dataset. The best result is shown in boldface.
ReferenceApproachAccuracy
Ataky and Lameiras Koerich [9]GLCM86.83
Haralick94.89
BiT97.87
Hazgui et al. [32]Shallow91.20
Nguyen et al. [33]Shallow97.73
Qi et al. [34]Shallow99.01
TiO + LDAShallow100.0
Table 6. Accuracy (%) of monolithic classifiers and ensemble methods with TiO and each descriptor employed individually on the test set of the CRC dataset. TiO (CV) represents the average accuracy and standard deviation (%) on a 10-fold cross-validation setup. The best results are shown in boldface.
Table 6. Accuracy (%) of monolithic classifiers and ensemble methods with TiO and each descriptor employed individually on the test set of the CRC dataset. TiO (CV) represents the average accuracy and standard deviation (%) on a 10-fold cross-validation setup. The best results are shown in boldface.
Texture DescriptorsClassification Algorithms
HistoBLightBLDACatBoost
TiO94.7194.7194.3793.80
TiO-CV94.28 (0.017)94.50 (0.019)93.37 (0.014)93.87 (0.018)
BiT93.2292.6186.6792.92
GLCM85.9886.9781.7987.20
Haralick91.6291.3192.3891.39
Table 7. Average accuracy (%) of shallow and deep approaches on the CRC dataset. The best results are marked in boldface.
Table 7. Average accuracy (%) of shallow and deep approaches on the CRC dataset. The best results are marked in boldface.
ReferenceApproach10-fold5-fold
Kather et al. [35]Shallow87.40
Sarkar and Acton [36]Shallow73.60
Ataky and Lameiras Koerich [9]Shallow92.96
TiO+LightBShallow94.5092.77
Wang et al. [37]CNN92.60
Pham [38]CNN84.00
Raczkowski et al. [39]CNN92.4092.20
Table 8. Image-level accuracy (%) of monolithic classifiers and ensemble methods with the TiO descriptor on 8 balanced classes of the BreakHis dataset. The best result for each magnification is shown in boldface. CV denotes the average accuracy and standard deviation (%) using a 10-fold cross-validation experimental setup.
Table 8. Image-level accuracy (%) of monolithic classifiers and ensemble methods with the TiO descriptor on 8 balanced classes of the BreakHis dataset. The best result for each magnification is shown in boldface. CV denotes the average accuracy and standard deviation (%) using a 10-fold cross-validation experimental setup.
MagnificationClassification Algorithms
HistoBLightBLDACatBoost
40×98.6498.1297.8398.04
40× (CV)98.47 (0.021)97.64 (0.015)98.12 (0.029)97.28 (0.026)
100×97.8597.5096.8196.67
100× (CV)97.28 (0.021)96.91 (0.015)95.56 (0.027)96.46 (0.024)
200×97.6298.7694.6397.28
200× (CV)96.82 (0.022)97.92 (0.014)95.93 (0.027)97.13 (0.024)
400×98.2298.0694.9296.98
400× (CV)97.81 (0.018)97.48 (0.022)94.11 (0.027)96.07 (0.025)
Table 9. Accuracy (%) of shallow and deep approaches on the BreakHis dataset. All works used the same data partitions for training and testing. The best result for each magnification is shown in boldface.
Table 9. Accuracy (%) of shallow and deep approaches on the BreakHis dataset. All works used the same data partitions for training and testing. The best result for each magnification is shown in boldface.
ReferenceMethodMagnification
40×100×200×400×
Spanhol et al. [40]Shallow (LBP)75.6073.0072.9071.20
Spanhol et al. [40]Shallow (GLCM)74.7076.8083.4081.70
Erfankhah et al. [41]Shallow (LBP)88.3088.3087.1083.40
Ataky and Lameiras Koerich [9]Shallow (BiT)97.5096.8095.8095.20
Alom et al. [42]CNN97.0097.5097.2097.20
Han et al. [43]CNN92.8093.9093.7092.90
Bayramoglu et al. [44]CNN83.0083.1084.6082.10
Spanhol et al. [45]CNN90.0088.4084.6086.10
TiO+HistoBShallow98.6497.8598.7698.22
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ataky, S.T.M.; Saqui, D.; de Matos, J.; de Souza Britto Junior, A.; Lameiras Koerich, A. Multiscale Analysis for Improving Texture Classification. Appl. Sci. 2023, 13, 1291. https://doi.org/10.3390/app13031291

AMA Style

Ataky STM, Saqui D, de Matos J, de Souza Britto Junior A, Lameiras Koerich A. Multiscale Analysis for Improving Texture Classification. Applied Sciences. 2023; 13(3):1291. https://doi.org/10.3390/app13031291

Chicago/Turabian Style

Ataky, Steve Tsham Mpinda, Diego Saqui, Jonathan de Matos, Alceu de Souza Britto Junior, and Alessandro Lameiras Koerich. 2023. "Multiscale Analysis for Improving Texture Classification" Applied Sciences 13, no. 3: 1291. https://doi.org/10.3390/app13031291

APA Style

Ataky, S. T. M., Saqui, D., de Matos, J., de Souza Britto Junior, A., & Lameiras Koerich, A. (2023). Multiscale Analysis for Improving Texture Classification. Applied Sciences, 13(3), 1291. https://doi.org/10.3390/app13031291

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop