Next Article in Journal
Dynamic Numerical Analysis of Displacement Restraining Effect of Inclined Earth-Retaining Structure during Embankment Construction
Previous Article in Journal
SPL Features Quantification and Selection Based on Multiple Multi-Level Objectives
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Local Ternary Cross Structure Pattern: A Color LBP Feature Extraction with Applications in CBIR

1
College of Information Science and Engineering, Northeastern University, Shenyang 110004, China
2
School of Software, Jiangxi Normal University, Nanchang 330022, China
3
College of Intelligence and Computing, Tianjin University, Tianjin 300350, China
4
School of Computer Engineering, Weifang University, Weifang 261061, China
*
Authors to whom correspondence should be addressed.
Appl. Sci. 2019, 9(11), 2211; https://doi.org/10.3390/app9112211
Submission received: 8 April 2019 / Revised: 21 May 2019 / Accepted: 24 May 2019 / Published: 29 May 2019
(This article belongs to the Section Optics and Lasers)

Abstract

:
With the advent of medical endoscopes, earth observation satellites and personal phones, content-based image retrieval (CBIR) has attracted considerable attention, triggered by its wide applications, e.g., medical image analytics, remote sensing, and person re-identification. However, constructing effective feature extraction is still recognized as a challenging problem. To tackle this problem, we first propose the five-level color quantizer (FLCQ) to acquire a color quantization map (CQM). Secondly, according to the anatomical structure of the human visual system, the color quantization map (CQM) is amalgamated with a local binary pattern (LBP) map to construct a local ternary cross structure pattern (LTCSP). Third, the LTCSP is further converted into the uniform local ternary cross structure pattern (LTCSPuni) and the rotation-invariant local ternary cross structure pattern (LTCSPri) in order to cut down the computational cost and improve the robustness, respectively. Finally, through quantitative and qualitative evaluations on face, objects, landmark, textural and natural scene datasets, the experimental results illustrate that the proposed descriptors are effective, robust and practical in terms of CBIR application. In addition, the computational complexity is further evaluated to produce an in-depth analysis.

1. Introduction

Along with the development of imaging equipment, a larger number of images have been extensively collected from various fields [1,2,3]. Meanwhile, CBIR technology has gradually become a hot research field, due to its applications in place recognition [4], image classification [5], and remote sensing [6]. Therefore, the problem of extracting effective, robust and practical features has attracted an increasing number of researchers. Thanks to these pioneers’ breakthroughs, many approaches [7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26] have been continuously proposed and extended for the task of feature extraction.
In the early days, a family of local binary pattern (LBP)-based methods [7,8,9,10,11,12,13,14,15,16,17] have been sequentially reported for the grayscale-based feature extraction. As a milestone, the LBP definition was initially authored by Ojala et al. [7], in which the referenced pixel and its nearest pixels were encoded as a binary string. Hereafter, Zhang et al. [8] extended the LBP to the local derivative pattern (LDP) descriptor for refining the magnitude difference. Subsequently, Guo et al. [9] designed a variant of the LBP named the completed LBP, and it is used for improving the robustness to rotation. After that, the LBP variance was developed by Guo et al. [10] for addressing the drawback of the global information loss. Further, Tan et al. [11] introduced an improvement of the LBP named the local ternary pattern (LTP), and it was integrated with the kernel principal component analysis to improve the robustness to illumination. After, Murala et al. [12] modified the LTP to the local tetra pattern (LTrP), in which the referenced pixel and its surrounding pixels were computed by comparing the vertical and horizontal direction differences. Afterwards, Subrahmanyam et al. [13] converted the LTP to the local maximum edge binary patterns (LMEBP), and the LMEBP was combined with the Gabor transform for CBIR and object tracking applications. In order to extend the LBP to the dynamic-texture application, Zhao et al. [14] studied the LBP histogram Fourier (LBP-HF) for video sequence recognition. Further, in order to improve the robustness to noise, Ren et al. [15] proposed the noise-resistant local binary pattern (NRLBP) to extract the local structure information. In order to retrieve the magnetic and resonance and computer tomography images, Subrahmanyam et al. [16] studied the local ternary co-occurrence patterns (LTCoP) encoding. Motivated by the concatenation strategy, Verma et al. [17] used the LBP feature map and the local neighborhood difference pattern (LNDP) to integrate the binary information into the local intensity difference for the natural scene and texture retrieval applications. Nevertheless, since all the above approaches are limited to the grayscale-based feature extraction, the color information unavoidably is lost.
In recent years, a family of color LBP descriptors [18,19,20,21,22,23,24,25,26,27,28] have been continuously explored for the color-based feature extraction. Among them, inspired by inter- and intra- channel encoding mechanisms, Mäenpää et al. [18] constructed the opponent color local binary patterns (OCLBP) for color textural classification. After that, Bianconi et al. [19] proposed an extension of OCLBP named the improved opponent color local binary patterns (IOCLBP), in which point-to-average thresholding replaced point-to-point thresholding. Further, the grayscale-based LTrP was extended to the local oppugnant color texture pattern (LOCTP) by Jeena Jacob et al. [20], and the LOCTP was extracted from the RGB, YCbCr, and HSV color models respectively. Inspired by the pair-based strategy, Qi et al. [21] presented the pairwise rotation-invariant co-occurrence LBP (PRICoLBP) by incorporating the R, G and B components. Similarly, Hao et al. [22] proposed the pairwise cross pattern (PCP), in which the color and LBP information were combined in pairwise and cross manners. With the help of the encoding-decoding technology, Dubey et al. [23] designed the multi-channel adder LBP (maLBP) and the multi-channel decoder LBP (mdLBP) to extract the LBP feature maps from the R, G, and B channels. In 2017, inspired by the vector quantization (VQ) strategy, Guo et al. [24] proposed the max and min color quantizer in order to extract the color information feature (CIF) in the RGB color model, and the CIF feature was combined with the LBP-based feature for the image classification and retrieval applications. In 2018, Somasekar et al. [25] integrated the Fuzzy C-Mean color clustering in the RGB color model with the LBP feature for the side-scan-sonar image enhancement. In the same year, the CIELAB color model was quantized by Singh et al. [26] to extract the color histogram (CH), and the CH feature was linearly combined with the orthogonal combination of LBP (OC-LBP) for the color image retrieval applications. More recently, Feng et al. [27] introduced the local parallel cross pattern (LPCP) which integrated the color and LBP information in parallel and cross manners. In order to capture the cross-channel information, Agarwal et al. [28] studied the multi-channel local ternary pattern (MCLTP) to extract the correlation from the H-V, S-V and V-V channels in a cross manner.
In this paper, we present the main following contributions:
  • We construct a five-level color quantizer, and it is applied to quantize the a* and b* components for the color quantization map extraction.
  • We integrate the color quantization map into the LBP feature map to extract a local ternary cross structure pattern (LTCSP).
  • We further extend the local ternary cross structure pattern to the uniform local ternary cross structure pattern and the rotation-invariant local ternary cross structure pattern for reducing the computational cost and improving the robustness.
  • We benchmark a series of experiments on face, landmark, object and textural datasets, and extensive experimental results demonstrate the effectiveness, robustness, and practicability of the proposed descriptor.
The rest of this paper is organized as follows. In Section 2, the local binary pattern definition and the color prior knowledge are briefly reviewed. Section 3 concretely details the feature extraction. Similarity measure and retrieval system are introduced in Section 4. Section 5 presents the experiments and discussion. Section 6 concludes this paper, and points out the possible future directions.

2. Related Work

2.1. Local Binary Pattern Definition

Firstly, the local binary pattern (LBP) definition was initially authored by Ojala et al. [7] for the gray-scale feature extraction. Given a referenced pixel P(s, t) and its surrounding pixels Pk(s, t), the LBP encoding at P(s, t) is formulated as follows:
L B P r , n ( s , t ) = k = 0 n 1 μ ( P ( s , t ) P k ( s , t ) ) × 2 k ,
μ ( m ) = { 1 , m 0 0 , m < 0 ,
where r is the radius of a circle, and n is the number of surrounding pixels in the circle with radius r.
Then, depending on the realistic application, over 90% of image micro-structures are encoded by only 23% of the LBP patterns. To cut down the computational cost, the LBP is simplified to the “uniform local binary pattern”, which is formulated as follows:
L B P r , n u n i ( s , t ) = U { k = 1 n | μ ( P ( s , t ) P k ( s , t ) ) μ ( P ( s , t ) P k 1 ( s , t ) ) | } ,
where U{·} represents a measure operator, and U{·} ≤ 2. P0 (s, t) is equivalent to Pn(s, t).
Finally, to improve robustness, the LBP further is converted into the rotation invariant local binary pattern L B P r , n r i ( s , t ) , which is expressed as follows [21]:
L B P r , n r i ( s , t ) = m i n { R O R ( L B P r , n ( s , t ) , k )   |   k 0 , 1 , , n 1 } ,
where ROR(LBPr,n(s, t), k) is a circular bit-wise right shift for x times on n-bit number LBPr,n(s, t).
Referring to [21,29], r and n are defined as 1 and 8 respectively. In the following, the LBP feature map, the uniform LBP feature map, and the rotation invariant LBP feature map are abbreviated as LBP, LBPuni, and LBPri respectively.

2.2. Color Quantization Scheme

Currently, the equal-interval color quantizer (EICQ) is considered as the most common used quantizer [26,30,31]. Among them, Singh et al. [26] proposed the color histogram (CH)-based scheme in which the number of quantization blocks for L*, a* and b* components was fixed at 3, 2 and 2. Similarly, Reta et al. [30] proposed the Lab color coherence vector (Lab-CVV) to quantize the L*, a* and b* components to 5, 4 and 4 respectively. Considering the HSV color model, Wan and Kuo [31] developed the multiresolution histogram representation (MHR)-based quantizer in which a variety of combinations of 2, 4, 8 and 16 blocks were designed for the H, S and V components quantization. Motive by the flexible stragegy, Liu et al. [32] proposed the flexible micro structure descriptor histogram (MSD)-based quantizer to quantize the L*, a* and b* component to 20, 3 and 3 blocks respectively. Similarly to the MSD-based quantizer, Liu et al. [33] presented the adaptable color difference histogram (CDH)-based quantizer in which the optimal number of quantization in the L* component was set to 10, and the number of quantization in the a* and b* components was fixed as 3. With help of the analysis of color distribution, Wan et al. [34] proposed standard vector quantization, a method of partitioning the vector space by minimizing the mean squared error. Motivated by unsupervised clustering techniques, Duda and Hart [35] designed the criterion-based schemes and the heuristic schemes. Inspired by the color clusters distribution, Xia and Kuo [36] proposed the octree with pruning color quantization which calculated the color information of each image respectively.

2.3. Color Prior Knowledge in the CIELAB Color Model

The CIELAB color model consists of three components, namely the white-black component of L* (ranging from 0 to 100), the yellow-blue component of a* (ranging from −128 to +127), and the red-green component of b* (ranging from −128 to +127) [37]. The CIELAB color model is not only an excellent splitter between color (represented by the a* and b* components) and intensity (represented by the L* component), but also is a perceptually uniform color space; that is to say there are the same amount of numerical changes between the CIELAB model and the visual perception in human color vision [38]. On the basis of the CIELAB color model, a color prior knowledge was originally introduced by Feng et al. [5], in which the frequency of pixels in the a* and b* components were explored and analyzed. Firstly, in Figure 1a,b the frequency of pixels is mostly distributed in the center of the a* and b* components on the Caltech-256 [39] dataset. Secondly, to verify the consistency, thousands of image datasets are calculated, and extensive experimental results demonstrate the consistency of this prior. Obviously, it can be summarized that most pixels focus on the middle third of the range. Thirdly, to verify the stability, a series of additional experiments are performed, which illustrate that the prior knowledge is stable even if the image dataset is changed. For instance, the probability distribution of the Caltech-256 dataset (see Figure 1a,b) is extremely approximate to 10% of the Caltech-256 dataset (see Figure 1c,d).

3. Feature Extraction

3.1. Five-Level Color Quantizer

Inspired by the color prior knowledge in the CIELAB color model, a novel five-level color quantizer is designed for the color quantization map (CQM) extraction. For convenience, the original [−128, +127] is mapped as 28. In our scheme, 28 is first subdivided into four blocks at Level 1, in which two blocks 28/3 are on both sides, and two refined blocks 27/3 are in the middle because most pixels focus on the middle third of the range. The corresponding indices are sequentially flagged as 0, 1, 2, and 3 at Level 1. Then, two refined blocks 27/3 are subdivided into to two refined 27/32 in the middle and two 28/32 on both sides from Level 1 to 2. In this manner, the pixels in the middle range can be further refined. The remaining blocks are duplicated from Level 1 to 2. Finally, two operators of “Duplicate” and “Subdivide” are sequentially repeated until the two middle blocks 27/35 at Level 5. We combine Level 1 to 5 to construct the five-level color quantizer. For clarity, the process is displayed in Figure 2, in which each level contains a group of blocks and indices. Mathematically, the quantization levels in the a* and b* components are flagged as Aa* and Ab*, where Aa*     {1, 2, …, 5} and Ab*     {1, 2, …, 5}. Meanwhile, the corresponding indices in the a* and b* components are flagged as Âa* and Âb*, where Âa*     {0, 1, …, 2(Aa* + 1) − 1} and Âb*     {0, 1, …, 2(Ab* + 1) − 1} respectively.
Moreover, in terms of the humans’ eye intensity perception mechanism [40], the original scope [0, +100] in the L* component is quantized into 3 blocks, e.g., [0, +25], [+26, +75] and [+76, +100]. Similarly, the quantization level in the L* component is flagged as AL*, where AL* = 1 and the index is defined as ÂL*     {0, 1, 2}.
An example of the proposed color quantization scheme is shown in Figure 3. From Figure 3a, the values in the CIELAB color model are set to L* = +87 in the L* component, a* = +84 in the a* component, and b* = −28 in the b* component. Correspondingly, the quantization levels are set to AL* = 1, Aa* = 2 and Ab* = 2 respectively. From Figure 3b, according to AL* = 1, the L* component is firstly quantized to 3 blocks, e.g., [0, +25], [+26, +75] and [+76, +100], and then the index of L* = +87 can be encoded as ÂL* = 2. From Figure 3c,d, considering Aa* = 2 and Ab* = 2 in the FLCQ quantizer, the a* and b* components are firstly quantized to six blocks, e.g., 28/3, 28/32, 27/32, 27/32, 28/32 and 28/3, and then the indices of a* = +84 and b* = −82 is encoded as Âa* = 5 and Âb* = 1 respectively. We combine the indices of ÂL*, Âa*, and Âb* to acquire the color quantization map (CQM), and the index of CQM is flagged as E, E     {0, 1, …, 3 × 2(Aa* + 1) × 2(Ab* + 1) − 1}.

3.2. Human Visual System

As documented in Gray’s Anatomy, the anatomical structure of the human visual system consists of the eyeball, optic nerve, lateral geniculate nucleus of thalamus, optic radiation and visual cortex [41]. Remarkably, it should be noted that the optic chiasm is a critical anatomical structure, in which a part of the visual cues between the left and right cerebral hemispheres are exchanged. According to the human visual system, the left and right eye balls first extract the low-level visual cues. Second, the extracted low-level visual cues are encoded and transmitted to the left and right optic nerves. Third, a part of the encoding visual information is crossed at the optic chiasm. Fourth, the crossed visual information is reconstructed at the left and right lateral geniculate nucleus of thalamus. Fifth, the reconstructed visual information is radiated at the left and right optic radiations. Finally, the radiated visual information is re-aggregated at the left and right visual cortices to form the high-level semantics perception. For more details, please refer to [41].

3.3. Local Ternary Cross Structure Pattern

According to the anatomical structure of the human visual system, a novel local ternary cross structure pattern (LTCSP) is proposed to integrate the color information into the LBP information as a whole. For an original map M(i, j), the reference point is flagged as M(i0, j0) and the eight nearest points are flagged as M(ip, jp), where p     {1, 2, …, 8}.
Firstly, the LBP feature map is computed as LBP(i, j) in Section 2.1, and the color quantization map is computed as CQM(i, j) in Section 3.1. Secondly, with the help of a thresholded polynomial selectivity indicator sign (·), the color quantization map and the LBP feature map are encoded into the color ternary map and the LBP ternary map respectively. Mathematically, the thresholded polynomial selectivity indicator sign (·) is defined as follows:
sign ( ϒ ( i 0 , j 0 ) , ϒ ( i p , j p ) ) = { 1   if   ϒ ( i 0 , j 0 ) > ϒ ( i p , j p )   0 if   ϒ ( i 0 , j 0 ) = ϒ ( i p , j p ) 1 if   ϒ ( i 0 , j 0 ) < ϒ ( i p , j p ) ,
where p     {1, 2, …, 8}, and ϒ ( i , j ) represents a feature map. For the LBP feature map, ϒ ( i , j ) is considered as LBP(i, j); for the color quantization map, ϒ ( i , j ) is considered as CQM(i, j). Thirdly, motivated by the optic chiasm, in which a part of the visual cues between the left and right cerebral hemispheres are exchanged, the eight nearest points in the color and LBP ternary maps are correspondingly crossed to extract the color and LBP cross maps. Fourthly, with the help of a counter ϑ {·} that computes the occurring numbers of ternary in the eight nearest points of the color and LBP cross maps, the maximum numbers are retained to construct the color and LBP structure maps respectively. Inspired by the max-pooling strategy, when there exist two maximum occurring numbers, the structure with the larger value is retained. Finally, the retained points in the color and LBP structure maps are computed as the feature vectors, and the values of the reference points are correspondingly calculated as the indices of the feature vectors. Mathematically, we define the LTCSP as follows:
L T C S P L B P ( L B P ( i 0 , j 0 ) ) = max p { 1 , 2 , 8 } ϑ { sign ( C Q M ( i 0 , j 0 ) , C Q M ( i p , j p ) ) } ,
L T C S P C Q M ( C Q M ( i 0 , j 0 ) ) = max p { 1 , 2 , 8 } ϑ { sign ( L B P ( i 0 , j 0 ) , L B P ( i p , j p ) ) } ,
For clarity, the schematic diagram of the LTCSP is illustrated in Figure 4, where the LTCSP is computed as [LTCSPCQM(250) = 6, LTCSPLBP(235) = 4]. Experimentally, the feature dimensionality of LTCSPCQM and LTCSPLBP are 3 × 2(Aa* + 1) × 2(Ab* + 1) and 256 respectively.
In order to select the optimal color quantization levels (Aa*, Ab*), we need to compare the average precision rates (APR) of 25 possible combinations according to different image datasets in the offline stage. Mathematically, given an image dataset D, the maximization APR is defined as follows:
argmax A a * , A b * APR ( D | A a * , A b * ) ,
where APR(D|Aa*, Ab*) represents the APR score, where Aa* ∈ {1, 2, …, 5} and Ab*     {1, 2, …, 5}. In Section 5.4., we provide the optimal color quantization levels (Aa*, Ab*).
Moreover, in order to reduce the computational cost, the LBP feature map LBP(i, j) in LTCSP can be replaced by the LBPuni feature map LBPuni(i, j) to construct the uniform local ternary cross structure pattern (LTCSPuni). Similarly, in order to improve the robustness to rotation, the LBP feature map LBP(i, j) in LTCSP can be replaced by the LBPri feature map LBPri(i, j) to construct the rotation-invariant local ternary cross structure pattern (LTCSPri).

4. Similarity Measure and Retrieval System

4.1. Similarity Measure

Given a query image provided by a user, the query and dataset images are first encoded as the query and dataset feature vectors respectively. Then, the similarity measure between the query and dataset feature vectors is performed. Finally, based on the sorting results of the similarity measure, the top similar images are returned to the user. Referring to [6,22,33,42], extended Canberra distance (ECD) is utilized in this paper, and the ECD is defined as follows:
E C D ( f d , f q ) = τ = 1 δ | f τ d f τ q | | f τ d + υ d | + | f τ q + υ q | ,
where ECD(·) denotes the result of extended Canberra distance. f τ d and f τ q are the feature vectors of the query and database images, and τ represents the dimensionality of the feature vector. υ d and υ q are defined as τ = 1 δ f τ d / δ and τ = 1 δ f τ q / δ respectively.

4.2. Retrieval System

User, query image, dataset images, feature extraction, query and dataset feature vectors, similarity measure, and returned images are totally combined to construct the proposed retrieval system. The schematic diagram of the proposed retrieval system is shown in Figure 5. From Figure 5, the proposed retrieval system can be divided into the offline stage and the online stage. In the offline stage, based on the optimal quantization levels (Aa*, Ab*) in Equation (8), all dataset images are sent to the feature extraction block to extract the dataset feature vectors only once. In the online stage, the user firstly inputs the query image. Secondly, the query image is sent to the feature extraction block to extract the query feature vector. Thirdly, the similarity measure block is performed between the query and dataset feature vectors. Finally, according to the similarity measure scores, the top-n similar images are considered as returned images to the user. If there exist several datasets, the query image would be adaptively encoded into different query feature vectors.

5. Experiments and Discussion

5.1. Evaluation Criteria

The precision rate, recall rate, average precision rate, and average recall rate are commonly adopted as the most used evaluation criteria. Among them, the precision and recall rates are used to evaluate the retrieval performance of a single query image, and they are defined as follows [3,5]:
P r e c i s i o n = Number   of   similar   images   returned Total   number   of   images   returned ,
R e c a l l = Number   of   similar   images   returned Number   of   all   relevant   images ,
Further, the average precision rate (APR) and average recall rate (ARR) are applied to evaluate the retrieval performance of the total number of query images, and they are computed as follows:
A P R = t = 1 T P r e c i s i o n ( t ) T ,
A R R = t = 1 T R e c a l l ( t ) T ,
where t represents the t-th query image, and T is the total number of query images.

5.2. Image Datasets

Six benchmark datasets, consisting of one face image dataset (Face-95 [43]), one object image dataset (ETHZ [44]), one landmark image dataset (ZuBuD [45]), two color textural image datasets (KTH-2a [46] and VisTex [47]), and one natural scene image dataset (Corel-100 [48]), are summarized in Table 1 to evaluate the effectiveness, robustness, and practicability of the proposed descriptors.
The Face-95 (No. 1) is a face image dataset, and it was captured by a S-VHS camcorder. The Face-95 consists of 1440 images in 72 persons. Each person contains 20 images in JPG format with size of 180 × 200. Note that all images are with face expressions, illumination changes, head scales and head turns. In Figure 6a, there are some samples from Face-95.
The ETHZ (No. 2) is an object image dataset, and it was taken with a Sony XC-77P camera. The ETHZ has 265 images in 53 natural objects. For each object, there are 72 images in BNG format with size of 320 × 240. Specially, each object is rotated by an arbitrary degree, so the ETHZ also can be used to evaluate the robustness of rotation. In Figure 6b, some samples from ETHZ are displayed.
The ZuBuD (No. 3) is a landmark image dataset, and it was produced by the Panasonic-NV-MX300 and Pentax-Opti430 cameras. The ZuBuD has 1005 images in 201 landmarks. Each landmark contains five images in JPG format with size of 640 × 480. Note that not only some occlusions (e.g., tree and car) are purposely included in images, but also all images are captured under different views points, weather conditions and seasons. Thus, the ZuBuD can be used for evaluating the effectiveness under the complex environment. In Figure 6c, some samples from ZuBuD are showned.
The VisTex (No. 4) and KTH-2a (No. 5) are two color textural image datasets. On one hand, the VisTex (No. 4) was acquired by collecting images from videos and photographs. The VisTex is having 640 images in 40 categories (e.g., fabric, flower, matel, terrain, bark, sand, leave, stone, and so on). Each category includes 16 images in PPM format with size of 128 × 128. In Figure 6d, there are some samples from VisTex. On the other hand, the KTH-2a (No. 5) was taken with an Olympus C-3030ZOOM camera, and it consists of 4608 images in 11 categories (e.g., cotton, brown bread, wood, wool, white bread, corduroy, linen, cracker, cork, aluminium foil and lettuce leaf). Each category contains 396/432 images in BNG format with size of 200 × 200. Specially, the KTH-2a also can be applied to evaluate whether the proposed descriptors are robust against rotation, scaling and illumination. In Figure 6e, there are some samples from KTH-2a.
The Corel-1000 (No. 6) is a natural scene image dataset, and it was collected by the SIMPLIcity system. The Corel-1000 has 1000 images in 10 natural scenes. Each natural scene contains 100 images in JPG format with size of 384 × 256 or 256 × 384. In Figure 6f, some samples from Corel-1000 are presented.

5.3. Experimental Details

All experiments were performed on a personal computer with an Intel Core i7-7700k [email protected] Ghz, a 16 GB DDR4 RAM@2400 MHz, and a 6 GB NVIDIA GTX1070ti. All six benchmark datasets in our experiments can be freely downloaded from [43,44,45,46,47,48]. Note that all images need to be transformed from the RGB to the grayscale space and the CIELAB color model for extracting the LBP feature map and the color quantization map respectively. For details, these transformations are referred to [38]. In addition, the open source implementation of the LBP feature map [9,10] can be downloaded from http://www.ee.oulu.fi/~gyzhao/. Referring to [6,22,23,27,40,49,50], all images in the dataset are considered as the query images to guarantee the accurateness and reproducibility. Referring to [6,22,23,27,40,49,50], the number of returned images on VisTex, KTH-2a, Face-95 and Corel-1000 is defined as 10, and the number of returned images on ETHZ and ZuBuD is defined as five because there are only five images in each class.

5.4. Evaluation of Color Quantization Levels

Table 2 reports the APR rates of LTCSP, LTCSPuni and LTCSPri with the optimal levels (Aa*, Ab*) over six datasets. For clarity, the best APR values with the optimal level (Aa*, Ab*) are documented in bold. Firstly, in Table 2, it can easily draw the conclusions that a single quantization level cannot satisfy the need of all image datasets. For example, LTCSP yields the highest APR rate on Face-95 when (Aa* = 5, Ab* = 5); (2) LTCSP acquires the top APR rate on ETHZ when (Aa* = 3, Ab* = 4); and (3) LTCSP products the highest APR rate on Corel-1000 when (Aa* = 3, Ab* = 3). Secondly, it can be noted that a relatively simple level also products the highest APR rate. For instance, when (Aa* = 3, Ab* = 2), LTCSPuni acquires the top APR rate on VisTex, and LTCSPuni and LTCSPri product the highest APR rate on Corel-1000. Depending upon all proven observations, it is necessary to optimally choose (Aa*, Ab*) from the five-level color quantizer. Additionally, we provide the APR values of about 25 possible combinations on six datasets in the supplementary file. In the following experiments, the optimal levels of Aa* and Ab* are adaptively adopted in the proposed descriptors according to different datasets.

5.5. Comparison with Other Hierarchical Quantization Schemes

Referring to [51,52,53], the mean square error (MSE) values obtained by the hierarchical quantization schemes of CH [26], Lab-CVV [30], CDH [32], MSD [33], LTCSP, LTCSPuni, and LTCSPri on six datasets are reported in Table 3. First, it can be clearly observed that the MSE values by the proposed LTCSP, LTCSPuni, and LTCSPri methods are obviously lower than all other descriptors on six datasets. These phenomena illustrate the hierarchical quantization schemes of the proposed LTCSP, LTCSPuni, and LTCSPri are superior to all other descriptors. Secondly, it can be concluded that the mean square errors by the proposed LTCSP, LTCSPuni, and LTCSPri methods are extremely close to each other on six datasets. These results demonstrate that the hierarchical quantization schemes of the proposed descriptors are stable and consistent between six datasets. Thirdly, it can be summarized that there exist obvious differences among different datasets. These phenomena demonstrate it is necessary to adaptively select different color quantization levels according to different datasets. In addition, we provide the MSE values of 25 color quantization levels (Aa*, Ab*) between the FLCQ and EICQ quantizers in the CIELAB color model on all six datasets in Appendix A.

5.6. Comparison with LBP-Based Methods

Table 4 details the evaluations of the APR and ARR rates resulting from the LBP, LBPuni, and LBPri methods and the proposed LTCSP, LTCSPuni, and LTCSPri methods. The best {APR, ARR} values are highlighted in bold. First, it can be clearly observed that the proposed LTCSP, LTCSPuni, and LTCSPri methods achieve remarkable enhancements as compared to the LBP, LBPuni, and LBPri methods on all six datasets. The foremost reason is that the proposed methods are beneficial to integrate the color information and the LBP information. Secondly, it is noted that LTCSP achieves the highest APR rate on VisTex and Corel-1000, LTCSPri generates the top APR rates on Face-95, ETHZ, ZuBuD, and KTH-2a datasets respectively. The possible reasons are summarized as follows: (1) there is no difference VisTex and Corel-1000; and there exist rotation differences on Face-95, ETHZ, ZuBuD, and KTH-2a. According to these encouraging results, it can be deduced that the proposed LTCSP, LTCSPuni, and LTCSPri methods are superior to the LBP-based methods on all six datasets.

5.7. Comparison with Other Color LBP Descriptor

Table 5 reports the evaluations of the APR and ARR rates obtained by the proposed descriptors and a series of state-of-the-art color LBP descriptors including OCLBP [18], IOCLBP [19], maLBP [23] mdLBP [23], OC-LBP + CH [26], LPCP [27] on all six datasets. The best values are highlighted in bold. For LTCSP, it not only reaches the higher results than all other previous color LBP descriptors on Face-95, ETHZ, ZuBuD and KTH-2a datasets, but also yields the highest {APR = 98.56%, ARR = 61.60%} on VisTex and {APR = 83.94%, ARR = 8.39%} on Corel-1000. In the case of “LTCSPuni”, it achieves a better performance than OCLBP, IOCLBP, mdLBP, maLBP, and OC-LBP + CH on all six datasets. For LTCSPri, it acquires the highest {APR = 98.39%, ARR = 48.69%} on Face-95, {APR = 94.72%, ARR = 94.72%} on ETHZ, {APR = 86.11%, ARR = 86.11%} on ZuBuD, and {APR = 99.19%, ARR = 2.37%} on KTH-2a respectively. But we also note that the {APR, ARR} rates of LTCSPuni and LTCSPri are slightly inferior to the LPCP on VisTex and Corel-1000. The main reason is that LPCP has a higher feature dimension. However, the issue can be addressed by adding more useful feature vectors. Based on these considerable results, the effectiveness of the proposed descriptors is demonstrated by comparing with six state-of-the-art color LBP descriptors. Furthermore, there exists illumination differences, head scale and head turn on Face-95, rotation differences on ETHZ, point view differences on ZuBuD, rotation, scaling and illumination differences on KTH-2a. Therefore, the proposed descriptors are also robust against rotation, scaling and illumination to some extent.
Figure 7 depicts the comparisons of the top-10 returned images acquired by the proposed methods and six previous color LBP methods on six benchmark datasets. Based on the similarity measure score, the top-10 returned images are sorted in descending order. Among them, the leftmost image in each row not only is the most similar image, but also is the query image. Clearly, when a returned image is the relational image, it is tagged in a green box; else it is tagged in a red box. In Figure 7a, the APR rate is 10% using OCLBP, 10% using IOCLBP, 10% using maLBP, 20% using mdLBP, 40% using mdLBP, 60% using OC-LBP + CH, 100% using LTCSP, 100% using LTCSPuni and 100% using LTCSPri respectively. From this figure, we can deduce that the proposed LTCSP, LTCSPuni and LTCSPri descriptors not only are efficient for face-based image retrial applications, but also are insensitive to head scales and head turns. In Figure 7b, the {APR, ARR} rate of relational images using OCLBP, IOCLBP, maLBP, mdLBP, OC-LBP + CH, LPCP, LTCSP, LTCSPuni and LTCSPri are {60%, 60%}, {60%, 60%}, {60%, 60%}, {60%, 60%}, {60%, 60%}, {80%, 80%}, {100%, 100%}, {100%, 100%} and {100%, 100%} respectively. From this comparison, it can be observed that LTCSP, LTCSPuni and LTCSPri are robust for object-based image retrial applications when the object of “toy plane” is rotated arbitrarily. In Figure 7c, the number of relational images using OCLBP, IOCLBP, maLBP, mdLBP, OC-LBP + CH, LPCP, LTCSP, LTCSPuni and LTCSPri are 6, 6, 6, 8, 7, 6, 6, 6, 10, 10 and 10. From Figure 7c, we summarize that the proposed descriptors are effective and robust for landmark-based image retrial applications even if some occlusions (e.g., tree and car) are purposefully included in images. As expected, from Figure 7d–f, LTCSP, LTCSPuni and LTCSPri still bring about higher APR rates than existing color LBP descriptors, apart from LPCP in Figure 7d. However, by comparing with the 10th returned images using LPCP, LTCSP, LTCSPuni and LTCSPri, it can be clearly seen that the proposed descriptors are more semantically similar with the leftmost query image. From these observations and analyses, the practicability and usability of LTCSP, LTCSPuni and LTCSPri are illustrated.
Table 6 reports the comparison of the feature dimensionality (d) and the memory cost (kB) comparison among the proposed descriptors and six former color LBP methods. Similar to LPCP, the items of 688/496/688/436/544/448 (d) and 5.38/3.88/5.38/3.41/4.25/3.50 (kB) represent LTCSP with 688 d and 5.38 kB conducts experiments on Face-95, 496 d and 3.88 kB on ETHZ, 688 d and 5.38 kB on ZuBuD, 436 d and 3.41 kB on VisTex, 544 d and 4.25 kB on KTH-2a, as well as 448 d and 3.50 kB on Corel-1000 respectively. As documented in Table 6, the feature dimension and memory cost of the proposed descriptors are obviously lower than OCLBP, IOCLBP, maLBP, mdLBP and LPCP on all six datasets (apart from LPCP on Corel-1000), yet LTCSP, LTCSPuni, and LTCSPri are also higher than OC-LBP + CH respectively. However, the superiorities of LTCSP, LTCSPuni, and LTCSPri are still summarized as follows:
  • The additional feature dimensionality and memory cost effectively improve the accuracy by a large marginal.
  • The LTCSP, LTCSPuni, and LTCSPri achieve the highest score on all six datasets.
  • The proposed methods achieve a trade-off compromise: adaptive feature dimensionality and acceptable memory cost, and competitive candidate in the real-world CBIR applications.

5.8. Comparison with Deep Learning (DL)-Based Models

Additionally, the proposed LTCSP, LTCSPuni, and LTCSPri descriptors are further compared with the emerging deep learning (DL)-based models including ALEX [54], GoogleNet [55], VGGm128 [56], VGGm1024 [56], VGGm2048 [56], and VGGm4096 [56]. Referring to [57,58], firstly, the last full-connected layers in the pre-trained models are converted into the corresponding feature vector. Secondly, the converted feature vectors are sent to perform the L2 normalization. Thirdly, the normalized feature vectors are used for computing the similarity measure score.
Figure 8 presents the comparisons among the proposed descriptors and the DL-based models. For LTCSP, LTCSPuni, and LTCSPri, it can be observed that the proposed descriptors product higher APRs rates than all DL-based models on VisTex, KTH-2a, Face-95, ETHZ and ZuBuD datasets. For Corel-1000, LTCSP, LTCSPuni, and LTCSPri product lower APR rates than all DL-based models. There are two main reasons for this phenomenon: (1) Corel-1000 is a natural scene image dataset that include more complex scene semantic information; and (2) all DL-based models are pre-trained on the ImageNet that is the natural scene dataset. Comparing with the DL-based models, the superiorities of LTCSP, LTCSPuni, and LTCSPri are summarized as follows:
  • The DL-based models rely heavily on expensive hardware configurations (e.g., RAM and GPU), yet the proposed descriptors can be easily embedded into cheap hardware devices (e.g., chip and microcontroller).
  • The DL-based models are sensitive to rotation, scaling and illumination differences, while the proposed descriptors are robust against rotation, scaling, and illumination differences to some extent.
  • The DL-based models need to be pre-trained on large-scale and annotated datasets (e.g., ImageNet), which seriously limits its applications.
  • LTCSP, LTCSPuni, and LTCSPri are superior to the DL-based models on five datasets out of six.

6. Conclusions

In this study, a series of color LBP descriptors namely local ternary cross structure pattern (LTCSP), uniform local ternary cross structure pattern (LTCSPuni) and rotation-invariant local ternary cross structure pattern (LTCSPri) is proposed for the CBIR applications. According to the experimental results, the effectiveness, robustness, and practicability of the proposed descriptors are evaluated and compared on face, landmark, object, natural scene and textural image datasets. Based on these considerable results, it can be concluded that the proposed methods achieve a trade-off/compromise among notable retrieval accuracy, adaptive feature dimensionality and acceptable memory cost, and they can be considered as a competitive candidate for real-world CBIR applications.
In the future, unsupervised feature selection [59] will be implemented to tackle the issue of the feature dimensionality and memory cost. In order to improve the robustness against illumination, the image normalization [60] will be exploited in the image pre-processing. In addition, manifold learning (ML) [61,62] and query expansion (QE) [63] will also be considered to further enhance the retrieval performance.

Supplementary Materials

The following are available online at https://www.mdpi.com/2076-3417/9/11/2211/s1.

Author Contributions

Q.F. conceived the research idea. Q.F. and Q.H. performed the experiments. Q.F., Q.H. and J.D. wrote the paper. Q.F., Q.H., Y.Y., J.D. and Y.W. gave many suggestions and helped revise this manuscript.

Funding

This research was funded by the National Nature Science Foundation of China, grant number [61871106, 61370152], the Fundamental Research Grant Scheme for the Central Universities, grant number [130204003], the Project of Shandong Province Higher Educational Science and Technology Program, grant number [J16LN68], the Weifang Science and Technology Development Plan Project (Nos. 2017GX006, 2018GX009, 2018GX004) and the National Key Technology Research and Development Programme of the Ministry of Science and Technology of China, grant number [2014BAI17B02].

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Table A1. The mean square error of the EICQ and FLCQ quantizers on Face-95.
Table A1. The mean square error of the EICQ and FLCQ quantizers on Face-95.
Quantization
Level Aa*
QuantizerQuantization Level Ab*
Ab* = 1Ab* = 2Ab* = 3Ab* = 4Ab* = 5
Aa* = 1EICQ1150.65877.69825.03812.65792.54
FLCQ355.51260.54254.78254.46254.45
Aa* = 2EICQ737.79464.83412.17399.79379.69
FLCQ156.9261.9556.2055.8855.86
Aa* = 3EICQ615.66342.70290.04277.66257.56
FLCQ149.7054.7348.9748.6548.64
Aa* = 4EICQ569.25296.29243.63231.25211.15
FLCQ149.3554.3848.6348.3148.29
Aa* = 5EICQ549.35276.39223.73211.35191.25
FLCQ149.3454.3748.6248.3048.28
Table A2. The mean square error of the EICQ and FLCQ quantizers on ETHZ.
Table A2. The mean square error of the EICQ and FLCQ quantizers on ETHZ.
Quantization
Level Aa*
QuantizerQuantization Level Ab*
Ab* = 1Ab* = 2Ab* = 3Ab* = 4Ab* = 5
Aa* = 1EICQ1458.631018.17874.3811.39778.25
FLCQ568.67301.03286.28285.81285.78
Aa* = 2EICQ1033.79593.33449.46386.55353.41
FLCQ344.2976.6661.9061.4361.41
Aa* = 3EICQ905.31464.86320.98258.07224.94
FLCQ347.0079.3764.6164.1564.12
Aa* = 4EICQ851.94411.49267.61204.70171.56
FLCQ346.6178.9864.2263.7563.73
Aa* = 5EICQ827.35386.90243.02180.11146.98
FLCQ346.6178.9764.2263.7563.72
Table A3. The mean square error of the EICQ and FLCQ quantizers on ZuBuD.
Table A3. The mean square error of the EICQ and FLCQ quantizers on ZuBuD.
Quantization
Level Aa*
QuantizerQuantization Level Ab*
Ab* = 1Ab* = 2Ab* = 3Ab* = 4Ab* = 5
Aa* = 1EICQ1698.311247.151106.261048.221020.12
FLCQ609.53359.00344.97344.27344.24
Aa* = 2EICQ1203.92752.76611.87553.82525.72
FLCQ302.6352.1138.0737.3737.34
Aa* = 3EICQ1041.70590.54449.65391.60363.50
FLCQ283.9533.4219.3818.6818.66
Aa* = 4EICQ971.45520.28379.40321.35293.25
FLCQ283.0932.5618.5217.8217.80
Aa* = 5EICQ935.66484.50343.61285.56257.46
FLCQ283.0432.5218.4817.7817.75
Table A4. The mean square error of the EICQ and FLCQ quantizers on VisTex.
Table A4. The mean square error of the EICQ and FLCQ quantizers on VisTex.
Quantization
Level Aa*
QuantizerQuantization Level Ab*
Ab* = 1Ab* = 2Ab* = 3Ab* = 4Ab* = 5
Aa* = 1EICQ1530.691167.551057.571011.01987.40
FLCQ556.59367.10358.62358.18358.17
Aa* = 2EICQ1053.23690.09580.11533.55509.94
FLCQ267.3177.8169.3368.9068.88
Aa* = 3EICQ898.62535.48425.50378.94355.33
FLCQ248.2758.7750.2949.8549.84
Aa* = 4EICQ832.16469.02359.04312.48288.87
FLCQ247.4657.9749.4949.0549.04
Aa* = 5EICQ798.37435.23325.25278.69255.08
FLCQ247.4257.9249.4449.0148.99
Table A5. The mean square error of the EICQ and FLCQ quantizers on KTH-2a.
Table A5. The mean square error of the EICQ and FLCQ quantizers on KTH-2a.
Quantization
Level Aa*
QuantizerQuantization Level Ab*
Ab* = 1Ab* = 2Ab* = 3Ab* = 4Ab* = 5
Aa* = 1EICQ1299.411090.021049.361031.491016.96
FLCQ438.11379.63373.66373.47373.47
Aa* = 2EICQ838.96629.56588.91571.04556.51
FLCQ176.84118.36112.39112.20112.20
Aa* = 3EICQ693.71484.32443.66425.79411.26
FLCQ160.37101.8995.9295.7495.73
Aa* = 4EICQ633.55424.16383.50365.64351.10
FLCQ159.52101.0495.0794.8994.88
Aa* = 5EICQ604.17394.77354.12336.25321.71
FLCQ159.49101.0195.0394.8594.84
Table A6. The mean square error of the EICQ and FLCQ quantizers on Corel-1000.
Table A6. The mean square error of the EICQ and FLCQ quantizers on Corel-1000.
Quantization
Level Aa*
QuantizerQuantization Level Ab*
Ab* = 1Ab* = 2Ab* = 3Ab* = 4Ab* = 5
Aa* = 1EICQ1363.971026.78928.64887.69865.82
FLCQ509.07342.97332.88332.43332.41
Aa* = 2EICQ946.85609.65511.52470.57448.70
FLCQ283.30117.20107.12106.67106.65
Aa* = 3EICQ822.90485.70387.57346.62324.75
FLCQ268.31102.2192.1291.6791.65
Aa* = 4EICQ768.42431.23333.09292.14270.27
FLCQ267.57101.4791.3990.9490.92
Aa* = 5EICQ741.21404.02305.88264.93243.06
FLCQ267.54101.4491.3590.9190.88
In order to illustrate the advantage of the proposed five-level color quantizer (FLCQ), the mean square error (MSE) values of the EICQ and FLCQ in the CIELAB color model are compared on all six datasets. To guarantee the experimental fairness, the same settings are set to the EICQ and FLCQ quantizers, apart from the number of quantization intervals. The lowest MSE values of the FLCQ and EICQ quantizers are highlighted in bold. Firstly, from Table A1, Table A2, Table A3, Table A4, Table A5 and Table A6, it can be summarized that the MSE values of the FLCQ quantizer are lower than EICQ on all six datasets. Secondly, it can be concluded that along with the refinement of levels (Aa*, Ab*), where Aa*, Ab* ∈ {1, 2, …, 5}, the MSE values of the EICQ quantizers drops much more obviously, yet the MSE values of the FLCQ quantizers from Aa* = 4 to 5 are decrease slightly. The results not only illustrate the stability of the FLCQ quantizers, but also demonstrate that it is suitable to stop the quantization level at Aa* = 5. Thirdly, under the same quantization interval and level in both quantizers, the MSE values are different from one another among six datasets. These results show that there exist obvious color probability distribution differences among different datasets. So, it is reasonable to adopt the FLCQ quantizer. As a consequence, we can observe that the FLCQ quantizer produces the lower MSE than the EICQ quantizer.

References

  1. Smeulders, A.W.M.; Worring, M.; Santini, S.; Gupta, A.; Jain, R. Content-based image retrieval at the end of the early years. IEEE Trans. Pattern Anal. Mach. Intell. 2002, 22, 1349–1380. [Google Scholar] [CrossRef]
  2. Zheng, L.; Yang, Y.; Tian, Q. SIFT meets CNN: A decade survey of instance retrieval. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 40, 1224–1244. [Google Scholar] [CrossRef] [PubMed]
  3. Irtaza, A.; Adnan, S.; Ahmed, K.; Jaffar, A.; Khan, A.; Javed, A.; Mahmood, M. An ensemble based evolutionary approach to the class imbalance problem with applications in CBIR. Appl. Sci. 2018, 8, 495. [Google Scholar] [CrossRef]
  4. Zeng, Z.; Zhang, J.; Wang, X.; Chen, Y.; Zhu, C. Place recognition: An overview of vision perspective. Appl. Sci. 2018, 8, 2257. [Google Scholar] [CrossRef]
  5. Zafar, B.; Ashraf, R.; Ali, N.; Lqbal, M.; Sajid, M.; Dar, S.; Ratyal, N. A novel discriminating and relative global spatial image representation with applications in CBIR. Appl. Sci. 2018, 8, 2242. [Google Scholar] [CrossRef]
  6. Feng, Q.; Hao, Q.; Chen, Y.; Yi, Y.; Wei, Y.; Dai, j. Hybrid histogram descriptor: A fusion feature representation for image retrieval. Sensors 2018, 18, 1943. [Google Scholar] [CrossRef]
  7. Ojala, T.; Pietikäinen, M.; Maenpaa, T. Multi resolution gray-scale and rotation invariant texture classification with local binary patterns. IEEE Trans. Pattern Anal. Mach. Intell. 2002, 24, 971–987. [Google Scholar] [CrossRef]
  8. Zhang, B.; Gao, Y.; Zhao, S.; Liu, J. Local derivative pattern versus local binary pattern: Face recognition with high-order local pattern descriptor. IEEE Trans. Image Process. 2010, 19, 533–544. [Google Scholar] [CrossRef]
  9. Guo, Z.; Zhang, L.; Zhang, D. A completed modeling of local binary pattern operator for texture classification. IEEE Trans. Image Process. 2010, 19, 1657–1663. [Google Scholar]
  10. Guo, Z.; Zhang, L.; Zhang, D. Rotation invariant texture classification using LBP variance (LBPV) with global matching. Pattern Recognit. 2010, 43, 706–719. [Google Scholar] [CrossRef]
  11. Tan, X.; Triggs, B. Enhanced local texture feature sets for face recognition under difficult lighting conditions. IEEE Trans. Image Process. 2010, 19, 1635–1650. [Google Scholar]
  12. Murala, S.; Maheshwari, R.P.; Balasubramanian, R. Local tetra patterns: A new feature descriptor for content-based image retrieval. IEEE Trans. Image Process. 2012, 21, 2874–2886. [Google Scholar] [CrossRef] [PubMed]
  13. Subrahmanyam, M.; Maheshwari, R.P.; Balasubramanian, R. Local maximum edge binary patterns: A new descriptor for image retrieval and object tracking. Signal. Process. 2012, 92, 1467–1479. [Google Scholar] [CrossRef]
  14. Zhao, G.; Ahonen, T.; Matas, J.; Pietikainen, M. Rotation-invariant image and video description with local binary pattern features. IEEE Trans. Image Process. 2012, 21, 1465–1477. [Google Scholar] [CrossRef]
  15. Ren, J.; Jiang, X.; Yuan, J. Noise-resistant local binary pattern with an embedded error-correction mechanism. IEEE Trans. Image Process. 2013, 22, 4049–4060. [Google Scholar] [CrossRef] [PubMed]
  16. Murala, S.; Wu, Q.M.J. Local ternary co-occurrence patterns: A new feature descriptor for MRI and CT image retrieval. Neurocomputing 2013, 119, 399–412. [Google Scholar] [CrossRef]
  17. Verma, M.; Raman, B. Local neighborhood difference pattern: A new feature descriptor for natural and texture image retrieval. Multimed. Tools Appl. 2018, 77, 11843–11866. [Google Scholar] [CrossRef]
  18. Mäenpää, T.; Pietikäinen, M. Texture analysis with local binary patterns. In Handbook of Pattern Recognition and Computer Vision; Word Scientific: Singapore, 2005; pp. 197–216. [Google Scholar]
  19. Bianconi, F.; Bello-Cerezo, R.; Napoletano, P. Improved opponent color local binary patterns: An effective local image descriptor for color texture classification. J. Electron. Imaging 2017, 27, 011002. [Google Scholar] [CrossRef]
  20. Jeena Jacob, I.; Srinivasagan, K.G.; Jayapriya, K. Local oppugnant color texture pattern for image retrieval system. Pattern Recognit. Lett. 2014, 42, 72–78. [Google Scholar] [CrossRef]
  21. Qi, X.; Xiao, R.; Li, C.; Qiao, Y.; Guo, J.; Tang, X. Pairwise rotation invariant co-occurrence local binary pattern. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 36, 2199–2213. [Google Scholar] [CrossRef] [PubMed]
  22. Hao, Q.; Feng, Q.; Wei, Y.; Sbert, M.; Lu, W.; Xu, Q. Pairwise cross pattern: A color-LBP descriptor for content-based image retrieval. In Proceedings of the Nineteenth Pacific Rim Conference on Multimedia, Hefei, China, 21–22 September 2018; pp. 290–300. [Google Scholar]
  23. Dubey, S.R.; Singh, S.K.; Singh, R.K. Multichannel decoded local binary patterns for content-based image retrieval. IEEE Trans. Image Process. 2016, 25, 4018–4032. [Google Scholar] [CrossRef] [PubMed]
  24. Liu, P.; Guo, J.; Chamnongthai, K.; Prasetyo, H. Fusion of color histogram and LBP-based features for texture image retrieval and classification. Inf. Sci. 2017, 390, 95–111. [Google Scholar] [CrossRef]
  25. Somasekar, M.; Sakthivel Murugan, S. Feature extraction of underwater images by combining Fuzzy C-Mean color clustering and LBP texture analysis algorithm with empirical mode decomposition. In Proceedings of the Fourth International in Ocean Engineering (ICOE2018), Chennai, India, 19 February 2018; pp. 453–464. [Google Scholar]
  26. Singh, C.; Walia, E.; Kaur, K.P. Enhancing color image retrieval performance with feature fusion and non-linear support vector machine classifier. Optik 2018, 158, 127–141. [Google Scholar] [CrossRef]
  27. Feng, Q.; Hao, Q.; Sbert, M.; Yi, Y.; Wei, Y.; Dai, j. Local parallel cross pattern: A color texture descriptor for image retrieval. Sensors 2019, 19, 315. [Google Scholar] [CrossRef] [PubMed]
  28. Agarwal, M.; Singhal, A.; Lall, B. Multi-channel local ternary pattern for content-based image retrieval. Pattern Anal. Appl. 2019, 22, 1–12. [Google Scholar] [CrossRef]
  29. Bianconi, F.; González, E. Counting local n-ary patterns. Pattern Recognit. Lett. 2018, 177, 24–29. [Google Scholar] [CrossRef]
  30. Reta, C.; Cantoral-Ceballos, J.A.; Solis-Moreno, I.; Gonzalez, J.A.; Alvarez-Vargas, R.; Delgadillo-Checa, N. Color uniformity descriptor: An efficient contextual color representation for image indexing and retrieval. J. Vis. Commun. Image Represent. 2018, 54, 39–50. [Google Scholar] [CrossRef]
  31. Wan, X.; Kuo, C.C. Content-based image retrieval using multiresolution histogram representation. In Proceedings of the SPIE: Digital Image Storage and Archiving Systems, Philadelphia, PA, USA, 21 November 1995; pp. 312–324. [Google Scholar]
  32. Liu, G.H.; Li, Z.Y.; Zhang, L.; Xu, Y. Image retrieval based on micro-structure descriptor. Pattern Recognit. 2011, 44, 2123–2133. [Google Scholar] [CrossRef] [Green Version]
  33. Liu, G.H.; Yang, J.Y. Content-based image retrieval using color difference histogram. Pattern Recognit. 2013, 46, 188–198. [Google Scholar] [CrossRef]
  34. Wan, X.; Kuo, C.C. Color distribution analysis and quantization for image retrieval. In Proceedings of the SPIE: Storage and Retrieval for Still Image and Video Databases IV, San Jose, CA, USA, 13 March 1996; pp. 8–17. [Google Scholar]
  35. Duda, R.O.; Hart, P.E. Pattern Classification and Scene Analysis; Wiley: New York, NY, USA, 1973; pp. 37–43. [Google Scholar]
  36. Wan, X.; Kuo, C.C. A new approach to image retrieval with hierarchical color clustering. IEEE Trans. Circ. Syst. Vid. 1998, 8, 628–643. [Google Scholar] [CrossRef]
  37. Hurvich, L.M.; Jameson, D. An opponent-process theory of color vision. Psychol. Rev. 1957, 64, 384–404. [Google Scholar] [CrossRef]
  38. Gonzalez, R.C.; Woods, R.E. Digital Image Processing, 3rd ed.; Publishing House of Electronics Industry: Beijing, China, 2010; pp. 455–456. ISBN 9787121102073. [Google Scholar]
  39. Caltech-256 Image Set. Available online: http://www.vision.caltech.edu/Image_Datasets/Caltech256/ (accessed on 8 August 2017).
  40. Zhang, M.; Zhang, K.; Feng, Q.; Wang, J.; Kong, J. A novel image retrieval method based on hybrid information descriptors. J. Vis. Commun. Image Represent. 2014, 25, 1574–1587. [Google Scholar] [CrossRef]
  41. Standring, S. Gray’s Anatomy: The Anatomical Basis of Clinical Practice, 41st ed.; Elsevier Limited: New York, NY, USA, 2016; pp. 686–708. ISBN 9780702068515. [Google Scholar]
  42. Guo, J.; Prasetyo, H.; Wang, N. Effective image retrieval system using dot-diffused block truncation coding features. IEEE Trans. Multimed. 2015, 17, 1576–1590. [Google Scholar] [CrossRef]
  43. Libor Spacek’s Facial Image Databases “Face 95 Image Database”. Available online: https://cswww.essex.ac.uk/mv/allfaces/faces95.html (accessed on 8 August 2014).
  44. ETH Zurich. Available online: http://www.vision.ee.ethz.ch/en/datasets/ (accessed on 8 August 2014).
  45. Zurich Buildings Database. Available online: http://www.vision.ee.ethz.ch/en/datasets/ (accessed on 8 August 2014).
  46. MIT Vision and Modeling Group. Available online: http://vismod.media.mit.edu/pub/ (accessed on 12 August 2014).
  47. KTH-TIPs2 Image Database. Available online: http://www.nada.kth.se/cvap/databases/kth-tips/download.html (accessed on 12 August 2014).
  48. Corel 1000 Image Database. Available online: http://wang.ist.psu.edu/docs/related/ (accessed on 12 August 2014).
  49. Guo, J.; Prasetyo, H. Content-based image retrieval using features extracted from halftoning-based block truncation coding. IEEE Trans. Image Process. 2015, 24, 1010–1024. [Google Scholar] [PubMed]
  50. Guo, J.; Prasetyo, H.; Su, H. Image indexing using the color and bit pattern feature fusion. J. Vis. Commun. Image Represent. 2013, 24, 1360–1379. [Google Scholar] [CrossRef]
  51. Orchard, M.T.; Bouman, C.A. Color quantization of Images. IEEE Trans. Signal. Process. 1991, 39, 2677–2690. [Google Scholar] [CrossRef]
  52. Kolesnikov, A.; Trichina, E.; Kauranne, T. Estimating the number of clusters in a numerical data set via quantization error modeling. Pattern Recognit. 2015, 48, 941–952. [Google Scholar] [CrossRef]
  53. Chen, Y.; Chang, C.; Lin, C.; Hsu, C. Content-based color image retrieval using block truncation coding based on binary ant colony optimization. Symmetry 2019, 11, 21. [Google Scholar] [CrossRef]
  54. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. In Proceedings of the Advances Neural Information Processing Systems, Lake Tahoe, NV, USA, 3–6 December 2012; pp. 1097–1105. [Google Scholar]
  55. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinvovich, A. Going deeper with convolutionals. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  56. Chatfield, K.; Simonyan, K.; Vedaldi, A.; Zisserman, A. Return of the devil in the details: Delving deep into convolutional nets. In Proceedings of the British Machine Vision Conference 2014, Nottinghamshire, UK, 1–5 September 2014. [Google Scholar]
  57. Napoletano, P. Hand-crafted vs. learned descriptors for color texture classification. In Proceedings of the International Workshop on Computational Color Imaging, Milan, Italy, 29–31 March 2017; pp. 259–271. [Google Scholar]
  58. Napoletano, P. Visual descriptors for content-based retrieval of remote-sensing images. Int. J. Romote Sens. 2018, 39, 1043–1376. [Google Scholar] [CrossRef]
  59. Yi, Y.; Zhou, W.; Liu, Q.; Luo, G.; Wang, J.; Fang, Y.; Zheng, C. Ordinal preserving matrix factorization for unsupervised feature selection. Signal. Process. Image Commun. 2018, 67, 118–131. [Google Scholar] [CrossRef]
  60. Cernadas, E.; Fernández-Delgado, M.; González-Rufino, E. Influence of normalization and color space to color texture classification. Pattern Recognit. 2017, 61, 120–138. [Google Scholar] [CrossRef]
  61. Yi, Y.; Wang, J.; Zhou, W.; Zheng, C.; Kong, J.; Qiao, S. Non-Negative matrix factorization with locality constrained adaptive graph. IEEE Trans. Circ. Syst. Vid. 2019. [Google Scholar] [CrossRef]
  62. Liu, S.; Wu, J.; Feng, L.; Qiao, H.; Liu, Y.; Lou, W.; Wang, W. Perceptual uniform descriptor and ranking on manifold for image retrieval. Inf. Sci. 2017, 424, 235–249. [Google Scholar] [CrossRef]
  63. Chum, O.; Mikulik, M.; Perdoch, M.; Matas, J. Total recall II: Query expansion revisited. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Colorado Springs, CO, USA, 20–25 June 2011; pp. 889–896. [Google Scholar]
Figure 1. The frequency of pixels on: (a,b) Caltech-256 and (c,d) 10% of Caltech-256.
Figure 1. The frequency of pixels on: (a,b) Caltech-256 and (c,d) 10% of Caltech-256.
Applsci 09 02211 g001
Figure 2. The details of the five-level color quantizer.
Figure 2. The details of the five-level color quantizer.
Applsci 09 02211 g002
Figure 3. An example of the proposed quantization scheme: (a) The values and its quantization levels in the CIELAB color model; (b) The quantization in the L* component; (c) The quantization in the a* component; and (d) The quantization in the b* component.
Figure 3. An example of the proposed quantization scheme: (a) The values and its quantization levels in the CIELAB color model; (b) The quantization in the L* component; (c) The quantization in the a* component; and (d) The quantization in the b* component.
Applsci 09 02211 g003
Figure 4. Schematic diagram of the local ternary cross structure pattern (LTCSP).
Figure 4. Schematic diagram of the local ternary cross structure pattern (LTCSP).
Applsci 09 02211 g004
Figure 5. Schematic diagram of the proposed retrieval system.
Figure 5. Schematic diagram of the proposed retrieval system.
Applsci 09 02211 g005
Figure 6. Some samples from six datasets: (a) Face-95; (b) ETHZ; (c) ZuBuD; (d) VisTex; (e) KTH-2a; and (f) Corel-1000.
Figure 6. Some samples from six datasets: (a) Face-95; (b) ETHZ; (c) ZuBuD; (d) VisTex; (e) KTH-2a; and (f) Corel-1000.
Applsci 09 02211 g006
Figure 7. The top-10 returned images using nine methods (the 1st row using OCLBP, the 2nd row using IOCLBP, the 3rd row using maLBP, the 4th row using mdLBP, the 5th row using OC-LBP + CH, the 6th row using LPCP, the 7th row using LTCSP, the 8th row using LTCSPuni, and the 9th row using LTCSPri) on: (a) Face-95; (b) ETHZ; (c) ZuBuD; (d) VisTex; (e) KTH-2a; and (f) Corel-1000 respectively.
Figure 7. The top-10 returned images using nine methods (the 1st row using OCLBP, the 2nd row using IOCLBP, the 3rd row using maLBP, the 4th row using mdLBP, the 5th row using OC-LBP + CH, the 6th row using LPCP, the 7th row using LTCSP, the 8th row using LTCSPuni, and the 9th row using LTCSPri) on: (a) Face-95; (b) ETHZ; (c) ZuBuD; (d) VisTex; (e) KTH-2a; and (f) Corel-1000 respectively.
Applsci 09 02211 g007aApplsci 09 02211 g007b
Figure 8. The comparisons of the APR rates among the proposed LTCSP, LTCSPuni and LTCSPri descriptors and the emerging deep learning (DL)-based models on six datasets.
Figure 8. The comparisons of the APR rates among the proposed LTCSP, LTCSPuni and LTCSPri descriptors and the emerging deep learning (DL)-based models on six datasets.
Applsci 09 02211 g008
Table 1. Summary of six image datasets.
Table 1. Summary of six image datasets.
NumberNameImage SizeClassImages in Each ClassImages TotalFormat
1Face-95180 × 20072201440JPG
2ETHZ320 × 240535265BNG
3ZuBuD640 × 48020151005JPG
4VisTex128 × 1284016640PPM
5KTH-2a200 × 20011396/4324608BNG
6Corel-1000384 × 256
or
256 × 384
101001000JPG
Table 2. The best APR (%) values of LTCSP, LTCSPuni and LTCSPri with the optimal levels (Aa*, Ab*) over six datasets.
Table 2. The best APR (%) values of LTCSP, LTCSPuni and LTCSPri with the optimal levels (Aa*, Ab*) over six datasets.
MethodPerformanceData Set
Face-95ETHZZuBuDVisTexKTH-2aCorel-1000
LTCSP(Aa*, Ab*)(5, 5)(3, 4)(5, 5)(4, 2)(5, 3)(3, 3)
APR (%)94.3190.5785.6398.5698.9683.94
LTCSPuni(Aa*, Ab*)(5, 5)(5, 3)(5, 4)(3, 2)(4, 3)(3, 2)
APR (%)97.1994.0485.9797.8199.1582.83
LTCSPri(Aa*, Ab*)(5, 5)(5, 3)(5, 5)(3, 3)(4, 5)(3, 2)
APR (%)97.3994.7286.1197.5399.1982.33
Table 3. The mean square errors obtained by the hierarchical quantization schemes of different descriptors on six datasets.
Table 3. The mean square errors obtained by the hierarchical quantization schemes of different descriptors on six datasets.
MethodData Set
Face-95ETHZZuBuDVisTexKTH-2aCorel-1000
CH6085.916714.117483.356716.295858.416386.27
Lab-CVV1024.931430.971622.651358.551090.411276.06
CDH391.83205.9990.02255.97487.37370.97
MSD385.29200.0083.31249.44481.69365.06
LTCSP48.2864.1517.7557.9795.0392.12
LTCSPuni48.2864.2217.7858.7795.07102.21
LTCSPri48.2864.2217.7558.7794.88102.21
Table 4. The evaluations the APR and ARR rates resulting from the proposed methods and the LBP-based methods on six datasets.
Table 4. The evaluations the APR and ARR rates resulting from the proposed methods and the LBP-based methods on six datasets.
MethodPerformanceDate Set
Face-95ETHZZuBuDVisTexKTH-2aCorel-1000
LBPAPR (%)63.4549.2861.4593.3791.5671.86
ARR (%)31.7349.2861.4558.362.197.19
LBPuniAPR (%)58.2544.3854.6390.8388.5668.94
ARR (%)29.1244.3854.6356.772.116.89
LBPriAPR (%)59.7845.9653.0789.7585.5266.73
ARR (%)29.8945.9653.0756.092.046.67
LTCSPAPR (%)94.3190.5785.6398.5698.9683.94
ARR (%)47.1590.5785.6361.602.368.39
LTCSPnuiAPR (%)97.1994.0485.9797.8199.1582.83
ARR (%)48.6094.0485.9761.132.378.28
LTCSPriAPR (%)97.3994.7286.1197.5399.1982.33
ARR (%)48.6994.7286.1160.962.378.23
Table 5. The evaluations of APR and ARR rates obtained by the proposed descriptors and six state-of-the-art color LBP descriptors on six datasets.
Table 5. The evaluations of APR and ARR rates obtained by the proposed descriptors and six state-of-the-art color LBP descriptors on six datasets.
MethodPerformanceDate Set
Face-95ETHZZuBuDVisTexKTH-2aCorel-1000
OCLBPAPR (%)64.4042.5756.4292.4290.6268.86
ARR (%)32.2042.5756.4257.762.166.89
IOCLBPAPR (%)66.4745.5161.0595.5994.2673.01
ARR (%)33.2445.5161.0559.752.257.30
maLBPAPR (%)67.9455.1759.4695.8092.2574.45
ARR (%)33.9755.1759.4659.872.207.45
mdLBPAPR (%)72.9761.4361.8597.0594.8876.02
ARR (%)36.4961.4361.8560.652.267.60
OC-LBP + CHAPR (%)80.5078.0463.9892.2095.3174.94
ARR (%)40.2578.0463.9857.632.277.49
LPCPAPR (%)92.3388.1584.8298.3398.7782.85
ARR (%)46.1688.1584.8261.462.368.29
LTCSPAPR (%)94.3190.5785.6398.5698.9683.94
ARR (%)47.1590.5785.6361.602.368.39
LTCSPnuiAPR (%)97.1994.0485.9797.8199.1582.83
ARR (%)48.6094.0485.9761.132.378.28
LTCSPriAPR (%)97.3994.7286.1197.5399.1982.33
ARR (%)48.6994.7286.1160.962.378.23
Table 6. Feature dimensionality (d) and memory cost (kB) among the proposed descriptors and six former descriptors.
Table 6. Feature dimensionality (d) and memory cost (kB) among the proposed descriptors and six former descriptors.
MethodFeature Dimensionality (d)Memory Cost (kB)
OCLBP153511.99
IOCLBP307224.00
maLBP10248.00
mdLBP204816.00
OC-LBP + CH1080.84
LPCP844/760/844/616/592/4246.59/5.94/6.59/4.81/4.63/3.31
LTCSP688/496/688/436/544/4485.38/3.88/5.38/3.41/4.25/3.50
LTCSPnui491/347/419/203/299/2033.84/2.71/3.27/1.59/2.34/1.56
LTCSPri468/324/468/288/396/1803.66/2.53/3.66/2.25/3.09/1.41

Share and Cite

MDPI and ACS Style

Feng, Q.; Wei, Y.; Yi, Y.; Hao, Q.; Dai, J. Local Ternary Cross Structure Pattern: A Color LBP Feature Extraction with Applications in CBIR. Appl. Sci. 2019, 9, 2211. https://doi.org/10.3390/app9112211

AMA Style

Feng Q, Wei Y, Yi Y, Hao Q, Dai J. Local Ternary Cross Structure Pattern: A Color LBP Feature Extraction with Applications in CBIR. Applied Sciences. 2019; 9(11):2211. https://doi.org/10.3390/app9112211

Chicago/Turabian Style

Feng, Qinghe, Ying Wei, Yugen Yi, Qiaohong Hao, and Jiangyan Dai. 2019. "Local Ternary Cross Structure Pattern: A Color LBP Feature Extraction with Applications in CBIR" Applied Sciences 9, no. 11: 2211. https://doi.org/10.3390/app9112211

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop