Next Article in Journal
A Comparative Study on Weighted Central Moment and Its Application in 2D Shape Retrieval
Previous Article in Journal
A Minimum-Entropy Based Residual Range Cell Migration Correction for Bistatic Forward-Looking SAR
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Local Structure Descriptor for Color Image Retrieval

Faculty of Software, Fujian Normal University, Qishan Community, Fuzhou 350108, China
Information 2016, 7(1), 9; https://doi.org/10.3390/info7010009
Submission received: 19 November 2015 / Revised: 28 January 2016 / Accepted: 15 February 2016 / Published: 22 February 2016
(This article belongs to the Section Information Processes)

Abstract

:
A novel local structure descriptor (LSD) for color image retrieval is proposed in this paper. Local structures are defined based on a similarity of edge orientation, and LSD is constructed using the underlying colors in local structures with similar edge direction. LSD can effectively combine color, texture and shape as a whole for image retrieval. LSH integrates the advantages of both statistical and structural texture description methods, and it possesses high indexing capability and low dimensionality. In addition, the proposed feature extraction algorithm does not need to train on a large scale training datasets, and it can extract local structure histogram based on LSD. The experimental results on the Corel image databases show that the descriptor has a better image retrieval performance than other descriptors.

1. Introduction

Images are one of the popular media formats for communication and understanding of human society. With the rapid development of internet and multimedia techniques, ever-increasing images are available to the public. Therefore, people desire to get a more efficient image indexing tool. Image search has become one of the crucial issues in computer vision. Generally, people retrieve images in three ways. Text-based methods use keywords that are annotated on images to search image, which are widely used by many company’s applications, such as Google and Baidu. However, this approach needs to label images manually, and manually annotating a large scale of images is time-consuming. Furthermore, the retrieval results may be inaccurate. Content-based image retrieval (CBIR) approach computes similarity between images by using low level features which describe the content of each image, and then the retrieved images have similarity in certain predefined threshold. However, low level features cannot describe image semantic concepts because of a huge semantic gap. In the last few years, many semantic-based image retrieval (SBIR) methods have been proposed, owing to the limits of relevant techniques, SBIR is an open problem so far [1,2]. To this day, CBIR is still one of the most effective image indexing methods.
As is well known, a process competing interactions among neurons will strengthen human visual attention, this means that a few elements of attention are selected and other irrelevant materials are suppressed by neurons [3]. There are close connections between visual features of each image and the human visual system, and the study on how to exploit visual attention mechanism for CBIR is a vital and challenging problem. To simulate visual processing procedure, we propose a novel local structure descriptor (LSD) for image retrieval in this paper. The novelty of this descriptor lies in: (1) due to sensitivity of human visual perception to the orientation and color, we use edge orientation and color of each image to describe the local structure of each image, then we use local structure descriptor to simulate visual processing procedure. (2) The descriptor can not only describe image feature, but also effectively combines color, texture, shape and color layout as a whole simultaneously without any image segmentation, training and learning. (3) The dimensionality of LSD feature vector is low, its time complexity is linear, and LSD is easy to implement.
In the LSD, we initially focus on how to define local structures. We adopt the feature integration method of two-stage model to define local structure [4]. In the pre-attentive stage, we extract primitive features in special modules of feature maps. In the attentive stage, we require focal attention to integrate the separate features to form objects. Subsequently, we move our focus toward describing local structures. We use a local structure histogram (LSH) to extract feature vectors of each image in HSV (Hue, Saturation, Value) color space. The local structure descriptor can describe color, texture and shape of each image simultaneously. Therefore, the LSD combines advantages of both statistical and structural texture descriptors. Finally, we focus on improving the performance of LSD by studying the effects of different color space, parameter settings, gradient operators, and distance metrics. The experimental results of the large image database show that the LSD gets higher precision and recall than representative visual representation descriptors.
Our contributions are as follows:
(a)
We design a new local structure descriptor to simulate human visual perception mechanism. The descriptor combines color, texture, shape and color layout as a whole. The dimensionality of its feature vector is low, which is very appropriate for large-scale image retrieval. The descriptor achieves better accuracy results on standard benchmarks than other descriptors.
(b)
The detail experimental research we carried out adds our understanding of the effects of different color space, parameter settings, and gradient operators of the studied descriptor.
The rest of this paper is organized as follows. Section 2 introduces related works. Section 3 describes the definition of LSD. The extraction of local structure histogram in HSV color space is presented in Section 4. In Section 5, similarity measurement is presented. The proposed descriptor is evaluated and compared with state-of-the-art descriptors on Corel image databases. The conclusion of this paper is presented in Section 7.

2. Related Works

Generally, researchers use different descriptors to represent color, texture and shape of image. They design different algorithms to extract image features and retrieve images. Because of simplicity and effectiveness, the color histogram has been extensively studied and used for image retrieval. Meanwhile, it is invariant to scale and orientation. However, it is difficult for color histogram to represent the image spatial structure. Instead, to exploit image spatial information, other color descriptors have been proposed [5,6]. Texture features present valuable information of smoothness, coarseness and regularity of objects such as fruits, skins, clouds, tree, etc., texture-based schemes have been widely studied in CBIR [7]. Manjunath has designed MPEG-7 edge histogram descriptor (EHD) according to the spatial distribution of edge, which is a very effective texture descriptor for image representation [8]. In addition, the outlines of different objects in an image are usually different, and shape features have been widely used in CBIR for its capability in describing the object outline. The edge orientation autocorrelogram (EOAC) [9] is one of the most successful methods.
Local image features have attracted more and more attention in recent years. Various feature algorithms were presented in order to highlight different properties of an image such as gray and edge, etc. These methods are often based on distribution, and color features can be combined with texture features. Lowe et al. propose a so-called scale-invariant feature transform (SIFT) descriptor, which is an effective algorithm to detect and describe local feature of an image [10]. Wang et al. propose a novel retrieval scheme based on structure elements’ descriptors (SEDs) [11]. SED effectively presents the spatial connection of color and texture by extracting the color and texture features of an image. Liu et al. propose an image feature representing method called color difference histogram (CDH) [12]. CDH is completely different from the existing histogram techniques; it pays more attention to perceptually uniform color difference between color and edge orientation. Murala et al. utilize the orientation map to obtain one order derivative on horizontal and vertical directions based on Local Tetra Patterns (LTrPs) for texture image retrieval and improve retrieval performance [13]. Meng et al. compute the feature of an image and its spatial correlation by using salient points in the patch of an image, and they propose the learning algorithm of multi-instances, this method improves the average retrieval accuracy of image [14]. Wang et al. extract the color feature of an image by Zernike color distributed moments and compute texture feature by contourlet transform. Then, they combine color with texture for image retrieval, and the algorithm obtains better image retrieval performance [15]. Local descriptors are tolerant to varied illuminations, distortions, transformations and are robust to occlusion. Meanwhile, they can describe different characteristics of object appearance or shape without any image segmentation. However, it is still an important challenge how to develop computational models that describe color, texture, shape and spatial structure simultaneously.
In order to improve further image retrieval performance, many feature combination methods and relevant feedback techniques have been proposed recently [16,17,18,19]. Lee et al. propose a novel retrieval scheme that extracts the image feature by fusing Advanced Speed-Up Robust Feature (ASURF) and Dominant Color Descriptor (DCD). The system can run in real-time on iPhone and find a natural color image for mobile image retrieval [16]. Kafai et al. describe Discrete Cosine Transform (DCT) hashing for creating index structures for face descriptors, and this method can efficiently reduce the cost of the linear search and improve retrieval efficiency and accuracy [17]. Yang et al. propose a semi-supervised Local Regression and Global Alignment (LRGA) algorithm for data ranking and a semi-supervised long-term Relevance Feedback (RF) algorithm for using data distribution and the history RF information, and then they integrate the two algorithms into multimedia content analysis and retrieval framework [18]. Spyromitros-Xioufis et al. use the framework of a Vector of Local Aggregated Descriptors (VLAD) and Product Quantization to develop an enhanced framework for large-scale image retrieval, the system significantly improves the performance for image retrieval [19].
Until now, the latest large-scale image retrieval frameworks use the bag-of-words (BoW) descriptor [20]. In these methods, retrieval system extracts local features (usually SIFT [10]) from each image and assigns each feature to the nearest visual word from a visual vocabulary. The feature vector of BoW is high dimensional and sparse for each image. BoW-based methods achieve better the accuracy of image indexing, but they cannot generalize to more than millions images datasets on a single machine owing to high computational complex and memory limits. Recently, a few scalable algorithms have been developed in [21,22,23]. The feature vectors of these methods are more discriminative than BoW, and these feature vectors combine with powerful compression techniques.
One of the most successful algorithms is proposed in [21]. This method uses SIFT features instead of BoW with highly discriminative Fisher Vector [22] or its simpler variant—VLAD [23]. Exploiting these optimized vector representation, greatly better retrieval accuracies are achieved.
In recent years, deep learning has reported encouraging results on CBIR tasks. For example, Wan et al. use a framework of deep learning for CBIR, they find that deep learning can mimic the human brain that is organized in a deep architecture and processes information through multiple stages of transformation and representation [24]. Ng et al. extract convolution features from different layers of the deep convolutional networks and adopt VLAD encoding to encode features into a vector, they conclude that intermediate layers or higher layers with finer scales produce better results for image retrieval [25]. Zhang et al. combine deep convolutional neural network (CNN) with learning hash functions, deep CNN is utilized to train the model, discriminative image features and hash functions are simultaneously optimized [26]. Lin et al. propose an effective deep learning framework to learn binary hash codes by employing a hidden layer for representation of the latent concepts that dominate the class labels in a point-wised manner [27]. Deep learning has many advantages. One of them is it can learn hierarchical features for different semantic abstraction, which effectively improve the performance of CBIR. However, deep learning needs some tricks in parameters tuning, and the feature dimensionality of the last layer is very high, which may affect its application in practice.

3. Local Structure Descriptors

Directly comparing image content is impractical for image search because the contents of image are significantly different. However, the structural information of the same class images often shows some similarity. We may regard that a semantic of a natural image is made up of many common local structures. If these local structures could be achieved and presented effectively, we could use them as a common base for different image matching. It is very important for image retrieval if shape is extracted along with color, texture and color layout of each image.
One key problem of the local structure descriptor is how to define local structure. Let g ( x , y ) be a color image of size M × N. Firstly, to detect local structure, we convert RGB color space into HSV color space, we quantize the color image into 72 colors in HSV color space, we use nine structure templates to detect edge orientation. The nine structure masks are nine 2 × 2 matrixes of different directions, which are displayed in Figure 1. Then, in the edge orientation map, we define the local structure. Thirdly, we construct the LSD using underlying color in local structures. Finally, we use local structure histogram to describe the LSD to represent image feature. The LSD not only could effectively represent each image, meanwhile, it could extract shape, texture, color and color layout feature of each image simultaneously. We present these steps in detail as follows:

3.1. Selection and Quantization of Color Space

HSV space is widely used for image feature extraction; it is made up of three components in terms of Hue (H), Saturation (S) and Value (V). HSV is usually modeled as a cylinder, the H component represents color type, its value is between 0–360°, with red at 0°, green at 120°, and blue at 240°. The S component describes the relative purity of color, its value is between 0–1. The V component represents the brightness of the color, its value is also between 0–1.
Human color perception could be imitated well by HSV color space l. In this paper, we select HSV color space, the color image is quantized into 72 bins. Specifically, we quantize the H, S and V components into 8, 3, 3 bins respectively. Hence, we obtain 8 × 3 × 3 = 72 color combinations. Let I ( x , y ) be the quantized color image, I ( x , y ) = i , i { 0 , 1 , 2 , ... , 71 } .

3.2. Edge Direction Detection

Edge direction plays an important role in image recognition. The boundaries and texture structure of object is represented by the direction map of each image, the direction map provides a large amount of semantic concepts for the image. Hence, the detection of edge direction is a vital processing operation. Edge direction is detected by many existing edge detection operators such as Sobel operator, Canny operator, etc. However, these detectors can only detect gray level images and cannot detect color image because a color image has three color channels. In this subsection, the following methods are used to detect edge orientation for color image in HSV color space.
In Cartesian space, we define the dot product of vectors a ( x 1 , y 1 , z 1 ) and b ( x 2 , y 2 , z 2 ) as follows:
a b = x 1 x 2 + y 1 y 2 + z 1 z 2
so that
cos ( a , b ) = a b | a | | b | = x 1 x 2 + y 1 y 2 + z 1 z 2 x 1 2 + y 1 2 + z 1 2 x 2 2 + y 2 2 + z 2 2
Because cylinder coordinate system is used in HSV color space, to calculate the angle between vectors, HSV color space should be transformed into the Cartesian coordinate system. ( H , S , V ) represents a point in HSV color space, ( H , S , V ) represents the transformation of ( H , S , V ) in Cartesian space, H = S cos ( H ) , S = S sin ( H ) and V = V . The Sobel operator is applied to each of the H , S , V channels of a color image in Cartesian space. We use two vectors a ( H x , S x , V x ) and b ( H y , S y , V y ) to denote the gradients along x and y direction, where the gradient in H channel along horizontal direction is denoted by H x , and so on. Their dot product and norm can be defined as:
| a | = ( H x ) 2 + ( S x ) 2 + ( V x ) 2
| b | = ( H y ) 2 + ( S y ) 2 + ( V y ) 2
and
a b = H x H y + S x S y + V x V y
The angle between a and b is:
cos ( a , b ) = a b | a | | b |
θ = arccos ( a , b ) = arccos [ a b | a | | b | ]
After we calculate the edge direction of each pixel, we quantize the edge orientation into m b i n s , and m { 6 , 12 , 18 , 24 , 30 , 36 } . We use θ ( x , y ) to represent the edge orientation map, as θ ( x , y ) = ϕ , ϕ { 0 , 1 , 2 , ... , m 1 } .

3.3. Definition and Extraction of Local Structure

The human visual system is sensitive to color and direction. Direction is a strong cue for topics described about an image. Obvious direction usually means definite pattern. However, there is no strong orientation and clear structure or specific pattern in many natural scenes. Although different contents are showed in the natural image, the common elements are included in these images. Different combination and spatial distribution of these fundamental elements lead to different local structures or patterns.
In order to find local structure with similar attributes such as edge direction and color, we use edge orientation map θ ( x , y ) to detect local structure, because edge direction is not sensitive to color and illumination variation, and it is robust to translation, scaling and small rotation. Let θ ( x , y ) be an edge direction map of size M × N, we shift nine local structure templates shown in Figure 1 from top to bottom and left to right throughout orientation map to detect fundamental local structure respectively. In order to obtain a single local structure map of whole image, we use the steps described to detect as follows:
(1)
Beginning from the point (0, 0), we shift 2 × 2 local structure template (a) from top-to-bottom and left-to-right throughout edge direction map θ ( x , y ) with a step length of two pixels along both vertical and horizontal directions. If the values of θ ( x , y ) in the corresponding structure template are equal, the values will be saved, otherwise, the values will be set zero. Then, we will obtain a local structure map C 1 ( x , y ) .
(2)
We use the other eight templates (b), (c), (d), (e), (f), (g), (h) and (i) throughout edge orientation map θ ( x , y ) to conduct the same operations as (1) step, respectively, we will obtain eight local structure maps C 2 ( x , y ) , C 3 ( x , y ) , C 4 ( x , y ) , C 5 ( x , y ) , C 6 ( x , y ) , C 7 ( x , y ) , C 8 ( x , y ) , C 9 ( x , y ) .
(3)
Using C(x,y) to denote the final local structure map, C(x,y) is obtained by fusing nine local structure maps based on the following rules:
C ( x , y ) = max { C 1 ( x , y ) , C 2 ( x , y ) , C 3 ( x , y ) , C 4 ( x , y ) , C 5 ( x , y ) , C 6 ( x , y ) , C 7 ( x , y ) , C 8 ( x , y ) , C 9 ( x , y ) }
The extraction and fusing process of the above local structure map is illustrated in Figure 2.

4. Feature Extraction

After the local structure map is extracted, the next step is to find a way to describe its feature. However, a critical problem is that we have to find a way that stimulates the visual system to some extent. Cortical visual system hypothesis is a framework for understanding cortical visual systems and visuomotor systems [4]. The framework is often called as the “what/where” and “what/how” pathways. It depicts two information processing streams which originate in the occipital cortex. The “what/where” pathways are responsible for color, shape, location and motion, respectively. The “how” pathway could be regarded as a generalization of the “where” pathway. The above analyses give an insight into how the relationship between image feature representation and visual attention mechanism is determined.
To present the content of image via imitating the visual attention mechanism to a certain extent, we consider the pixels with similar color in the local structure images as the stimuli in the visual field. We carry out the LSH method in accordance with the following rules: (1) what the local structure is; (2) where the local structure is; (3) how the local structure correlates with others. (1) and (2) are about the semantics of image based on perceptive space, and (3) is about the spatial representation of the local structures in a human brain.
After the local structure map C(x,y) is extracted from the edge direction map θ ( x , y ) , it is used as a mask to obtain the underlying colors from the quantized image I(x,y). We will achieve the local structure map by reserving only the colors in the local structure image and other colors outside the image are set to empty. In the subsection, we apply f ( x , y ) to denote the local structure image. The value of local structure map f ( x , y ) can be denoted as f ( x , y ) = i , i { 0 , 1 , 2 , ... , 71 } . We use local structure histogram LSH to extract feature vector of the image in HSV color space. In order to extract color, texture and shape feature of the f ( x , y ) simultaneously, LSH will be calculated on every bin in HSV color space. LSH will be extracted by using two steps strategy described as follows:
(1)
When f ( x , y ) = i , i { 0 , 1 , 2 , ... , 71 } , count the number of LSD on nine LSD maps. In particular, when the local structure descriptor that denotes no direction has been counted, the other eight local structure descriptors should be counted again, because no direction means that every direction is possible.
(2)
Calculate LSH based on the number of LSD.
The full color image is composed of local structure images. The attribute of the human visual system is fitted by local structure maps, because local structure images choose only a few points of attention and suppress the irrelevant material. Hence, the LSH not only depicts “what” and “how” colors and directions are utilized, but it also designates “where” and “how” the color and direction components are distributed within a certain range in the visual scenery. Hence, the different combination and spatial distribution of the local structures could be described by LSD, LSD could describes strong discriminative features such as texture, color, shape features and color layout information.
The process of extracting LSH is illustrated in Figure 3. Figure 3a displays the procedure of extracting LSH on an image when the value is 1, and the outcome of LSH is {1, 0, 1, 0, 0, 0, 0, 0, 0}. Figure 3b–e are the procedure of extracting LSH when the color values are 2, 3, 4, and 5. The final LSH of the image is {1, 0, 1, 0, 0, 0, 0, 0, 0, 2, 1, 0, 1, 0, 0, 1, 0, 0, 1, 2, 2, 1, 1, 1, 1, 2, 1, 0, 1, 1, 0, 0, 0, 0, 0, 0, 1, 1, 1, 1, 1, 1, 1, 1, 1}.
In a natural image, LSH is extracted from 72 bins in HSV color space, all the nine local structure elements are extracted in each bin, respectively, so there are 72 × 9 = 648 bins extracted in HSV color space.

5. Similarity Measurement

For each image in a database, M bins feature vector T = [ T 1 , T 2 , ... , T M ] of the LSH is computed according to the above described method and stored in the database. Let Q = [ Q 1 , Q 2 , ... , Q M ] be M bins, the feature vector of a query image, the similarity measurement between a query image and any an image in the database can be defined as follows:
D ( Q , T ) = i = 1 M | Q i T i |
The above formula is called L 1 distance, which is the accumulative summary of two vectors difference, without square or square root operation on vectors. Hence, its time complexity is linear and is very suitable for large-scale image search.

6. Experiments and Results Analysis

6.1. Image Database

Two Corel image databases are used to test image retrieval performance in this paper, which are widely used for CBIR. The first one is Corel-1000 dataset, which includes 10 categories. There are 1000 natural images from diverse semantics such as scenes, horses, elephants, human, buses, flowers, buildings, mountains, foods and dinosaurs. The second one is Corel-10000 dataset, which includes 100 categories. There are 10,000 natural images from diverse semantics such as beaches, buses fishes and sunsets, etc. Experimental images contain different topics.
In following experiments, 10 images from each category in Corel-1000 dataset are randomly selected and used as query images. The precision and recall percentage for each category are calculated. Then, average precision-recall pair percentage is calculated by the precision-recall pair percentage of 10 random images. In Corel-10000 dataset, we randomly choose 20 categories from 100 categories image. Then, 10 images are randomly selected from each category to use them as query images, and average precision and recall are computed.

6.2. Performance Measurements

We adopt Precision and Recall to evaluate the performance of the proposed method; these two metrics are the most commonly used metrics for evaluating image retrieval performance. Precision and Recall is defined as follows:
P r e c i s i o n = I N N
Re c a l l = I N M
where I N is the number of similar image retrieved, M is the total number of similar images in image database, and N is the total number of images retrieved. In the following experiments, to get the precision and recall values in Table 1, Table 2 and Table 3, N and M are set 12 and 100, respectively, for two datasets.

6.3. Retrieval Results

In the following experiments, we adopt a different number of quantization level for texture and color to evaluate the retrieval performance in RGB and HSV color space. The RGB color space is quantized to 16, 32, 64 and 128 bins, respectively, The HSV color space is quantized to 72, 108, 128 and 192 bins, respectively, and the texture orientation is quantized to 6, 12, 18, 24, 30 and 36 bins, respectively. As can be seen from the results in Table 1 and Table 2, the proposed LSD has better performance in HSV color space than in RGB color space.
In order to test the results of the proposed direction detection operator, we use several classical edge detectors to detect gradient magnitude and direction, and the experimental results are listed in Table 3. We should note that the proposed operator works on the full color image, while the other four operators work on the gray level images. It can be seen from the retrieval results in Table 3 that the proposed direction operator achieves better results because it utilizes the color information that is ignored by the other detectors in direction detection.
We then validate the performance of the proposed distance metric and other popular distance metrics or similarity measurements. It can be seen from the retrieval results in Table 4 that our distance metric achieves better retrieval results than other distance metrics or similarity measurements. Meanwhile, comparing the results of L 1 distance and L 2 distance, we note that they have comparative performance with almost exactly precision and recall. However, L 1 distance is very simple to calculate with linear time requirement, unlike L 2 distance, it requires square or square root operations, which is costly for vector operation. Hence, our L 1 distance saves much computational cost and is very suitable for large-scale image database.
We compare the LSD representation with the state-of-the-art representations such as BoW [20], VLAD [22] and compressed Fisher kernel [23] on Corel-1000 and Corel-10000 datasets. These state-of-the-art representations are all parameterized by parameter k. Parameter k represents the number of centroids for BoW and VLAD, and to the number of mixture components in the Fisher kernel representation. We set k = 8000 for BoW and k = 64 for VLAD and compressed Fisher kernel, respectively. We use the same feature detection and description procedure as in [28] (Hessian–Affine extractor and the SIFT descriptor) since the code and the features are available online. We reduce the dimensionalities of these features from D = 8000 to D = 648 through PCA (Principal Component Analysis). We use the Flickr60k dataset to learn the PCA and the GMM (Gaussian Mixture Model) vocabularies. The Fisher vectors used in our experiments were computed by taking the gradient with respect to the mean parameters only. The average retrieval precision and recall curves of the above local descriptors are plotted in Figure 4. It can be seen from the retrieval results in Figure 4, LSD obtains better retrieval results than BoW [20], VLAD and Compressed Fisher kernel [22,23]. Though BoW is a good local feature representation, it ignores spatial information, BoW achieves the lowest retrieval results in the above descriptors. Fisher kernel and VLAD combine the strengths of generative and discriminative approaches, their feature representations are more discriminative and get better results than BoW. LSD imitates human visual mechanism to some extent, and it combines color, texture, shape and color layout as a whole. Meanwhile, LSD and its feature vector can be automatically extracted without any image segmentation, training and learning, they can be implemented simply, and the dimensionality of its feature vector is low, therefore, LSD is very suitable for large-scale image retrieval. The descriptor outperforms the previous best reported accuracy on standard benchmarks.
Figure 5 and Figure 6 show the retrieval examples on Corel-1000 and Corel-10000 datasets. In Figure 5, the query image is a bus image, it has clear shape feature and similar color. All the top 12 retrieved images belong to the bus category, which shows a good match of shape and color to query image. In Figure 6, the query image is a horse image, all of the top retrieved images are the horse category, which shows a good match of texture, shape and color to query the image.

7. Conclusions

A novel local structure descriptor for color image retrieval is proposed in this paper. LSD can effectively describe the shape, texture, color, and color layout of each image without any image segmentation, learning and training. Hence, its implementation is easy and its time complexity is linear. LSD can be regarded as a generalized visual attribute descriptor because the descriptor imitates human visual mechanisms to some extent. The dimensionality of LSD feature vector is only 648, which is very suitable for large-scale image search. Our experimental results on two Corel image databases show that the proposed descriptor has strong discriminative capability for color, texture and shape features and outperforms the BoW, VLAD and Compressed Fisher kernel descriptors significantly.

Acknowledgments

This research is partially sponsored by the National Natural Science Foundation of China (No. 60805016), a major project of the Industrial Science and Technology of Fujian Province of China (No. 2013H0020).

Conflicts of Interest

The author declares no conflict of interest.

References

  1. Liu, Y.; Zhang, D.; Lu, G.; Ma, W.Y. A survey of content-based image retrieval with high-level semantics. Pattern Recognit. 2007, 40, 262–282. [Google Scholar] [CrossRef]
  2. Zhang, D.S.; Islam, M.M.; Lu, G.J. A review on automatic image annotation techniques. Pattern Recognit. 2012, 45, 346–362. [Google Scholar] [CrossRef]
  3. Desimone, R. Visual attention mediated by biased competition in extrastriate visual cortex. Philos. Trans. R. Soc. B 1998, 353, 1245–1255. [Google Scholar] [CrossRef] [PubMed]
  4. Reily, R.C.O. The what and how of prefrontal cortical organization. Trends Neurosci. 2010, 33, 355–361. [Google Scholar]
  5. Huang, J.; Kumar, S.R.; Mitra, M. Image indexing using color correlograms. In Proceedings of the 1997 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Juan, Puerto Rico, 17–19 June 1997; pp. 762–768.
  6. Quellec, G.; Lamard, M.; Cazuguel, G.; Cochener, B.; Roux, C. Fast wavelet-based image characterization for highly adaptive image retrieval. IEEE Trans. Image Process. 2012, 21, 1613–1623. [Google Scholar] [CrossRef] [PubMed]
  7. Gonzalez, R.C.; Woods, R.E. Digital Image Processing, 3rd ed.; Prentice Hall: Upper Saddle River, NJ, USA, 2007; pp. 395–398. [Google Scholar]
  8. Manjunath, B.S.; Ohm, J.R.; Vasudevan, V.V.; Yamada, A. Color and texture descriptors. IEEE Trans. Circuit Syst. Video Technol. 2001, 11, 703–715. [Google Scholar] [CrossRef]
  9. Mahmoudi, F.; Shanbehzadeh, J. Image retrieval based on shape similarity by edge orientation autocorrelogram. Pattern Recognit. 2003, 36, 1725–1736. [Google Scholar] [CrossRef]
  10. Lowe, D.G. Distinctive image features from scale-invariant key points. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar] [CrossRef]
  11. Wang, X.Y.; Wang, Z.Y. A novel method retrieval based on structure elements’ descriptor. J. Vis. Commun. Image Represent. 2013, 24, 63–74. [Google Scholar] [CrossRef]
  12. Liu, G.H.; Yang, J.Y. Content-based image retrieval using color difference histogram. Pattern Recognit. 2013, 46, 188–198. [Google Scholar] [CrossRef]
  13. Murala, S.; Maheshwari, R.P.; Balasubramanian, R. Local tetra patterns: a new feature descriptor for content-based image retrieval. IEEE Trans. Image Process. 2012, 21, 2874–2886. [Google Scholar] [CrossRef] [PubMed]
  14. Meng, F.J.; Guo, B.L.; Wu, X.X. Localized Image Retrieval Based on Interest Points. Proced. Eng. 2012, 29, 3371–3375. [Google Scholar]
  15. Wang, X.Y.; Yang, H.Y.; Li, D.M. A New Content-Based Image Retrieval Technique Using Color and Texture Information. Comput. Electr. Eng. 2013, 39, 746–761. [Google Scholar] [CrossRef]
  16. Lee, Y.H.; Kim, Y. Efficient image retrieval using advanced SURF and DCD on mobile platform. Multimed. Tools Appl. 2015, 74, 2289–2299. [Google Scholar] [CrossRef]
  17. Kafai, M.; Eshghi, K.; Bhanu, B. Discrete Cosine Transform Locality-Sensitive Hashes for Face Retrieval. IEEE Trans. Multimed. 2014, 16, 1090–1103. [Google Scholar] [CrossRef]
  18. Yang, Y.; Nie, F.P.; Xu, D. A Multimedia Retrieval Framework based on Semi-Supervised Ranking and Relevance Feedback. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 34, 723–742. [Google Scholar] [CrossRef] [PubMed]
  19. Spyromitros-Xioufis, E.; Papadopoulos, S. A Comprehensive Study over VLAD and Product Quantization in Large-Scale Image Retrieval. IEEE Trans. Multimed. 2014, 16, 1713–1728. [Google Scholar] [CrossRef]
  20. Sivic, J.; Zisserman, A. Video Google: A text retrieval approach to object matching in videos. In Proceedings of the Ninth IEEE International Conference on Computer Vision, Nice, France, 13–16 October 2003; pp. 1470–1477.
  21. Jegou, H.; Perronnin, F.; Douze, M.; Sanchez, J.; Perez, P.; Schmid, C. Aggregating local image descriptors into compact codes. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 34, 1704–1716. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  22. Jégou, H.; Douze, M.; Schmid, C.; Pérez, P. Aggregating local descriptors into a compact image representation. In Proceedings of the 2010 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), San Francisco, CA, USA, 13–18 June 2010; pp. 3304–3311.
  23. Perronnin, F.; Liu, Y.; Sánchez, J.; Poirier, H. Large-scale image retrieval with compressed fisher vectors. In Proceedings of the 2010 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), San Francisco, CA, USA, 13–18 June 2010; pp. 3384–3391.
  24. Wan, J.; Wang, D.; Hoi, S.C.H.; Wu, P. Deep Learning for Content-Based Image Retrieval: A Comprehensive Study. ACM Int. Conf. Multimed. 2014. [Google Scholar] [CrossRef]
  25. Ng, J.Y.-H.; Yang, F.; Davis, L.S. Exploiting Local Features from Deep Networks for Image Retrieval. In Proceedings of the IEEE International Conference on Vision and Pattern Recognition, Deep Vision Workshop, Boston, MA, USA, 7–12 June 2015; pp. 53–61.
  26. Zhang, R.; Lin, L.; Zhang, R.; Zuo, W.; Zhang, L. Bit-Scalable Deep Hashing with Regularized Similarity Learning for Image Retrieval and Person Re-identification. IEEE Trans. Image Process. 2015, 24, 4766–4779. [Google Scholar] [CrossRef] [PubMed]
  27. Lin, K.; Yang, H.-F.; Hsiao, J.-H.; Chen, C.-S. Deep Learning of Binary Hash Codes for Fast Image Retrieval. In Proceedings of the IEEE International Conference on Vision and Pattern Recognition, Deep Vision Workshop, Boston, MA, USA, 7–12 June 2015; pp. 27–35.
  28. Jegou, H.; Douze, M.; Schmid, C. Hamming embedding and weak geometric consistency for large scale image search. In Computer Vision—ECCV 2008; Springer: Berlin/Heidelberg, Germany, 2008; Volume 5302, pp. 304–317. [Google Scholar]
Figure 1. The definition of structure templates in different directions: (a) 0°; (b) 90°; (c) 45°; (d) 135°; (e) 225°; (f) 270°; (g) 315°; (h) 360°; (i) no direction.
Figure 1. The definition of structure templates in different directions: (a) 0°; (b) 90°; (c) 45°; (d) 135°; (e) 225°; (f) 270°; (g) 315°; (h) 360°; (i) no direction.
Information 07 00009 g001
Figure 2. Local structures’ map extraction and fusion. (a) shows the local structure map extracted C 1 ( x , y ) . Maps C 2 ( x , y ) , C 3 ( x , y ) , C 4 ( x , y ) , C 5 ( x , y ) , C 6 ( x , y ) , C 7 ( x , y ) , C 8 ( x , y ) and C 9 ( x , y ) can be extracted similarly in (b)–(i). (j) shows the fusion of nine Maps to form the final local structure map C ( x , y ) .
Figure 2. Local structures’ map extraction and fusion. (a) shows the local structure map extracted C 1 ( x , y ) . Maps C 2 ( x , y ) , C 3 ( x , y ) , C 4 ( x , y ) , C 5 ( x , y ) , C 6 ( x , y ) , C 7 ( x , y ) , C 8 ( x , y ) and C 9 ( x , y ) can be extracted similarly in (b)–(i). (j) shows the fusion of nine Maps to form the final local structure map C ( x , y ) .
Information 07 00009 g002aInformation 07 00009 g002b
Figure 3. LSH extraction algorithm. (a)–(e) are the process of extracting LSH of nine bins, respectively. The number of which under the local structure is the number of the local structure of every bin.
Figure 3. LSH extraction algorithm. (a)–(e) are the process of extracting LSH of nine bins, respectively. The number of which under the local structure is the number of the local structure of every bin.
Information 07 00009 g003aInformation 07 00009 g003b
Figure 4. The comparison of average retrieval performance of four descriptors. (a) Corel-1000 datasets (b) Corel-10000 datasets.
Figure 4. The comparison of average retrieval performance of four descriptors. (a) Corel-1000 datasets (b) Corel-10000 datasets.
Information 07 00009 g004
Figure 5. Retrieval result for a bus.
Figure 5. Retrieval result for a bus.
Information 07 00009 g005
Figure 6. Retrieval result for horses.
Figure 6. Retrieval result for horses.
Information 07 00009 g006
Table 1. The average retrieval performance of the local structure descriptor (LSD) under diverse color and direction quantization levels on Corel-1000 in RGB color space.
Table 1. The average retrieval performance of the local structure descriptor (LSD) under diverse color and direction quantization levels on Corel-1000 in RGB color space.
Color Quantization LevelTexture Orientation Quantization Level
Precision (%)Recall (%)
6121824303661218243036
12893.4393.3293.1392.8292.9193.129.409.389.369.329.339.35
6492.5092.4291.8691.9791.9992.259.289.279.219.229.229.25
3289.8590.0689.5089.4189.6489.668.968.978.908.918.948.93
1682.1282.6382.3882.5582.2282.358.028.128.098.118.048.05
Table 2. The average retrieval performance of the LSD under diverse color and direction quantization levels on Corel-1000 in HSV color space.
Table 2. The average retrieval performance of the LSD under diverse color and direction quantization levels on Corel-1000 in HSV color space.
Color Quantization LevelTexture Orientation Quantization Level
Precision (%)Recall (%)
6121824303661218243036
19299.0598.9198.7398.4498.6998.5610.3110.3010.2810.2510.2710.26
12899.0998.5798.8598.7398.6098.4910.3310.2610.2810.2810.2510.24
10898.8898.5498.4698.7298.7298.1610.1810.1210.1210.1610.1510.09
7298.2098.2498.3098.4598.5498.1210.0610.0510.0810.0910.1410.05
Table 3. The retrieval results of LSD with different gradient operators for orientation detection.
Table 3. The retrieval results of LSD with different gradient operators for orientation detection.
DatasetsPerformanceGradient Operator
Proposed OperatorSobelRobertLOGPrewitt
Corel-1000Precision (%)98.2097.8397.1596.2197.52
Recall (%)10.069.969.989.819.95
Corel-10000Precision (%)52.2651.8551.5451.1851.62
Recall (%)5.875.755.735.685.74
Table 4. The average results of LSD with different distance metrics.
Table 4. The average results of LSD with different distance metrics.
DatasetsPerformanceDistance or Similarity Measurement
L 1 EuclidianHistogram Intersection
Corel-1000Precision (%)98.2098.2176.88
Recall (%)10.0610.069.72
Corel-10000Precision (%)52.2652.2630.42
Recall (%)5.875.873.16

Share and Cite

MDPI and ACS Style

Zeng, Z. A Novel Local Structure Descriptor for Color Image Retrieval. Information 2016, 7, 9. https://doi.org/10.3390/info7010009

AMA Style

Zeng Z. A Novel Local Structure Descriptor for Color Image Retrieval. Information. 2016; 7(1):9. https://doi.org/10.3390/info7010009

Chicago/Turabian Style

Zeng, Zhiyong. 2016. "A Novel Local Structure Descriptor for Color Image Retrieval" Information 7, no. 1: 9. https://doi.org/10.3390/info7010009

APA Style

Zeng, Z. (2016). A Novel Local Structure Descriptor for Color Image Retrieval. Information, 7(1), 9. https://doi.org/10.3390/info7010009

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop