Next Article in Journal
Shear-Jamming in Two-Dimensional Granular Materials with Power-Law Grain-Size Distribution
Next Article in Special Issue
Information-Theoretic Data Discarding for Dynamic Trees on Data Streams
Previous Article in Journal
The New Genetics and Natural versus Artificial Genetic Modification
Previous Article in Special Issue
Kernel Spectral Clustering for Big Data Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Stochasticity: A Feature for the Structuring of Large and Heterogeneous Image Databases

by
Abdourrahmane M. Atto
1,*,
Yannick Berthoumieu
2 and
Rémi Mégret
2
1
LISTIC (Laboratory of Informatics, Systems, Information and Knowledge Processing), University of Savoie, B.P. 80439, 74944 Annecy le Vieux Cedex, France
2
IMS (Laboratory of Material to System Integration), CNRS UMR 5218, University of Bordeaux, IPB, ENSEIRB-MATMECA, 351 cours de la libération, 33400 Talence, France
*
Author to whom correspondence should be addressed.
Entropy 2013, 15(11), 4782-4801; https://doi.org/10.3390/e15114782
Submission received: 26 August 2013 / Revised: 27 September 2013 / Accepted: 29 October 2013 / Published: 4 November 2013
(This article belongs to the Special Issue Big Data)

Abstract

:
The paper addresses image feature characterization and the structuring of large and heterogeneous image databases through the stochasticity or randomness appearance. Measuring stochasticity involves finding suitable representations that can significantly reduce statistical dependencies of any order. Wavelet packet representations provide such a framework for a large class of stochastic processes through an appropriate dictionary of parametric models. From this dictionary and the Kolmogorov stochasticity index, the paper proposes semantic stochasticity templates upon wavelet packet sub-bands in order to provide high level classification and content-based image retrieval. The approach is shown to be relevant for texture images.

1. Introduction

The diversity of real world images has led researchers to take advantage of various mathematical tools in order to extract relevant image features or retrieve suitable information in images. For instance, probabilistic models, geometry properties and functional analysis have raised much dissertation in the last few decades. In practice, the selection of appropriate features is driven by the class of images of interest. In this paper, the class of images we deal with can be conceptually defined through its departure from the class of regular images.
From the literature, a regular image is defined as either smooth or geometrically regular [1]: the image is composed of different smooth regions delimited by singularity curves. In contrast, a non-regular image is such that: when splitting the image into smaller and smaller subimages (sub-surfaces), almost every subimage is non-regular in that it is expected to contain many delimitation curves.
From the above consideration, a texture can be identified as either geometrically regular, i.e., composed of different or repetitive regions that are smooth, except along their delimitation curves, or non-regular. Geometrically regular textures can be well characterized by local or global regularity measurements, such as Holder exponents [1,2,3,4,5,6,7] or spectral measurements [8,9,10,11,12,13,14]. In contrast, regularity measurements fail to be efficient for the characterization of non-regular textures, since measuring very low regularity parameters is not straightforward.
The approach proposed below to characterize non-regular textures involves the most relevant features that have proven useful in texture analysis [2,12,13]. These features are considered jointly in the framework of (1) stochasticity, a concept that relies on, but is not limited to, coarseness and roughness, and (2) wavelet packet transform, a representation that provides time-frequency, directionality, as well as other suitable statistical properties mentioned below.
The stochasticity degree (or randomness appearance) is hereafter measured by using the Kolmogorov stochasticity parameter [15]. This parameter is applied under the assumption that data are independent and identically distributed (iid), and their cumulative distribution function is completely specified. Recent works related to this parameter concern measuring the randomness degree of discrete sequences from dynamical systems and number theory [16], as well as measuring the contribution of randomness in the cosmic microwave background [17]. In these works, the Kolmogorov parameter has been used for specific datasets that are expected to comply with the underlying iid assumption, with a known distribution function.
In a more general framework involving real world textures, this iid assumption is very restrictive, due to non-stationarity, correlation and other more intricate statistical dependencies that occur among real world images.
The contributions of the present paper with respect to [16,17] concern relaxing these restrictive assumptions by:
(1)
Considering the wavelet packet transform, a transform that makes it possible to distribute many random processes as stationary, independent and identically distributed sequences (see, for instance, [18,19,20]);
(2)
Considering a dictionary of parametric models that are relevant with respect to the statistical distribution of the wavelet packet coefficients.
Under the above considerations, stochasticity can be measured even for correlated and non-stationary data through their wavelet packet representations. Furthermore, from the order structure that characterizes wavelet packet bases, we derive two different semantic templates for supporting a high level texture description: (1) generating a semantic stochasticity template upon a fixed wavelet packet basis; and (2) learning a wavelet packet tree structure that supports the best stochastic bases of training samples, provided that a critical stochasticity value is fixed.
The presentation of this paper is as follows. Section 2 introduces the Kolmogorov stochasticity parameter and assesses its relevance in detecting deterministic regular patterns. Section 3 addresses texture classification by using stochasticity templates. Section 4 presents the application of stochasticity analysis to standard content-based image retrieval by providing the stochastic structuring of databases. Section 5 provides semantic-based texture retrieval concepts and experimental results. Finally, Section 6 provides a conclusion to the paper.

2. Stochasticity Measurements

2.1. Kolmogorov Stochasticity Index: Deterministic Pattern

Let x = ( x 1 , x 2 , , x N ) be a sample set that follows from iid random variables with probability density function (pdf), f, and cumulative distribution function (cdf) F. The Kolmogorov stochasticity parameter [15] is:
κ x , F = sup t F x , N ( t ) - F ( t )
where F x , N is the empirical cdf of the N-sample sequence, x.
Standard approaches for testing random generators are based on binary hypothesis testing (stochastic or not) and focus on the asymptotic properties of N κ ( x , F ) when N tends to infinity. In contrast with these approaches, we assume no binary hypothesis, since we will use the heights of κ ( x , F ) to compare textures in terms of their randomness appearances, whatever the values of the stochasticity indices.
Note that κ ( x , F ) 1 for datasets that are stochastic with respect to F (a consequence of the Glivenko-Cantelli theorem). This implies that any x satisfying κ ( x , F ) = 1 is non-stochastic with respect to F. For instance, since we are dealing with a dictionary of continuous cdfs, we will say that a constant sequence is deterministic with respect to this dictionary: for such a sequence, the reader can check that sup F x , N ( t ) - F ( t ) = 1 , as far as N 2 . Furthermore, we have that the presence of a value with large occurrence in a dataset can be qualified as a deterministic pattern, since it impacts, as well, sup F x , N ( t ) - F ( t ) .
The following section addresses the relevance of κ ( x , F ) in pointing out deterministic patterns, in comparison with other stochasticity measures available in the literature.

2.2. The Relevance of the Kolmogorov Stochasticity Parameter in Detecting Deviations From a Specified Distribution

The results presented in this section concern the sensitivity of different stochasticity measures when data with a given stochasticity degree are corrupted with elementary deterministic patterns with increasing sizes.
There are basically two criteria that distinguish stochasticity measures:
(1)
The norm used, which can be cumulative or uniform;
(2)
The distribution, which can be specified as pdf or cdf.
Some examples issued from random generator testing are a) the Kolmogorov-Smirnov test [21], based on the uniform ( ) norm and comparing two cdfs in a binary hypothesis testing, b) the chi-squared test [22], based on the cumulative 2 norm and comparing two pdfs in a binary hypothesis testing problem.
Let us consider an N-size dataset (texture image for instance) with a stochasticity degree, η, with respect to a given distribution model. Assume that these data are affected by a deterministic pattern in the sense that a proportion, K / N , of the data is set to a constant value, where K is the size of the pattern under consideration. Since a stochasticity measure can be seen as a dissimilarity measure between distribution functions, a relevant stochasticity measure is such that its stochasticity parameter should increases as K increases (the randomness appearance of the texture has decreased).
In the following experiments, a deterministic pattern consisting of the insertion of K occurrences of a fixed value is introduced into datasets, and the relevance of different stochasticity measures is tested when the size K of this pattern increases. These experiments are performed upon the detail wavelet coefficients of textured images. These coefficients are expected to be very small in smooth regions and large in the neighborhood of edges. Increasing the number of null coefficients, if any, by forcing K large coefficients to zero (deterministic pattern) results in smoothing some edges of the image. This implies reducing the intrinsic stochasticity of the data when K increases. A relevant stochasticity measure should depart from the initial stochasticity degree when K increases.
Table 1. Experimental setup for testing the relevance of the uniform ( ) norm versus the cumulative 2 norm and Kullback-Leibler divergence (KLD) in stochasticity measurements. The quantities involved in the computation of the relative stochasticity value (RSV) are the empirical distribution and the model. cdf: cumulative distribution function.
Table 1. Experimental setup for testing the relevance of the uniform ( ) norm versus the cumulative 2 norm and Kullback-Leibler divergence (KLD) in stochasticity measurements. The quantities involved in the computation of the relative stochasticity value (RSV) are the empirical distribution and the model. cdf: cumulative distribution function.
For 0 K 150 , do:
Compute the wavelet coefficients, c j , n j , n , of the input image.
Introduce a deterministic pattern among the coefficients of a sub-band
by setting the K largest coefficients to zero (notation: c j , n K j , n ).
Compute the stochasticity parameters:
Check the distribution type from the variable “specification”
Casespecification is “cdf”, then:
Compute RSV ( K ) = F c j , n K , N - F θ c j , n K F c j , n 0 , N - F θ c j , n 0
Casespecification is “pdf”, then:
Compute RSV ( K ) = f c j , n K , N - f θ c j , n K f c j , n 0 , N - f θ c j , n 0
End
Compare the measurements obtained: for a relevant stochasticity measure,
RSV is a non-decreasing function of K.
We consider the experimental setup presented in Table 1: different combinations between norms ( 2 , ) and distribution specifications (cdf, pdf) are used for testing stochasticity measures. In this table, | | · | | specifically denotes either the and 2 norms. The Kullback-Leibler Divergence (KLD) is also used for comparison purposes. The Kullback-Leibler similarity measure between random variables, X 1 and X 2 , having probability distribution functions f X 1 and f X 2 is defined as:
K ( X 1 , X 2 ) = K ( X 1 | | X 2 ) + K ( X 2 | | X 1 )
with K ( X i | | X j ) = R f X i ( x ) log f X i ( x ) f X j ( x ) d x , i , j = 1 , 2 .
In addition, if c j , n denotes the wavelet packet coefficients obtained at sub-band W j , n , then c j , n K corresponds to the dataset obtained by setting the K largest values of c j , n to zero. In particular, c j , n 0 = c j , n .
Figure 1. Relative stochasticity values for the image “Fabric.0004” from the VisTeX database. The RSV of a relevant stochasticity measurement must be an increasing function of the size K of the deterministic pattern. We have K = 0 , 10 , 20 , , 150 and N = 68 , 644 . (a) cdf-based stochasticity measures; (b) pdf-based stochasticity measures.
Figure 1. Relative stochasticity values for the image “Fabric.0004” from the VisTeX database. The RSV of a relevant stochasticity measurement must be an increasing function of the size K of the deterministic pattern. We have K = 0 , 10 , 20 , , 150 and N = 68 , 644 . (a) cdf-based stochasticity measures; (b) pdf-based stochasticity measures.
Entropy 15 04782 g001
Departure from the initial stochasticity value of the wavelet coefficients is measured with respect to the generalized Gaussian distributions. In addition, Gaussian, triangle and Epanechnikov kernels have been used for the estimation of the empirical pdfs involved in Table 1. The results provided in Figure 1 and Figure 2 are obtained with a Gaussian kernel, and the wavelet decomposition has been performed with a Daubechies wavelet function of order seven. These results concern the images “Fabric.0004” and “Fabric.0018” from the VisTeX database (see Figure 1). Results are similar for other textures from the VisTeX database and for other kernels (concerning pdf-based measures).
Figure 2. Relative stochasticity values for the image “Fabric.0018” from the VisTeX database. The RSV of a relevant stochasticity measurement must be an increasing function of the size K of the deterministic pattern. We have K = 0 , 10 , 20 , , 150 and N = 68 , 644 . (a) cdf-based stochasticity measures; (b) pdf-based stochasticity measures.
Figure 2. Relative stochasticity values for the image “Fabric.0018” from the VisTeX database. The RSV of a relevant stochasticity measurement must be an increasing function of the size K of the deterministic pattern. We have K = 0 , 10 , 20 , , 150 and N = 68 , 644 . (a) cdf-based stochasticity measures; (b) pdf-based stochasticity measures.
Entropy 15 04782 g002
As can be seen in Figure 1 and Figure 2, the uniform norm on the cdfs (Kolmogorov strategy) is the sole strategy that guarantees a non-decreasing relative stochasticity value (RSV; see Table 1) when the size K of the pattern increases. Cumulative measures ( 2 , KLD), as well as pdf-based specifications, are not very relevant for stochasticity assessment, because of non-increasing deviations from the initial stochasticity degree: the local information is blurred through the averaging effect induced by cumulative measures or through neighborhood consideration when computing pdfs. Moreover, the same conclusion as above holds true when the experiments are performed on synthetic random numbers and without the use of wavelet transform.
From now on, we assume that the stochasticity parameter is of the Kolmogorov type: a uniform norm that applies for comparison of the empirical cdf with the distribution model. Section 3.1 addresses the choice of different bounds on this parameter for generating a semantic stochasticity template. This makes it possible to classify textures by mapping their sequences of stochasticity values on the stochasticity templates under consideration.

3. Classification From Wavelet Packet-Based Stochasticity Templates

3.1. Kolmogorov Stochasticity Measure versus Error-Bounds from Image Estimation

As shown in Section 2.2, the Kolmogorov parameter is relevant for detecting deterministic patterns in stochastic datasets and vice versa. For the main purpose of this paper, it is convenient to specify stochasticity bounds that make it possible to classify textures depending on their stochasticity degrees.
In the following, we derive different semantic classes consisting of stochasticity categories by fixing bounds on sup F x , N ( t ) - F ( t ) . This is performed by dealing with sup F x , N ( t ) - F ( t ) < η i as a problem of the estimation of an unknown function from observed samples and by fixing η i , so as to guarantee a peak signal-to-noise ratio (PSNR) greater than Ω i dBs , where ( Ω i ) i are bounds taken from standards on PSNR quality from image denoising and compression problems.
The PSNR (peak signal-to-noise ratio, in decibel units, dB) is given by:
PSNR = 10 log 10 d 2 / MSE
where d is the dynamic of the image (the range of the pixel values d = 255 for eight-bit coded images) and MSE denotes the mean squared error.
Proposition 1 
Consider the problem of fitting F x , N ( t i ) by F ( t i ) for i = 1 , 2 , , N . Then, in order to have a PSNR greater than Ω d B s , it suffices that
η 2 d × 10 - Ω / 10
The dynamic of F x , N is one. Now, we set Ω 0 = 30 dBs , Ω 1 = 35 dBs and Ω 2 = 40 dBs (a minimum of 30 dBs is required for an image denoising or compression algorithm to be relevant). These values are associated with the constants:
η i = 10 - Ω i / 10 , i = 0 , 1 , 2
hereafter referred to as the lower bounds for high, good and fair quality set indicators. We will then use four stochasticity classes below:
Definition 1 (Semantic stochasticity classes) 
A sample set x is said to be strongly stochastic (respectively stochastic, quasi-stochastic, non-stochastic) with respect to a continuous cdf, F, if κ x , F 0 , η 2 (respectively κ x , F η 2 , η 1 , κ x , F η 1 , η 0 , κ x , F η 0 , + ).

3.2. Texture Classification by Using Stochasticity Templates Upon Wavelet Packet Bases

Wavelet packet bases constitute a general framework for studying dictionaries of functional bases. Indeed, they offer a large family of functional bases with several properties, depending on whether we decide to split a given sub-band or not [23,24,25,26], among others. Best basis algorithms for the representation of signals involve seeking for functional atoms satisfying a given benchmark. In the wavelet framework, this benchmark is usually expressed in terms of the energy of the coefficients, the sparsity, or a number above a threshold, and entropy measurements [23,27,28]. Recent works on best basis algorithms concern compressive sensing and are related to the sparsity benchmark for piecewise regular images [29].
Hereafter, the best basis is computed upon the wavelet packet transform and under the stochasticity criterion: starting from the root node, W 0 , 0 , this consists of splitting every wavelet packet node W j , n recursively, unless the stochasticity of a node has reached the fixed stochasticity bound. Indeed, the statistical properties of the wavelet packet coefficients (in particular, the higher order cumulant decay, see [18,20], among others) ensure that the Kolmogorov parameter decrease for a large class of stationary and non-stationary random processes. The corresponding algorithm is hereafter called BSB-WP (Best Stochastic Basis upon Wavelet Packets).
Table 2. VisTeX “Fabric” texture classification from Best Stochastic Basis upon Wavelet Packets (BSB-WP) stochasticity measurements. The “Fabric” textures are given in Figure 3. In this table, “Det” designates wavelet packet details and “Approx” designates wavelet approximations.
Table 2. VisTeX “Fabric” texture classification from Best Stochastic Basis upon Wavelet Packets (BSB-WP) stochasticity measurements. The “Fabric” textures are given in Figure 3. In this table, “Det” designates wavelet packet details and “Approx” designates wavelet approximations.
Quasi-stochasticStochasticStrongly-stochastic
Det.Approx.Det.Approx.Det.Approx.
Fabric.18
Fabric.07
Fabric.17
Fabric.04
Fabric.09
Fabric.11
Fabric.15
Fabric.00
Fabric.14
We run the BSB-WP algorithm in order to classify the “Fabric” textures of the VisTeX database (see Figure 3). Stochasticity is measured with respect to cdfs pertaining to the exponential class and the four semantic classes given in Definition 1 (non-stochastic, quasi-stochastic, stochastic and strongly-stochastic semantic classes). We used a maximum depth, J * , fixed to four for the wavelet packet decomposition and the Daubechies wavelet of order seven. Table 2 summarizes the results obtained.
Figure 3. Texture “Fabrics” from the VisTeX database.
Figure 3. Texture “Fabrics” from the VisTeX database.
Entropy 15 04782 g003
One can note that these results are consistent with the visual perception of randomness appearance. Furthermore, we derive from these results that textures “Fabric.18” and “Fabric.07” can be well characterized by using probabilistic distribution modeling applied on the nodes involved in their best bases. In contrast, probabilistic distribution modeling is not relevant for describing non-stochastic textures, such as “Fabric.00” and “Fabric.14”, because their wavelet packet coefficient distribution deviates significantly from the continuous cdfs used. Note that the latter textures are regular, and thus, the appropriate criterion for their characterization needs to be based on regularity: regular images are sparse in the wavelet packet domain, and the sparsity criterion is thus expected to be more relevant.
The following provides some examples for illustrating BSB-WP texture characterization.
Example 1 
Texture “Fabric.07” is intrinsically quasi-stochastic: κ F a b r i c . 07 < η 0 . BSB-WP provides a basis (see the basis composed of framed sub-bands in Figure 4) where all sub-bands involved in the representation are stochastic: κ c 1 , n [ F a b r i c . 07 ] < η 1 for every n = 0 , 1 , 2 , 3 . The BSB-WP basis with a higher stochasticity property ( κ < η 2 ) has not been found up to decomposition level J * .
Example 2 
Texture “Fabric.09” is not intrinsically stochastic: κ F a b r i c . 09 > η 0 . BSB-WP provides a basis, where the texture can represented as a deterministic (smooth) approximation and stochastic details, κ c j , n [ F a b r i c . 09 ] < η 1 , for every nodes, ( j , n ) , involved in the tree of Figure 5, with n 0 .
Example 3 
Texture “Fabric.14” is not stochastic. κ-measurements are out of the stochasticity bounds for the input texture, as well as for many of its wavelet packet sub-bands up to decomposition level J * . In addition, this texture presents a “singular” path in the sense given in [18]. Indeed, depending on the input process, some paths are such that no cdf regularization can be expected. In this case, stochasticity measures can increase as the decomposition level increases. This occurs for the path with sub-bands marked in oval frames in Figure 6.
Table 2 and Figure 3 highlight that stochasticity measurements in the wavelet domain are sensitive to the roughness/coarseness/coherence of textures and reflect the “randomness-appearance” of textures.
Figure 4. 100 × κ for texture “Fabric.07” from the VisTeX database. The BSB-WP is composed of framed sub-bands.
Figure 4. 100 × κ for texture “Fabric.07” from the VisTeX database. The BSB-WP is composed of framed sub-bands.
Entropy 15 04782 g004
Figure 5. 100 × κ for texture “Fabric.09” from the VisTeX database. The BSB-WP is composed of framed sub-bands. The texture is represented as the sum of a smooth approximation and stochastic details.
Figure 5. 100 × κ for texture “Fabric.09” from the VisTeX database. The BSB-WP is composed of framed sub-bands. The texture is represented as the sum of a smooth approximation and stochastic details.
Entropy 15 04782 g005
Figure 6. 100 × κ for texture “Fabric.14” from the VisTeX database at decomposition level 2. Many sub-bands remain non-stochastic. In addition, the stochasticity parameter, κ, does not systematically decrease as the decomposition level increases in some detail paths (see the path with the oval boxes).
Figure 6. 100 × κ for texture “Fabric.14” from the VisTeX database at decomposition level 2. Many sub-bands remain non-stochastic. In addition, the stochasticity parameter, κ, does not systematically decrease as the decomposition level increases in some detail paths (see the path with the oval boxes).
Entropy 15 04782 g006
Table 3. Retrieval results per texture classes for 40 classes issued from the VisTeX database. Experimental results are performed without (blind approach) and with stochastic structuring, respectively. Stochastic texture classes are given in red, whereas non-stochastic classes are colored in blue, when considering stochastic structuring.
Table 3. Retrieval results per texture classes for 40 classes issued from the VisTeX database. Experimental results are performed without (blind approach) and with stochastic structuring, respectively. Stochastic texture classes are given in red, whereas non-stochastic classes are colored in blue, when considering stochastic structuring.
Blind approachStochastic structuring
TextureGGWBLPRTTextureGGWBLPRT
Bark.0069.5367.5876.95Bark.0069.9268.3678.13
Bark.0671.0970.7062.89Bark.0685.5585.9464.06
Bark.0868.7567.1956.64Bark.0869.5368.3656.64
Bark.0943.3642.9770.70Bark.0948.0547.2773.05
Bric.0198.8398.8375.39Bric.0198.8398.8376.56
Bric.0484.3882.8187.50Bric.0484.7783.2088.28
Bric.0591.8089.4581.64Bric.0592.9789.8483.98
Buil.0976.5697.6687.89Buil.0976.5697.6688.28
Fabr.0087.8986.7278.13Fabr.0094.9291.8078.13
Fabr.0486.7286.7281.64Fabr.0489.8487.8984.38
Fabr.0798.0598.0575.39Fabr.0798.0598.0575.39
Fabr.0910010080.47Fabr.0910010080.47
Fabr.1192.5892.5857.81Fabr.1192.5892.5857.81
Fabr.1410010089.45Fabr.1410010089.84
Fabr.1592.5892.1957.03Fabr.1592.9792.9757.03
Fabr.1792.5896.0985.16Fabr.1792.5896.0985.16
Fabr.1894.9291.8047.27Fabr.1894.9291.8047.27
Flow.0566.8065.6380.47Flow.0568.3666.4180.86
Food.0096.4896.4884.77Food.0010010087.50
Food.0578.9178.9172.27Food.0581.6481.6472.27
Food.0899.6110086.33Food.0899.6110086.33
Gras.0198.8398.8352.73Gras.0198.8398.8353.13
Leav.0874.2273.4480.86Leav.0882.0383.2080.86
Leav.1061.7260.9476.17Leav.1064.8463.6777.34
Leav.1167.5866.8087.11Leav.1173.0572.6687.50
Leav.1278.1378.5253.91Leav.1298.0597.2753.91
Leav.1672.2771.4887.11Leav.1672.2771.4887.11
Meta.0083.2082.4259.38Meta.0083.2082.4259.38
Meta.0210010086.33Meta.0210010086.33
Misc.0296.0995.7056.64Misc.0296.0995.7056.64
Sand.0096.4897.6651.56Sand.0096.4897.6651.56
Ston.0173.8374.6178.13Ston.0173.8374.6178.13
Ston.0493.7592.9749.22Ston.0493.7592.9749.22
Terr.1055.0853.5287.89Terr.1063.2862.5088.28
Tile.0162.1161.7291.41Tile.0162.1161.7291.41
Tile.0499.6199.6194.14Tile.0499.6199.6194.14
Tile.0798.0597.6683.98Tile.0799.2298.8383.98
Wate.0510010056.25Wate.0510010056.25
Wood.0157.4256.6475.39Wood.0161.3361.3375.78
Wood.0210010066.41Wood.0210010067.58

4. Content-Based Image Retrieval with Stochastic Structuring

In what follows, T denotes a texture database assumed to be heterogeneous in the sense that it contains both stochastic and regular textures. We consider the problem of structuring the elements of T by using the stochasticity degree. The structuring proposed is a splitting of database T into two metaclasses: stochastic versus regular textures. This structuring will be used as a pre-classification for level 1 content-based image retrieval (CBIR), based on parametric modeling of the statistical distributions of the wavelet coefficients. In this standard 1 CBIR [30], the query is completely specified through the statistical distributions of texture pixel values.

4.1. Stochastic Structuring

The structuring is performed with respect to the stochasticity measurements in the wavelet domain. The wavelet transform used is the stationary wavelet transform (SWT). Indeed, the SWT is appreciated for its shift-invariance property and is known to be relevant for the level 1 CBIR under consideration [31].
We consider the Edgeworth expansions of order four for modeling the SWT approximation sub-bands and the generalized Gaussian, Pareto and Weibull distributions for modeling the detail of the SWT coefficients. Model validation regarding the above issues can be found in [31]. The symmetric Kullback-Leibler divergence is used as a similarity measure between the statistical distributions given above.
Experimental tests concern 40 texture classes of the VisTeX database. The database structuring for these classes is given, in terms of stochastic versus regular textures, in Table 3: this structuring yields a stochastic metaclass composed of 22 texture classes and a regular metaclass composed of 18 texture classes.

4.2. Content-Based Image Retrieval on Structured Databases

This section provides CBIR experimental results on structured databases, in comparison with the results obtained without stochastic structuring. The experimental setup is the one used in [31]: any texture class (among the 40 texture classes considered) is composed of 16 images obtained by splitting every large texture image into 16 non-overlapping subimages. Thus, we have a test database, T , composed of 640 images, among which 352 images forming a database structure, T 1 , are issued from a stochastic class; whereas the 288 remaining textures constitute a database structure, T 2 , associated with regular texture classes, with T = T 1 T 2 .
We then run CBIR from parametric modeling and similarity measurements, as described in [31], with the symlet wavelet of order eight. Experimental tests are performed independently on the tree database structures, T 1 , T 2 a n d T . For a given structure, performance measurements concern the retrieval rates, when a query is any subimage of the structure under consideration. Retrieval rates per class are given in Table 3 concerning T 1 and T 2 . Average retrieval rates per structures T 1 , T 2 a n d T are given in Table 4 for comparison purposes.
Table 4. The average values of texture-specific retrieval for the whole database, T , the database composed of stochastic textures, T 1 , and the database composed of regular textures T 2 , with T 1 T 2 = T . Experimental results performed without stochastic structuring (blind approach) are given for comparison purposes.
Table 4. The average values of texture-specific retrieval for the whole database, T , the database composed of stochastic textures, T 1 , and the database composed of regular textures T 2 , with T 1 T 2 = T . Experimental results performed without stochastic structuring (blind approach) are given for comparison purposes.
Stochastic textures  ( T 1 )
Blind approachStochastic structuring
GGWBLPRTGGWBLPRT
88.1287.8266.0590.4590.0266.67
Regular textures  ( T 2 )
Blind approachStochastic structuring
GGWBLPRTGGWBLPRT
78.9579.6083.1881.1081.8183.51
Whole texture database  ( T )
Blind approachStochastic structuring
GGWBLPRTGGWBLPRT
83.9984.1273.7686.2486.3374.25
From Table 4, it follows that the retrieval is more concise when the search focuses either on T 1 or on T 2 than on the whole structure, T . Since T 1 and T 2 have low cardinality, the structuring also eases the search. In addition, from Table 4 and when comparing the role played by the distribution type on the metaclass, it follows that the more relevant family is:
  • the GG family for modeling the stochastic textures,
  • the PRT family for modeling the regular textures,
  • the WBL family for modeling the whole database containing both regular and stochastic textures.
The above remarks confirm the suitability of separating a heterogeneous database into structures with approximately the same stochasticity degrees.

5. Content-Based Stochasticity Retrieval

This section addresses stochasticity considerations for CBIR feature selection in texture databases. This CBIR takes into account the inference made in Section 3 for deriving different stochasticity templates. It is worth noting that the stochasticity degree can be seen as an index aggregating many low-level texture features (statistical distributions) in order to derive a high-level feature: the randomness-like appearance of a dataset. In this respect, we are concerned with the level 2 CBIR [30]. The motivation in using a stochasticity criterion for level 2 CBIR is the following.
Consider a geometrically regular image (a human face, textures “Fabric.00” and “Fabric.14” given in Figure 3, etc). For such an image, the form (through primitives) and the regularity are known to be relevant features for content description [10,32]. Consider now a stochastic texture (see, for instance, “Fabric.18” and “Fabric.07”). Such a texture is not regular and has no structured components that can be taken as feature descriptors. In contrast, the randomness appearance measured by the stochasticity parameter is appealing in differencing textures “Fabric.18” and “Fabric.07”: stochasticity is an index that addresses the intrinsic coherence (non-coherence being close to stochasticity) of the texture.
The following provides two CBIR strategies based on stochasticity measurements and referred to as content-based stochasticity retrieval (CBSR): (1) CBSR by learning the stochasticity tree structure characterizing the BSB-WP of some texture training samples; and (2) CBSR by generating the stochasticity template from a set of training texture samples, given a fixed wavelet packet basis.

5.1. Content-Based Stochasticity Retrieval by Learning the Stochasticity Tree Structure

In this section, we consider a set of M texture classes indexed by integer m , 1 m M . For a given class, m, we assume that samples ( m k ) k = 1 , 2 , K m are available (learning database). Let B Best [ m k ] denote the BSB-WP associated with sample m k at the fixed stochasticity degree, η.
Consider the smallest (infimum) and the largest (supremum) wavelet packet tree structures of the BSB-WP trees associated with samples ( m k ) k of texture class m. These trees define some wavelet packet bases denoted respectively by:
B inf [ m ] = inf B Best [ m 1 ] , B Best [ m 2 ] , , B Best [ m K m ]
B sup [ m ] = sup B Best [ m 1 ] , B Best [ m 2 ] , , B Best [ m K m ]
In this respect, we will say that the tree structure describing the behavior of the best stochastic representations of the samples of texture m have lower bound B inf [ m ] and upper bound B sup [ m ] .
The CBSR principle considered in this section is the following: an arbitrary sample belongs to stochasticity class m if its best basis at stochasticity degree η, denoted by B, is such that B inf [ m ] B B sup [ m ] .
Consider the set of “Fabric” textures from the VisTeX database (see Figure 3). From the classification obtained in Table 2, we focuses on “Fabric.0007” and “Fabric.0018”, which are closer on the basis of their stochasticity degrees. We set the stochasticity degree to η 2 . We then consider the following experimental setup: each image is split into 16 non-overlapping subimages; K = 8 images among them (the eight upper-half subimages) are used as the training set. The remaining 16 subimages, eight subimages of ‘Fabric.0007” and eight subimages of “Fabric.0018”, the lower-half subimages, are put together to form the test database.
We run the following CBSR strategy:
  • Learn the stochasticity tree structure for any of the “Fabric” textures by computing B inf and B sup from its eight, samples available from the learning database.
  • Retrieve, from the test database, the samples that belong to the semantic class of any of the “Fabric” textures, that are the samples having stochasticity bases bounded by the infimum and supremum bases associated with the class.
  • Sort the samples thus obtained, and compute texture-specific retrieval.
From the experiments carried out, we have that:
  • The learned basis structure corresponding to “Fabric.0007” is any basis, B, such that:
    n = 0 , 1 , 2 , 3 W 1 , n B W 2 , 0 n = 4 , 5 , , 4 3 - 1 W 3 , n
  • The learned basis structure for “Fabric.0018” is any basis B, such that:
    W 2 , 0 j = 1 , 2 n = 1 , 2 , 3 W j , n B W 3 , 0 j = 1 , 2 , 3 n = 1 , 2 , 3 W j , n
The retrieval rates obtained from the test database are such that:
  • Query “Fabric.0007”, associated with the lowest randomness degree among the two classes, reduces the search database from 16 to eight, including seven good retrievals/eight.
  • Query “Fabric.0018”, associated with the highest randomness degree among the two classes, reduces the search database from 16 to seven, including seven good retrievals/eight.
From these experiments, stochasticity, as a feature measuring the intrinsic coherence of a texture, can be used to generate a stochasticity tree structure representing the BSB-WPs, a texture observed through some training samples.

5.2. Content-Based Stochasticity Retrieval by Learning the Stochasticity Bounds

Depending on constraints, such as computational load or dealing with a large number of semantic classes, it may sometimes be desirable to fix the decomposition basis. In this section, we consider a fixed wavelet packet basis B = p = 1 , 2 , , L W J p , n p and propose high-level CBSR by computing, from training samples, the subspace within which the stochasticity parameters are expected to lie.
Assume the availability of K samples (subimages) for every texture class, m, considered, with 1 m M (the training set for this texture). Let us denote by κ m ( J p , n p ) the value of the stochasticity parameter (see Equation (1)) associated with the sub-band, W J p , n p , coefficients of subimage m . The sequence κ m ( J p , n p ) = 1 , 2 , , K represents the behavior of the stochasticity parameters of the projection of texture m samples on sub-band W J p , n p . Let us denote κ min m ( J p , n p ) = min κ m ( J p , n p ) , = 1 , 2 , , K and κ max m ( J p , n p ) = max κ m ( J p , n p ) , = 1 , 2 , , K . Define the stochasticity hypercube associated with the samples of texture m on basis B by:
H L m = p = 1 L κ m min ( J p , n p ) , κ m max ( J p , n p )
The CBSR principle considered in this section is the following: it is decided that a query sample admitting stochasticity parameters, κ ( J 1 , n 1 ) , κ ( J 2 , n 2 ) a n d , κ ( J L , n L ) , on basis B belongs to class m if the vector κ ( J 1 , n 1 ) , κ ( J 2 , n 2 ) , , κ ( J L , n L ) H L m .
In this respect, a texture can be characterized by the hypercube defined from the lower and upper bounds of the stochasticity parameters of its sample coefficients on the basis, B. This hypercube defines the semantic class of the texture.
Assume that a new sample of the texture is available. Then, we can re-evaluate the stochasticity bounds when some of the additional stochasticity parameters of this sample are out of the texture stochasticity hypercube. In addition, depending on the distribution of the stochasticity parameters, the user can discard those behaving as outliers in order to tighten the stochasticity hypercube and avoid overlapping with stochasticity hypercubes associated with other semantic classes. This re-evaluation is known to be useful in integrated CBIR systems [33].
The following provides CBSR experimental results obtained for M = 40 textures from the VisTeX database. The experimental setup used is described below:
  • First, we construct the learning database by using the top-half of the images: each top-half image is split into K = 8 non-overlapping subimages (128 × 128 pixels per subimage). These K subimages are used to compute the stochasticity hypercube H L m for m = 1 , 2 , , 40 .
  • Then, we constitute the test database by using the down-half of the images: each down-half image is split into eight non-overlapping subimages. Thus, the test database is composed of 8 × M subimages.
  • In order to increase the number of experiments, we have also permutated the roles played by the learning and the test database (the top-half becomes the down-half and vice-versa).
We run this procedure when the decomposition is performed by using a wavelet basis with J * = 2 . The stochasticity is measured with respect to dictionary D (Table 5a) and with respect to a single distribution family: the GG distributions (Table 5b). Specifically, in these tables, we have that two stochasticity coordinates out of H L m are tolerated, that is, two stochasticity parameters that are out-of-bounds are tolerated among a set of 3 * J * + 1 = 7 stochasticity parameters κ ( J 1 , n 1 ) , κ ( J 2 , n 2 ) , , κ ( J L , n L ) with L = 7 . In Table 5, TPR denotes the true positive rate defined as the ratio (the fraction of relevant queries per class):
TPR [ m ] = Number of admissible subimages that are issued from texture m Total number of relevant subimages
and FAR denotes the false alarm rate per class:
FAR [ m ] = Number of admissible subimages that are not issued from texture m Total number of subimages that are not issued from texture m
Table 5. (a) True positive rate (TPR) and false alarm rate (FAR) for content-based stochasticity retrieval (CBSR) by learning the stochasticity bounds. Dictionary D is used for stochasticity measurements; (b) True positive rate (TPR) and false alarm rate (FAR) for CBSR by learning the stochasticity bounds. The GG family is used for stochasticity measurements.
(a)
(a)
TextureTPRFAR
Bark.0062.5008.97
Bark.0637.5007.69
Bark.0862.5000.80
Bark.0975.0010.58
Bric.0131.2501.44
Bric.0437.5002.56
Bric.0550.0006.89
Buil.0937.5000.96
Fabr.0043.7501.92
Fabr.0450.0007.05
Fabr.0762.5001.12
Fabr.0937.5000.48
Fabr.1156.2501.44
Fabr.1450.000
Fabr.1568.7500.80
Fabr.1756.2501.12
Fabr.1868.7500.48
Flow.0531.2507.21
Food.0062.5001.92
Food.0531.2503.21
Food.0856.250
Grass.0137.5004.81
Leav.0868.7512.66
Leav.1043.7509.13
Leav.1156.2504.65
Leav.1243.7504.65
Leav.1662.5001.92
Meta.0050.0001.92
Meta.0268.7500.32
Misc.0262.5000.64
Sand.0031.2502.24
Ston.0156.2508.01
Ston.0462.5001.92
Terr.1050.0008.01
Tile.0131.2502.40
Tile.0437.5000.96
Tile.0725.000
Wate.0562.5002.40
Wood.0156.2512.18
Wood.0256.2509.29
(b)
(b)
TextureTPRFAR
Bark.0010018.27
Bark.0675.0013.78
Bark.0868.7503.37
Bark.0975.0020.83
Bric.0162.5003.85
Bric.0468.7504.01
Bric.0575.0008.97
Buil.0956.2502.56
Fabr.0062.5002.88
Fabr.0450.0013.62
Fabr.0781.2501.44
Fabr.0975.0000.64
Fabr.1181.2502.40
Fabr.1487.5000.16
Fabr.1587.5002.08
Fabr.1787.5003.05
Fabr.1881.2501.12
Flow.0556.2511.06
Food.0093.7502.56
Food.0550.0004.17
Food.0875.0000.48
Grass.0150.0007.85
Leav.0875.0016.03
Leav.1056.2514.58
Leav.1162.5006.89
Leav.1243.7504.81
Leav.1675.0002.56
Meta.0068.7502.72
Meta.0287.5000.64
Misc.0268.7500.96
Sand.0062.5003.37
Ston.0162.5013.94
Ston.0468.7503.53
Terr.1068.7516.35
Tile.0137.5004.17
Tile.0468.7502.08
Tile.0731.250
Wate.0581.2507.69
Wood.0193.7524.20
Wood.0275.0015.22
As can be seen in these tables, stochasticity-based retrieval is relevant for most textures given in this database. Low TPRs occur when texture is very regular (Example: “Tile.0001”, “Tile.0007”); see Figure 7. High FARs occur when textures have non-homogeneous subimages (Example: “Wood.0001”): the bounds define a large interval, which is expected to contain stochasticity values related to many other textures; see Figure 7.
Figure 7. Textures “Tile.0001”, “Tile.0007”, “Wood.0001” from the VisTeX album.
Figure 7. Textures “Tile.0001”, “Tile.0007”, “Wood.0001” from the VisTeX album.
Entropy 15 04782 g007
Experiments on the Brodatz album yield approximately the same results. The global TPR is 69 % for the Brodatz album (respectively 70 % for the VisTeX album) and the global FAR is 10 % for the Brodatz album (respectively 7 % for the VisTeX album), when GG modeling is used for stochasticity measurements. Tables concerning the Brodatz album are omitted, because the tests are concerned with 111 textures.

6. Conclusions

The paper has addressed texture image description and understanding through stochasticity or randomness appearance. The framework used for measuring stochasticity is that of the wavelet bases, because of their suitable statistical properties. The Kolmogorov stochasticity parameter is shown to be relevant for pointing out deterministic smooth patterns from wavelet coefficients of textures. The relevance of the stochasticity consideration is proven to be efficient for classification, database structuring and content-based image retrieval involving textured images.
Open issues related to this work may concern the analysis and the interpretation of the sequence of wavelet sub-band stochasticity parameters. In this work, we have considered the whole stochasticity hypercube obtained from the minimum and the maximum values of the stochasticity parameters of texture training samples. However, more investigations need to be performed in order to derive, among these sequences of parameters, some clusters or the manifold that describes the observed stochasticity sequences well.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Korostelev, A.P.; Tsybakov, A.B. Minimax Theory of Image Reconstruction; Springer: New York, NY, USA, 1993. [Google Scholar]
  2. Tamura, H.; Mori, S.; Yamawaki, T. Textural features corresponding to visual perception. IEEE Trans. Syst. Man Cybern. 1978, 8, 460–473. [Google Scholar] [CrossRef]
  3. Vasselle, B.; Giraudon, G. A multiscale regularity measure as a geometric criterion for image segmentation. Mach. Vision Appl. 1994, 7, 229–236. [Google Scholar] [CrossRef]
  4. Manjunath, B.S.; Ohm, J.; Vasudevan, V.V.; Yamada, A. Color and texture descriptors. IEEE Trans. Circuits Syst. Video Technol. 1998, 11, 703–715. [Google Scholar] [CrossRef]
  5. Chetverikov, D. Pattern regularity as a visual key. Image Vision Comput. 2000, 18, 975–985. [Google Scholar] [CrossRef]
  6. Fujii, K.; Sugi, S.; Ando, Y. Textural properties corresponding to visual perception based on the correlation mechanism in the visual system. Psychol. Res. 2003, 67, 197–208. [Google Scholar] [CrossRef] [PubMed]
  7. Trujillo, L.; Olague, G.; Legrand, P.; Lutton, E. Regularity based descriptor computed from local image oscillations. Opt. Express 2007, 15, 6140–6145. [Google Scholar] [CrossRef] [PubMed]
  8. Daugmann, J.G. Uncertainty relation for resolution in space, spatial frequency, and orientation optimized by two-dimensional visual cortical filters. J. Opt. Soc. Am. A 1985, 2, 1160–1169. [Google Scholar] [CrossRef]
  9. Liu, F.; Picard, R.W. Periodicity, directionality, and randomness: Wold features for image modeling and retrieval. IEEE Trans. Pattern Anal. Mach. Intell. 1996, 18, 722–733. [Google Scholar] [CrossRef]
  10. Wang, J.Z.; Li, J.; Wiederhold, G. SIMPLIcity: Semantics-sensitive integrated matching for picture libraries. IEEE Trans. Pattern Anal. Mach. Intell. 2001, 23, 947–963. [Google Scholar] [CrossRef]
  11. Stitou, Y.; Turcu, F.; Berthoumieu, Y.; Najim, M. Three-dimensional textured image blocks model based on wold decomposition. IEEE Trans. Signal Process. 2007, 55, 3247–3261. [Google Scholar] [CrossRef]
  12. Unser, M. Texture classification and segmentation using wavelet frames. IEEE Trans. Image Process. 1995, 4, 1549–1560. [Google Scholar] [CrossRef] [PubMed]
  13. Mondal, D.; Percival, D. Wavelet variance analysis for random fields on a regular lattice. IEEE Trans. Image Process. 2012, 21, 537–549. [Google Scholar] [CrossRef] [PubMed]
  14. Geilhufe, M.; Percival, D.; Stern, H. Two-dimensional wavelet variance estimation with application to sea ice SAR images. Comput. Geosci. 2013, 54, 351–360. [Google Scholar] [CrossRef]
  15. Kolmogorov, A.N. Sulla determinazione empirica di una legge di distribuzione. Giornale dell’Istituto Italiano degli Attuari 1933, 4, 83–91. (in Italian). [Google Scholar]
  16. Arnold, V. Orbits’ statistics in chaotic dynamical systems. Nonlinearity 2008, 21, T109–T112. [Google Scholar] [CrossRef]
  17. Gurzadyan, V.G.; Kocharyan, A.A. Kolmogorov stochasticity parameter measuring the randomness in the cosmic microwave background. Astron. Astrophys. 2008, 492, L33–L34. [Google Scholar] [CrossRef]
  18. Atto, A.M.; Berthoumieu, Y. Wavelet packets of nonstationary random processes: Contributing factors for stationarity and decorrelation. IEEE Trans. Inf. Theory 2012, 58, 317–330. [Google Scholar] [CrossRef]
  19. Atto, A.M.; Pastor, D.; Mercier, G. Wavelet packets of fractional brownian motion: Asymptotic analysis and spectrum estimation. IEEE Trans. Inf. Theory 2010, 56, 4741–4753. [Google Scholar] [CrossRef] [Green Version]
  20. Atto, A.M.; Pastor, D. Central limit theorems for wavelet packet decompositions of stationary random processes. IEEE Trans. Signal Process. 2010, 58, 896–901. [Google Scholar] [CrossRef] [Green Version]
  21. Massey, J.F.J. The Kolmogorov-Smirnov test for goodness of fit. J. Am. Stat. Assoc. 1951, 253, 68–78. [Google Scholar] [CrossRef]
  22. Kenney, J.F.; Keeping, E.S. Mathematics of Statistics, Part II; Van Nostrand Company, Inc: Princeton, NJ, USA, 1951. [Google Scholar]
  23. Wickerhauser, M.V. Inria Lectures on Wavelet Packet Algorithms. In Problèmes Non-Linéaires Appliqués, Ondelettes et Paquets D’Ondes; Lions, P.-L., Ed.; INRIA: Roquencourt, France, 1991; pp. 31–99. [Google Scholar]
  24. Daubechies, I. Ten Lectures on Wavelets; SIAM: Philadelphie, PA, USA, 1992. [Google Scholar]
  25. Coifman, R.R.; Wickerhauser, M.V. Entropy-based algorithms for best basis selection. IEEE Trans. Inf. Theory 1992, 38, 713–718. [Google Scholar] [CrossRef]
  26. Hess-Nielsen, N.; Wickerhauser, M.V. Wavelets and time-frequency analysis. Proc. IEEE 1996, 84, 523–540. [Google Scholar] [CrossRef]
  27. Laine, A.; Fan, J. Texture classification by wavelet packet signatures. IEEE Trans. Pattern Anal. Mach. Intell. 1993, 15, 1186–1191. [Google Scholar] [CrossRef]
  28. Chang, T.; Kuo, C.-C.J. Texture analysis and classification with tree-structured wavelet transform. IEEE Trans. Image Process. 1993, 2, 429–441. [Google Scholar] [CrossRef] [PubMed]
  29. Peyré, G. Best basis compressed sensing. IEEE Trans. Signal Process. 2010, 58, 2613–2622. [Google Scholar] [CrossRef]
  30. Eakins, J.; Graham, M. Content-Based Image Retrieval; Technical Report; University of Northumbria: Newcastle, UK, 1999. [Google Scholar]
  31. Atto, A.M.; Berthoumieu, Y. How to Perform Texture Recognition from Stochastic Modeling in the Wavelet Domain. In Proceedings of IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP), Prague, Czech Republic, 22–27 May 2011; pp. 4320–4323.
  32. Liu, Y.; Zhang, D.; Lu, G.; Ma, W.-Y. A survey of content-based image retrieval with high-level semantics. Pattern Recognit. 2007, 40, 262–282. [Google Scholar] [CrossRef]
  33. Smeulders, A.W.M.; Worring, M.; Gupta, A.; Jain, R. Content-based image retrieval at the end of the early years. IEEE Trans. Pattern Anal. Mach. Intell. 2000, 22, 1349–1380. [Google Scholar] [CrossRef]

Share and Cite

MDPI and ACS Style

Atto, A.M.; Berthoumieu, Y.; Mégret, R. Stochasticity: A Feature for the Structuring of Large and Heterogeneous Image Databases. Entropy 2013, 15, 4782-4801. https://doi.org/10.3390/e15114782

AMA Style

Atto AM, Berthoumieu Y, Mégret R. Stochasticity: A Feature for the Structuring of Large and Heterogeneous Image Databases. Entropy. 2013; 15(11):4782-4801. https://doi.org/10.3390/e15114782

Chicago/Turabian Style

Atto, Abdourrahmane M., Yannick Berthoumieu, and Rémi Mégret. 2013. "Stochasticity: A Feature for the Structuring of Large and Heterogeneous Image Databases" Entropy 15, no. 11: 4782-4801. https://doi.org/10.3390/e15114782

Article Metrics

Back to TopTop