Next Article in Journal
A Feature Subset Selection Method Based On High-Dimensional Mutual Information
Next Article in Special Issue
On the Thermodynamics of Classical Micro-Canonical Systems
Previous Article in Journal
Some Convex Functions Based Measures of Independence and Their Application to Strange Attractor Reconstruction
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Optimal Multi-Level Thresholding Based on Maximum Tsallis Entropy via an Artificial Bee Colony Approach

School of Information Science and Engineering, Southeast University, Nanjing 210096, China
*
Author to whom correspondence should be addressed.
Entropy 2011, 13(4), 841-859; https://doi.org/10.3390/e13040841
Submission received: 2 March 2011 / Revised: 17 March 2011 / Accepted: 29 March 2011 / Published: 13 April 2011
(This article belongs to the Special Issue Tsallis Entropy)

Abstract

:
This paper proposes a global multi-level thresholding method for image segmentation. As a criterion for this, the traditional method uses the Shannon entropy, originated from information theory, considering the gray level image histogram as a probability distribution, while we applied the Tsallis entropy as a general information theory entropy formalism. For the algorithm, we used the artificial bee colony approach since execution of an exhaustive algorithm would be too time-consuming. The experiments demonstrate that: 1) the Tsallis entropy is superior to traditional maximum entropy thresholding, maximum between class variance thresholding, and minimum cross entropy thresholding; 2) the artificial bee colony is more rapid than either genetic algorithm or particle swarm optimization. Therefore, our approach is effective and rapid.

1. Introduction

Thresholding is a popular method for image segmentation. From a grayscale image, bi-level thresholding can be used to create binary images, while multilevel thresholding determines multiple thresholds which divide the pixels into multiple groups. Thresholding is widely used in many image processing applications such as optical character recognition [1], infrared gait recognition [2], automatic target recognition [3], detection of video changes [4], and medical image applications [5].
There are many image thresholding studies in the literature. In general, thresholding methods can be classified into parametric and nonparametric methods. For parametric approaches, the gray-level distribution of each group is assumed to obey a Gaussian distribution, and then the approaches attempt to find an estimate of the parameters of Gaussian distribution that best fits the histogram. Wang et al. [6] integrated the histogram with the Parzen window technique to estimate the spatial probability distribution. Fan et al. [7] approximated the histogram with a mixed Gaussian model, and estimated the parameters with an hybrid algorithm based on particle swarm optimization and expectation maximization. Zahara et al. [8] fitted the Gaussian curve by Nelder-Mead simplex search and particle swarm optimization. To resolve the histogram Gaussian fitting problem, Nakib et al. used an improved variant of simulated annealing adapted to continuous problems [9].
Nonparametric approaches find the thresholds that separate the gray-level regions of an image in an optimal manner based on some discriminating criteria. Otsu’s criterion [10], which selects optimal thresholds by maximizing the between class variance, is the most popular method. However, inefficient formulation of between class variance makes the method quite time-consuming in multilevel threshold selection. To overcome this problem, Chung et al. [11] presented an efficient heap and quantization based data structure to realize a fast implementation. Huang et al. [12] proposed a two-stage multi-threshold Otsu method. Wang et al. [13] proposed applying an improved shuffled frog-leaping algorithm to the three dimensional Otsu thresholding. Besides the criterion of the between class variance, other criteria are also investigated. Horng et al. [14] applied the maximum entropy thresholding (MET) criterion and use the honey bee mating optimization method for fast thresholding. Yin et al. [15] applied the minimum cross entropy thresholding (MCET), and presented a recursive programming technique to reduce by an order of magnitude the computation of the MCET objective function. Hamza [16] proposed a nonextensive information-theoretic measure called Jensen-Tsallis divergence for image edge detection.
Recent developments in statistical mechanics based on Tsallis entropy have intensified the interest of investigating it as an extension of Shannon’s entropy [17]. It appears in order to generalize the Boltzmann/Gibbs’ traditional entropy to nonextensive physical systems. In the new theory a new parameter q measuring the degree of nonextensivity is introduced, and it is system dependent. In this paper, we applied Tsallis entropy to take replace of traditional MET.
Meanwhile, the computation time will increase sharply when the number of thresholds increases, so the traditional exhaustive method does not work. For a 4-level thresholding of a Lena image of size 512-by-512, the computation time will be over three days [14]. Therefore, scholars tend to use intelligence optimization methods to find the best threshold quickly. Nakib et al. [18] proposed a new image thresholding model based on multi-objective optimization, and then used an improved genetic algorithm to solve the model. Maitra et al. [19] proposed an improved variant of particle swarm optimization (PSO) for the task of image multilevel thresholding. Both GA and PSO can jump from local minima, but their efficiencies are not satisfactory.
The Artificial Bee Colony (ABC) algorithm was originally presented by Dervis Karaboga [20] under the inspiration of the collective behavior of honey bees. It shows better performance in function optimization problems compared with GA, differential evolution (DE), and particle swarm optimization (PSO) [21]. As is known, normal global optimization techniques conduct only one search operation in one iteration, for example the PSO carries out a global search in the beginning stage and a local search in the ending stage. The ABC approach features the advantage that it conducts both global search and local search in each iteration, and as a result the probability of finding the optimal result is significantly increased, which effectively avoids local optima to a large extent [22]. In this paper, we chose ABC to solve the multi-level thresholding model.
The rest of this paper is organized as follows: Section 2 presents the fundamental concepts of nonextensive systems, and compares conventional Shannon entropy with Tsallis entropy. Section 3 presents the detailed multi-level thresholding model, and gives the formula of the objective function. Section 4 introduces the ABC algorithm with explanations of the essential components, the principles, and the procedures. Experiments in Section 5 first compare the proposed maximum Tsallis entropy criterion with other traditional criteria (maximum entropy, maximum between class, and minimum cross entropy thresholding), followed by a comparison between the ABC algorithm, GA and PSO. Finally the influence of nonextensive parameter q of Tsallis entropy is discussed. Section 6 gives the concluding remarks of this paper.

2. Nonextensive Entropy

The entropy is basically a thermodynamic concept associated with the order of irreversible processes from a traditional point of view. Shannon redefined the entropy concept of Boltzmann/Gibbs as a measure of uncertainty regarding the information content of a system [23]. The Shannon entropy is defined from the probability distribution, where pi denotes the probability of each state i. Therefore, the Shannon entropy can be described as:
Entropy 13 00841 i001
where L denotes the total number of states. This formalism is restricted to the domain of validity of the Boltzmann-Gibbs-Shannon (BGS) statistics, which only describes Nature when the effective microscopic interactions and the microscopic memory are short ranged [24]. Suppose a physical system can be decomposed into two statistical independent subsystems A and B, the Shannon entropy has the extensive property (additivity):
Entropy 13 00841 i002
However, for a certain class of physical systems which entail long-range interactions, long time memory, and fractal-type structures, it is necessary to use nonextensive entropy. Tsallis has proposed a generalization of BGS statistics, and its form can be depicted as:
Entropy 13 00841 i003
where the real number q denotes an entropic index that characterizes the degree of nonextensivity. Above expression will meet the Shannon entropy in the limit q → 1. The Tsallis entropy is nonextensive in such a way that for a statistical dependent system. Its entropy is defined with the obey of pseudo additivity rule:
Entropy 13 00841 i004
Three different entropies can be defined with regard to different values of q. For q < 1, the Tsallis entropy becomes a subextensive entropy where Sq(A + B) < Sq(A) + Sq(B); for q = 1, the Tsallis entropy reduces to an standard extensive entropy where Sq(A + B) = Sq(A) + Sq(B); for q > 1, the Tsallis entropy becomes a superextensive entropy where Sq(A + B) > Sq(A) + Sq(B).
The Tsallis entropy can be extended to the fields of image processing, because of the presence of the correlation between pixels of the same object in a given image. The correlations can be regarded as the long-range correlations that present pixels strongly correlated in luminance levels and space fulfilling.

3. Thresholding Model

Assume that an image can be represented by L gray levels. The probabilities of pixels at level i is denoted by pi; so pi ≥ 0 and p1 + p2 + … pL = 1. If the image is divided into two classes, CA and CB by a threshold at level t, where class CA consists of gray levels from 1 to t and CB contains the rest gray levels from t + 1 to L, the cumulative probabilities can be defined as:
Entropy 13 00841 i005
Therefore, the normalization of probabilities pA and pB can be defined as:
Entropy 13 00841 i006
Entropy 13 00841 i007
Now, the Tsallis entropy for each individual class is defined as:
Entropy 13 00841 i008
Entropy 13 00841 i009
The Tsallis entropy Sq(t) of each individual class is dependent on the threshold t. The total Tsallis entropy of the image is written as follows [25]:
Entropy 13 00841 i010
The task is to maximize the total Tsallis entropy between class CA and CB. When the value of Sq(t) is maximized, the corresponding gray-level t* is regarded as the optimum threshold value:
Entropy 13 00841 i011
For bi-level thresholding, the best threshold can be obtained by a cheap computation effort, however, for multi-level thresholding (suppose the number of class is m), the problems will encounter following two challenges: 1) The optimization functions become more complicated. Considering a tri-level thresholding, the optimization function can be depicted as:
Entropy 13 00841 i012
This causes little trouble since computers can calculate the function rapidly. 2) The parametric space is enlarged exponentially. The number of Sq(t) calculated is equal to:
Entropy 13 00841 i013
We choose m varying from 2 to 7 as examples and show the results in Table 1. It indicates the number of function calculated also increase from 255 to 3.6 × 1011. Therefore, the exhaustive algorithm failed for multi-level thresholding, we will introduce the ABC method for fast computation in next section.
Table 1. Computation burden increases with the number of classes.
Table 1. Computation burden increases with the number of classes.
No of Classes m234567
No of Sq(t) calculated25532,3852.73×1061.72×1088.64×1093.60×1011

4. Artificial Bee Colony

4.1. Essential Components

In nature, each bee only performs a single task, whereas, through a variety of information communication ways between bees such as waggle dances and special odor, the entire bee colony can easily find food resources that produce relative high amounts of nectar, hence realize its self-organizing behavior [21]. In order to introduce the self-organization model of forage selection that lead to the emergence of collective intelligence of honey bee colony, we should define three essential components [26]:
  • Food Resource: The value of a food source depends on different parameters, such as its proximity to the nest, richness of concentration of energy, and the ease of extracting this energy. For the simplicity, the profitability of a food source can be represented with a single quantity.
  • Unemployed Foragers: If it is assumed that a bee has no knowledge about the food sources in the search field, it initializes its search as an unemployed forager looking for a food source to exploit. There are two types of unemployed forages: (I) Scouts: If the bee starts searching spontaneously without any knowledge, it will be a scout bee. The percentage of scout bees varies from 5% to 30% according to the information into the nest. The mean number of scouts averaged over conditions is about 10% in Nature [27]. (II) Onlookers: The onlookers wait in the nest and find a food source through the information shared by employed foragers. If the onlookers attend a waggle dance done by some other bee, the bees become recruits and start searching by using the knowledge from the waggle dance.
  • Employed Foragers: Employed foragers are associated with a particular food source, which they are currently exploiting or are “employed” at. They carry with them information about this particular source, its distance and direction from the nest, and the profitability of the source. They share this information with a certain probability. After the employed foraging bee loads a portion of nectar from the food source, it returns to the hive and unloads the nectar to the food area in the hive. There are three possible options related to the residual amount of nectar for the foraging bee: (I) Unemployed Forager: If the nectar amount decreases to a low level or becomes exhausted, the foraging bee abandons the food source and become an unemployed bee. (II) Employed Forager 1: Otherwise, the foraging bee might dance and then recruit nest mates before returning to the same food source. (III) Employed Forager 2: Or it might continue to forage at the food source without recruiting other bees.

4.2. Principles and Procedures

At the initial stage, all the bees without any prior knowledge play the role of unemployed bees. There are two options for such a bee, viz. it can be either a scout or an onlooker. After finding the food source, the bee utilizes its own capability to memorize the information and immediately starts exploiting it. Hence, the unemployed bee becomes an employed forager. The foraging bee takes a load of nectar from the source and returns to the hive, unloading the nectar to a food store. When the profit of the food source is lower than the threshold, it abandons the food source and again becomes an unemployed bee. When the profit of the food source is higher than the threshold, it can continue its foraging, either recruiting more bees or by itself.
In this way, an artificial bee colony is introduced by mimicking the behavior of natural bees to construct a relative good solution of realistic optimization problems. The colony of artificial bees contains three groups of bees: employed bees, onlookers and scouts. The first half of the colony consists of the employed artificial bees and the second half includes the onlookers [21]. For every food source, there is only one employed bee, viz., the number of employed bees is equal to the number of food sources. The employed bee of an abandoned food source becomes a scout. The search carried out by the artificial bees can be summarized as follows [21]: (I) Employed bees determine a food source within the neighborhood of the food source in their memory. (II) Employed bees share their information with onlookers within the hive and then the onlookers select one of the food sources. (III) Onlookers select a food source within the neighborhood of the food sources chosen by themselves. (IV) An employed bee of which the source has been abandoned becomes a scout and starts to search a new food source randomly.
The detailed main steps of the algorithm are given below:
  • Step 1. Initialize the population of solutions xij (here i denotes the ith solution, and j denotes the jth epoch, i = 1, …, SN, here SN denotes the number of solutions) with j = 0:
    Entropy 13 00841 i014
    here LB & UB represents the lower and upper bounds, which can be infinity if not specified. Then, evaluate the population via the specified optimization function;
  • Step 2. Repeat, and let j = j + 1;
  • Step 3. Produce new solutions (food source positions) υij in the neighborhood of xij for the employed bees using the formula:
    Entropy 13 00841 i015
    here xkj is a randomly chosen solution in the neighborhood of xij to produce a mutant of solution xij, Φ is a random number in the range [−1, 1]. Evaluate the new solutions;
  • Step 4. Apply the greedy selection process between the corresponding xij and υij;
  • Step 5. Calculate the probability values Pij for the solutions xij by means of their fitness values using the equation:
    Entropy 13 00841 i016
    here f denotes the fitness value;
  • Step 6. Normalize Pij values into [0, 1];
  • Step 7. Produce the new solutions (new positions) υij for the onlookers from the solutions xij using the same equation as in Step 3, selected depending on Pij, and evaluate them;
  • Step 8. Apply the greedy selection process for the onlookers between xij and υij;
  • Step 9. Determine the abandoned solution (source), if it exists, and replace it with a new randomly produced solution xij for the scout using the equation:
    Entropy 13 00841 i017
    Here φij is a random number in [0, 1].
  • Step 10. Memorize the best food source position (solution) achieved so far;
  • Step 11. Go to Step 2 until termination criteria is met.

5. Experiments

The experiments were carried out on a P4 IBM platform with a 3 GHz processor and 2 GB memory, running under the Windows XP operating system. The algorithm was developed via the signal processing toolbox, image processing toolbox, and global optimization toolbox of Matlab 2010b.

5.1. Effectiveness of Tsallis Entropy

We compared the maximum Tsallis entropy thresholding (MTT) to other criteria including maximum between class variance thresholding (MBCVT), maximum entropy thresholding (MET), and minimum cross entropy thresholding (MCET). Taking m = 2 as the example, suppose PA and PB denote the probability of class CA and CB, respectively, ωA and ωB denote the corresponding the cumulative probabilities, μA and μB denote the corresponding mean value, the formalism of objective function of different criteria are shown in Table 2. Here, we add a minus symbol to the objective function of MCET to transform the minimization to maximization.
Table 2. Objective function of different criteria (m = 2).
Table 2. Objective function of different criteria (m = 2).
CriteriaObjectiveType
METS(A) + S(B)Maximize
MBCVTωA × ωB × (μAμB)2Maximize
MCETμA × ωA × log(μA) − μB × ωB × log(μB)Minimize
μA × ωA × log(μA) + μB × ωB × log(μB)Maximize
MTTSq(A) + Sq(B) + (1 − q) × Sq(A) × Sq(B)Maximize
We use some synthetic images to demonstrate the robustness of the MTT criteria. In the synthetic image, a square (gray value 192) is put on a background (gray value 64), and then Gaussian noises (variances = 0.01) was added to the whole image, so the histogram of the image is two Gaussian peaks. The size of the image is 256-by-256 size, and the size of the foreground square is 180-by-180, so that the heights of the two Gaussian peaks of foreground and background are nearly the same.
In addition, we reduce the length of the foreground square to change the height ratio of the two Gaussian peaks, and lower the gray level of the background to enlarge the distance of the two Gaussian peaks. q is determined as 4 for all cases of synthetic images.
The results are chosen and shown in Figure 1, Figure 2 and Figure 3. In Figure 1, the two Gaussian peaks are nearly the same. The MET, MBCVT, and MTT can find the optimal threshold, while MCET locates the threshold a bit left. In Figure 2, we shrink the size of the foreground object, so the peak of the background is much higher than the peak of the foreground object. The MET and the MTT perform well with threshold as 128 while MBCVT and MCET obtained the threshold nearly at the position of the higher Gaussian peak.
Figure 1. Segmentation of the synthetic image.
Figure 1. Segmentation of the synthetic image.
Entropy 13 00841 g001
Figure 2. Segmentation of the synthetic image (foreground is smaller).
Figure 2. Segmentation of the synthetic image (foreground is smaller).
Entropy 13 00841 g002
Figure 3. Segmentation of the synthetic image (background is darker).
Figure 3. Segmentation of the synthetic image (background is darker).
Entropy 13 00841 g003
In Figure 3, we lower the gray level of the background to only 26 so that the distance between two peaks are further. In this case we observe that MBCVT and MTT obtain the right threshold just between the two peaks while MET finds a higher threshold and MCET finds a lower threshold. The results are listed in Table 3 which indicates that the MTT succeeds in all cases.
Table 3. Result of segmentation of synthetic image (√ denotes success).
Table 3. Result of segmentation of synthetic image (√ denotes success).
ImageMETMBCVTMCETMTT
Two equal Gaussian peaks
One peak is higher
Larger distance between two peaks
Afterwards, we compared the criteria with several real images, three of which are selected to demonstrate the effectiveness of MTT criterion. Figure 4 is an image of a flower with an inhomogeneous distribution of light around it. The Tsallis entropy criterion is very useful in such applications where we define a value for the parameter q to adjust the thresholding level to the correct point. In Figure 4 the image was segmented with q equal to 4. The thresholds of MET, MBCVT, MCET, and MTT are 126, 78, 42, and 14, respectively. It indicates that MTT obtains a cleaner segmentation without loss of any part of the flower.
Figure 4. Segmentation of the rose image.
Figure 4. Segmentation of the rose image.
Entropy 13 00841 g004
Figure 5 is an image of a tree in the right and leaves in the top-left corner. The q was set to 5. The thresholds of MET, MBCVT, MCET, and MTT are 114, 52, 35, and 63, respectively. Figure 5 shows that MTT extracts leaves in the top-left corner more completely than those by MET, MBCVT, and MCET. Therefore, the MTT outperforms again.
Figure 5. Segmentation of the tree image.
Figure 5. Segmentation of the tree image.
Entropy 13 00841 g005
Figure 6 shows an image of Pluto. The q was set to 4. The thresholds of MET, MBCVT, MCET, and MTT are 111, 183, 170, and 238, respectively. It indicates that the MTT method obtains a fuller Pluto segmentation.
Figure 6. Segmentation of the Pluto image.
Figure 6. Segmentation of the Pluto image.
Entropy 13 00841 g006

5.2. Rapidness of the ABC

The aforementioned simulation experiments have proven the superiority of the MTT criterion. In this section, we generalize the bi-level thresholding to multi-level thresholding which needs the proposed ABC algorithm for fast computation. A synthetic image (A central square and five surrounding borders) and three real images of 256-by-256 pixels with corresponding histograms were used and shown in Figure 7. In order to evaluate and analyze the performance of ABC, we use the GA and PSO as the comparative algorithms. Each algorithm was run 30 times to eliminate the randomness.
Figure 7. Four test images and the histograms: (a) synthetic (b) Lena; (c) cameraman; (d) peppers.
Figure 7. Four test images and the histograms: (a) synthetic (b) Lena; (c) cameraman; (d) peppers.
Entropy 13 00841 g007aEntropy 13 00841 g007b
A typical run of the resulting segmentation for the four test images based on MTT are shown in Table 4, Table 5, Table 6 and Table 7, respectively, for classes m = 3 to 6 with segmented thresholds listed below. The selected thresholds by the ABC are nearly equivalent to the ones obtained by GA and PSO, which reveals that the segmentation results depend heavily on the criterion function selected.
Table 4. Multi-level segmentation for synthetic image.
Table 4. Multi-level segmentation for synthetic image.
mGAPSOABC
3 Entropy 13 00841 i018 Entropy 13 00841 i019 Entropy 13 00841 i020
89 172 85 171 85 170
4 Entropy 13 00841 i021 Entropy 13 00841 i022 Entropy 13 00841 i023
63 127 188 64 126 185 62 128 189
5 Entropy 13 00841 i024 Entropy 13 00841 i025 Entropy 13 00841 i026
52 99 157 205 49 100 155 205 50 101 155 204
6 Entropy 13 00841 i027 Entropy 13 00841 i028 Entropy 13 00841 i029
40 85 128 169 213 42 86 127 172 215 42 85 128 170 214
Table 5. Multi-level segmentation for Lena image.
Table 5. Multi-level segmentation for Lena image.
mGAPSOABC
3 Entropy 13 00841 i030 Entropy 13 00841 i031 Entropy 13 00841 i032
68 180 66 188 70 185
4 Entropy 13 00841 i033 Entropy 13 00841 i034 Entropy 13 00841 i035
73 112 192 67 113 187 71 113 188
5 Entropy 13 00841 i036 Entropy 13 00841 i037 Entropy 13 00841 i038
70 110 136 192 68 109 139 194 68 108 138 192
6 Entropy 13 00841 i039 Entropy 13 00841 i040 Entropy 13 00841 i041
68 111 136 168 187 69 112 134 169 188 69 110 135 169 189
Table 6. Multi-level segmentation for cameraman image.
Table 6. Multi-level segmentation for cameraman image.
mGAPSOABC
3 Entropy 13 00841 i042 Entropy 13 00841 i043 Entropy 13 00841 i044
102 199 95 199 99 195
4 Entropy 13 00841 i045 Entropy 13 00841 i046 Entropy 13 00841 i047
29 85 190 28 92 196 28 88 192
5 Entropy 13 00841 i048 Entropy 13 00841 i049 Entropy 13 00841 i050
27 93 144 195 31 88 141 197 29 90 141 195
6 Entropy 13 00841 i051 Entropy 13 00841 i052 Entropy 13 00841 i053
33 93 121 138 193 34 92 121 141 192 32 91 120 140 192
Table 7. Multi-level segmentation for peppers image.
Table 7. Multi-level segmentation for peppers image.
mGAPSOABC
3 Entropy 13 00841 i054 Entropy 13 00841 i055 Entropy 13 00841 i056
29 70 25 61 25 65
4 Entropy 13 00841 i057 Entropy 13 00841 i058 Entropy 13 00841 i059
23 64 111 24 69 106 26 66 108
5 Entropy 13 00841 i060 Entropy 13 00841 i061 Entropy 13 00841 i062
31 67 103 198 29 68 104 201 28 68 105 199
6 Entropy 13 00841 i063 Entropy 13 00841 i064 Entropy 13 00841 i065
30 70 106 147 204 31 70 104 146 201 30 70 104 148 203
Table 8 shows the average computation time of the three different methods for the four images with m = 3 to 6. As indicated in this table, with an increase in the class number m, the runtime tends to increase slowly due to the more complicated criterion function. For the Lena image, the computation algorithm of ABC increases from 1.6611 s to 1.7740 s when m increases from 3 to 6. Clearly, the efficiency of the intelligence optimization is far higher than traditional exhaustive method. Besides, the ABC costs the least time among all three algorithms.
Table 8. Average computation time (30 runs).
Table 8. Average computation time (30 runs).
ImagemGAPSOABC
Synthetic31.88491.84791.8210
41.94871.89301.8316
51.97801.95071.8709
62.04532.03671.9807
Lena31.91431.89311.6611
41.90361.91151.6594
52.01761.95001.7121
62.15982.05771.7740
Cameraman31.87551.77081.7083
41.91681.85581.7003
51.96861.87221.7975
62.01751.99331.9191
Peppers31.88901.85401.5321
41.96361.84951.5527
51.99851.93501.6924
62.13952.01221.6484

5.3. Influence of the Parameter q

The final experiment is to demonstrate the influence of parameter q on the segmentation result. We use the rose image as the example and let q = 0.2, 0.5, 1, and 4. The segmentation results are shown in Figure 8. The rose image is a super-extensive system since the correlation of the pixels of the flower are long-range, so the MTT segmentation of q = 0.2 (sub-extensive) and q = 1(extensive) performed a non-satisfying segmentation result, and the MTT criterion of q = 4 segmented the rose image best. However, when q = 8, we do not get a good result since it corresponds to a too long-range correlation image.
Figure 8. of long-range correlation image with variations of parameter q.
Figure 8. of long-range correlation image with variations of parameter q.
Entropy 13 00841 g008
Another example is a texture image of which a large amount of foreground objects cluster together as shown in Figure 9, so the correlation between pixels of the same object is short range. In this case, the small q (q = 1) performs best.
Figure 9. Segmentation of short-range correlation image with variations of parameter q.
Figure 9. Segmentation of short-range correlation image with variations of parameter q.
Entropy 13 00841 g009
From the above two examples, we can see that the parameter q in an image is usually interpreted as a quantity characterizing the degree of nonextensivity. For image segmentation, the nonextensivity of the image system can be justified by the presence of correlations between pixels of the same object in the image. The correlations can be regarded as a long-range correlation in the case of the image that presents pixels strongly correlated in gray-levels and space fulfilling.

6. Conclusions

In this paper, we compared the maximum Tsallis entropy (MTT) with traditional criteria such as MBCVT, MET, and MCET. We found the segmentation by MTT can give better results than other criteria in many cases. When MTT is applied to multi-level thresholding, the ABC algorithm is introduced and it requires less time than GA and PSO. Therefore, the proposed method is both effective and rapid.
In the experimental section, we obtained all the values of q by a trial-and-error method. We found q = 4 performs good for most natural images, but for short-range correlation images, especially textured imaged, reducing the value of q will give better segmentation, and for some extremely-long correlation images, increasing the value of q is better. Nevertheless, there is no reliable evidence of a formal relationship between the parameter q and the image category.
The q parameter in the Tsallis entropy is used as an adjustable value, which plays an important role as a tuning parameter in the image segmentation. The future work will mainly focus on: 1) designing an automatic q-adjusting method and applying the Tsallis entropy to other industrial fields; 2) testing other nonextensive entropy including the Jensen-Tsallis divergence [16] and Renyi entropy [28].

Acknowledgment

The research was conducted with the following support: (1) National Natural Science Foundation of China (#60872075); (2) National Technical Innovation Project Essential Project Cultivate Project (#706928) and (3) Nature Science Fund in Jiangsu Province (#BK2007103). We gratefully appreciate the help of Jackie Penn, who reviewed the whole paper and corrected the English grammar.

References

  1. Lázaro, J.; Martín, J.L.; Arias, J.; Astarloa, A.; Cuadrado, C. Neuro semantic thresholding using OCR software for high precision OCR applications. Image Vision Comput. 2010, 28(4), 571–578. [Google Scholar] [CrossRef]
  2. Xue, Z.; Ming, D.; Song, W.; Wan, B.; Jin, S. Infrared gait recognition based on wavelet transform and support vector machine. Patt. Recog. 2010, 43(8), 2904–2910. [Google Scholar] [CrossRef]
  3. Anagnostopoulos, G.C. SVM-based target recognition from synthetic aperture radar images using target region outline descriptors. Nonlinear Anal.-Theor. Meth. App. 2009, 71(12), e2934–e2939. [Google Scholar] [CrossRef]
  4. Hsiao, Y.-T.; Chuang, C.-L.; Lu, Y.-L.; Jiang, J.-A. Robust multiple objects tracking using image segmentation and trajectory estimation scheme in video frames. Image Vision Comput. 2006, 24(10), 1123–1136. [Google Scholar] [CrossRef]
  5. Doelken, M.T.; Stefan, H.; Pauli, E.; Stadlbauer, A.; Struffert, T.; Engelhorn, T.; Richter, G.; Ganslandt, O.; Doerfler, A.; Hammen, T. 1H-MRS profile in MRI positive- versus MRI negative patients with temporal lobe epilepsy. Seizure 2008, 17(6), 490–497. [Google Scholar] [CrossRef] [PubMed]
  6. Wang, S.; Chung, F.-l.; Xiong, F. A novel image thresholding method based on Parzen window estimate. Patt. Recog. 2008, 41(1), 117–129. [Google Scholar] [CrossRef]
  7. Fan, S.-K.S.; Lin, Y. A multi-level thresholding approach using a hybrid optimal estimation algorithm. Patt. Recog. Lett. 2007, 28(5), 662–669. [Google Scholar] [CrossRef]
  8. Zahara, E.; Fan, S.-K.S.; Tsai, D.-M. Optimal multi-thresholding using a hybrid optimization approach. Patt. Recog. Lett. 2005, 26(8), 1082–1095. [Google Scholar] [CrossRef]
  9. Nakib, A.; Oulhadj, H.; Siarry, P. Non-supervised image segmentation based on multiobjective optimization. Patt. Recog. Lett. 2008, 29(2), 161–172. [Google Scholar] [CrossRef]
  10. Farrahi Moghaddam, R.; Cheriet, M. A multi-scale framework for adaptive binarization of degraded document images. Patt. Recog. 2010, 43(6), 2186–2198. [Google Scholar] [CrossRef]
  11. Chung, K.-L.; Tsai, C.-L. Fast incremental algorithm for speeding up the computation of binarization. Appl. Math. Comput. 2009, 212(2), 396–408. [Google Scholar] [CrossRef]
  12. Huang, D.-Y.; Wang, C.-H. Optimal multi-level thresholding using a two-stage Otsu optimization approach. Patt. Recog. Lett. 2009, 30(3), 275–284. [Google Scholar] [CrossRef]
  13. Wang, N.; Li, X.; Chen, X.-h. Fast three-dimensional Otsu thresholding with shuffled frog-leaping algorithm. Patt. Recog. Lett. 2010, 31(13), 1809–1815. [Google Scholar] [CrossRef]
  14. Horng, M.-H. A multilevel image thresholding using the honey bee mating optimization. Appl. Math. Comput. 2010, 215(9), 3302–3310. [Google Scholar] [CrossRef]
  15. Yin, P.-Y. Multilevel minimum cross entropy threshold selection based on particle swarm optimization. Appl. Math. Comput. 2007, 184(2), 503–513. [Google Scholar] [CrossRef]
  16. Hamza, A.B. Nonextensive information-theoretic measure for image edge detection. J. Electron. Imag. 2006, 15(1), 1–8. [Google Scholar]
  17. Parvan, A.S. Critique of multinomial coefficients method for evaluating Tsallis and Rényi entropies. Physica A 2010, 389(24), 5645–5649. [Google Scholar] [CrossRef]
  18. Nakib, A.; Oulhadj, H.; Siarry, P. Image thresholding based on Pareto multiobjective optimization. Eng. Appl. Artif. Intell. 2010, 23(3), 313–320. [Google Scholar] [CrossRef]
  19. Maitra, M.; Chatterjee, A. A hybrid cooperative-comprehensive learning based PSO algorithm for image segmentation using multilevel thresholding. Expert Syst. Appl. 2008, 34(2), 1341–1350. [Google Scholar]
  20. Karaboga, N.; Kalinli, A.; Karaboga, D. Designing digital IIR filters using ant colony optimisation algorithm. Eng. Appl. Artif. Intell. 2004, 17(3), 301–309. [Google Scholar]
  21. Karaboga, D.; Basturk, B. On the performance of artificial bee colony (ABC) algorithm. Appl. Soft Comput. 2008, 8(1), 687–697. [Google Scholar]
  22. Xu, C.; Duan, H.; Liu, F. Chaotic artificial bee colony approach to Uninhabited Combat Air Vehicle (UCAV) path planning. Aerosp. Sci. Technol. 2011, in press. [Google Scholar]
  23. Guida, A.; Nienow, A.W.; Barigou, M. Shannon entropy for local and global description of mixing by Lagrangian particle tracking. Chem. Eng. Sci. 2010, 65(10), 2865–2883. [Google Scholar]
  24. Campos, D. Real and spurious contributions for the Shannon, Rényi and Tsallis entropies. Physica A 2010, 389(18), 3761–3768. [Google Scholar]
  25. Portes de Albuquerque, M.; Esquef, I.A.; Gesualdi Mello, A.R. Image thresholding using Tsallis entropy. Patt. Recog. Lett. 2004, 25(9), 1059–1065. [Google Scholar]
  26. Kang, F.; Li, J.; Xu, Q. Structural inverse analysis by hybrid simplex artificial bee colony algorithms. Comput. Struct. 2009, 87(13–14), 861–870. [Google Scholar]
  27. Özbakir, L.; Baykasoglu, A.; Tapkan, P. Bees algorithm for generalized assignment problem. Appl. Math. Comput. 2010, 215(11), 3782–3795. [Google Scholar]
  28. Zarezadeh, S.; Asadi, M. Results on residual Rényi entropy of order statistics and record values. Inform. Sci. 2010, 180(21), 4195–4206. [Google Scholar]

Share and Cite

MDPI and ACS Style

Zhang, Y.; Wu, L. Optimal Multi-Level Thresholding Based on Maximum Tsallis Entropy via an Artificial Bee Colony Approach. Entropy 2011, 13, 841-859. https://doi.org/10.3390/e13040841

AMA Style

Zhang Y, Wu L. Optimal Multi-Level Thresholding Based on Maximum Tsallis Entropy via an Artificial Bee Colony Approach. Entropy. 2011; 13(4):841-859. https://doi.org/10.3390/e13040841

Chicago/Turabian Style

Zhang, Yudong, and Lenan Wu. 2011. "Optimal Multi-Level Thresholding Based on Maximum Tsallis Entropy via an Artificial Bee Colony Approach" Entropy 13, no. 4: 841-859. https://doi.org/10.3390/e13040841

Article Metrics

Back to TopTop