Next Article in Journal
Quantum Information and Foundations
Previous Article in Journal
Analysis of the Influence of the Conduction Sub-Model Formulation on the Modeling of Laser-Induced Incandescence of Diesel Soot Aggregates
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Context-Aware Superpixel and Bilateral Entropy—Image Coherence Induces Less Entropy

1
School of Information Science and Technology, Northwest University, Xi’an 710027, China
2
School of Computer Science and Technology, Xi’an University of Posts and Telecommunications, Xi’an 710121, China
3
Shaanxi Key Laboratory of Network Data Analysis and Intelligent Processing, Xi’an University of Posts and Telecommunications, Xi’an 710121, China
4
State-Province Joint Engineering and Research Center of Advanced Networking and Intelligent Information Services, School of Information Science and Technology, Northwest University, Xi’an 710127, China
*
Authors to whom correspondence should be addressed.
Entropy 2020, 22(1), 20; https://doi.org/10.3390/e22010020
Submission received: 24 November 2019 / Revised: 18 December 2019 / Accepted: 20 December 2019 / Published: 23 December 2019

Abstract

:
Superpixel clustering is one of the most popular computer vision techniques that aggregates coherent pixels into perceptually meaningful groups, taking inspiration from Gestalt grouping rules. However, due to brain complexity, the underlying mechanisms of such perceptual rules are unclear. Thus, conventional superpixel methods do not completely follow them and merely generate a flat image partition rather than hierarchical ones like a human does. In addition, those methods need to initialize the total number of superpixels, which may not suit diverse images. In this paper, we first propose context-aware superpixel (CASP) that follows both Gestalt grouping rules and the top-down hierarchical principle. Thus, CASP enables to adapt the total number of superpixels to specific images automatically. Next, we propose bilateral entropy, with two aspects conditional intensity entropy and spatial occupation entropy, to evaluate the encoding efficiency of image coherence. Extensive experiments demonstrate CASP achieves better superpixel segmentation performance and less entropy than baseline methods. More than that, using Pearson’s correlation coefficient, a collection of data with a total of 120 samples demonstrates a strong correlation between local image coherence and superpixel segmentation performance. Our results inversely support the reliability of above-mentioned perceptual rules, and eventually, we suggest designing novel entropy criteria to test the encoding efficiency of more complex patterns.

1. Introduction

Superpixels are groups of coherent pixels. Instead of the huge amount of pixels in the uniform grid, only several superpixels in irregular shape are the elementary unit of an image, which serve in a number of subsequent computer vision methods. At the very beginning, superpixel designers take inspiration from Gestalt grouping rules [1]. Although the correlation between such perceptual rules and visual efficiency is not well interpreted yet, Gestalt researchers, based on behavioral evidence, hypothesized the minimum principle that illustrates the efficiency of visual processes [2]. On the other hand, the employment of superpixels has greatly reduced image redundancy and has significantly improved the processing efficiency of a number of computer vision methods in practice, including object segmentation [3,4,5], recognition [6], location [7] and tracking [8], and so forth.
While the underlying mechanisms of visual efficiency are not figured out yet, vision researchers continuously discover mechanisms supporting that stimuli coherence depicted by Gestalt grouping rules accelerates vision formation, even at the very early stages of perception (i.e., in the visual working memory) [9,10,11]. However, since the human brain is extremely complex, it is impossible to evaluate the efficiency of transmitting, encoding, or processing visual information in vivo so far. Only a few pioneer researchers attempt to quantitatively figure out that question in virtue of information theory [12,13]. They hypothesized that the human visual system transmits visual information with the utmost economy either with respect to time or bandwidth.
As an alternative of the brain, an increasing number of computer vision systems serve as experimental platforms in testing research hypotheses [14,15,16]. Those artificial vision systems even excel humans on a number of practical tasks, owing to the advance of computer vision techniques; and they can be a reliable tool assisting humans in routine tasks [17,18,19,20]. Therefore, a superpixel method that completely follows perceptual rules can be exploited to test the Gestalt minimum hypothesis. Nevertheless, there are limitations on conventional superpixel algorithms and the image entropy.
Firstly, conventional superpixel algorithms do not completely follow Gestalt grouping rules. Normalized cuts is an Eigen-based method [1], which is very time-consuming. Next, Achanta et al. (2012) proposed the simple linear iterative clustering (SLIC) which iteratively aggregates pixels based on the K-means clustering in a 5D Euclidean space [21]. Because SLIC can only utilize local image characteristics, which is less effective, Li and Chen proposed linear spectral clustering (LSC) which can capture perceptually important global features [22]. Nevertheless, these methods only generate the flat image partition rather than hierarchical image structures like a human does. Besides, those algorithms should be initialized by manually specifying the total number of superpixels, which could hardly suit to various images under diverse environments.
Secondly, conventional image entropy [23] merely involves the overall intensity statistics which cannot respond to intensity patterns over spatial, as shown in Figure 1. Such intensity or visual patterns over spatial are definitely crucial to any biological or artificial vision system. First of all, the ventral and dorsal two visual pathways transmit and ’manipulate’ the physical appearance and spatial information, respectively [24], viz., visual patterns involve the visuospatial domain. Second and more apparently, Gestalt grouping rules involve stimuli patterns of both aspects [25]. Last but not the least, surrounding objects (i.e., visual context) modulate the firing rate of neurons representing the target object [26,27], which affects the visual awareness inherently.
In this paper, we propose a context-aware superpixel (CASP) algorithm and the bilateral entropy (BE), aiming to test the encoding efficiency of Gestalt grouping rules. CASP iteratively aggregates pixels from coarse to fine and each clustering iteration explicitly follows the similarity and proximity grouping rules. The coarse-level superpixels with contextual information guide to generating superpixels in a fine-level. Besides CASP, we propose BE with two aspects, that is, conditional intensity entropy and spatial occupation entropy which involve local image coherence, spatial occupation respectively. With extensive experiment results, CASP demonstrates promising superpixel segmentation performance, compared with SLIC and LSC; while CASP will achieve the smallest entropy of all three criteria when it segments coherent objects out precisely. Besides, further results demonstrate a strong correlation between local image coherence and superpixel segmentation performance, based on Pearson’s correlation coefficient, in a collection of data with a total of 120 samples. Our results support the visual system inclines to encode visual information in the most efficient fashion.
Our paper is organized as follows. In Section 2, we will describe the workflow of CASP. In Section 3, we will give a detailed description of BE. In Section 4, we will demonstrate the effectiveness of CASP with respect to both segmentation performance and encoding efficiency. Finally, in Section 5, we will discuss and conclude our works.

2. Context-Aware Superpixel (CASP)

We design CASP to generate a hierarchy of large to small superpixels that represent the input image from coarse to fine. To achieve this, CASP iteratively partitions an image, and each iteration contains 3 stages, including (1) similarity clustering, which aggregates pixels with similar physical appearance; and (2) proximity clustering, which defines proximity by connectedness and isolates spatially unconnected groups; and (3) merging stage, which merges tiny groups into the most similar neighbor.
The workflow of CASP is vividly illustrated in Figure 2, each column, with three blue boxes denoting three clustering stages respectively, demonstrates one clustering iteration; and the image below each box is outcome label map of the corresponding clustering stage. From the left column to the right, CASP generates superpixels from coarse to fine. The preceding clustering iteration generates coarse-level aggregations that serve as the context to subsequent clustering. The coarse-level outcomes are next entered into the next iteration of clustering separately to adopt fine-level ones. By this, CASP can gradually pop out the less salience image regions (i.e., circular halo in the background region), as shown in Figure 2.
In the first stage, CASP employs expectation maximization (EM) clustering in simulating the Gestalt similarity grouping rule. Denote an image with N pixels, I ( x ) = I ( x 1 ) , , I ( x i ) , , I ( x N ) , EM clustering aggregates them into two coherent groups with labels 1 , 2 , which intensity distributions are represented by two Gaussian models, with a parameter set θ t , that has maximized the lower-bound of log-likelihood [28],
arg max θ t , * 1 , 2 p I ( x i ) , θ t , log p I ( x i ) , θ t , * ,
where θ t , = μ t , , δ t , consistes of mean μ t , k and standard deviation (STD) δ t , of the cluster at t-th level.
The input pixels are split into two parts, which are denoted by two colors in the label map of 1st similarity clustering, shown in Figure 2. Next, proximity clustering defines connectedness as a type of proximity for simulating Gestalt proximity grouping rule. It detects outcome connectivity of first stage and gives each isolated part a unique label. The label map hence becomes more colorful. After proximity clustering, the adopted clusters may be tiny, even only have 1 pixel. The third stage hereby merges tiny groups into the most similar neighbor. Finally, CASP derives superpixels in a specific coherence level.
Initialized by four parameters. They heuristically adapt the total number of superpixels to diverse images. They will be introduced in Section 4.1 in detail.

3. Bilateral Entropy

Gestalt grouping rules do not only shed light on how inner consciousness represents afferent visual stimuli but also are the strategies to aggregate coherent pixels into superpixels. They emphasize the crucial function of both physical appearance and spatial information. However, conventional image entropy cannot respond to intensity patterns over spatial,
H IM ( I ( x ) ) = - I ( x ) = 0 255 p ( I ( x ) ) log 2 p ( I ( x ) ) ,
where H IM ( I ( x ) ) merely measures overall intensity statistics. If images share the same intensity histogram, it will generate an identical value, regardless of their content differences.
To evaluate the encoding efficiency of superpixels or image coherence, we propose bilateral entropy that extends the conventional image entropy for concurrently considering the local image coherence and spatial occupation,
H BE I ( x ) , L ( x ) = - I ( x ) = 0 , L ( x ) = 1 255 , K p I ( x ) , L ( x ) log 2 p I ( x ) , L ( x ) ,
where the joint probability distribution
p I ( x ) , L ( x ) = p I ( x ) | L ( x ) p L ( x ) ,
thus, we can derive the bilateral entropy in the form,
H BE I ( x ) , L ( x ) = - I ( x ) = 0 , L ( x ) = 1 255 , K p ( I ( x ) | L ( x ) ) p ( L ( x ) ) log 2 { p ( I ( x ) | L ( x ) ) p ( L ( x ) ) } ,
where L ( x ) denotes the label of superpixels; p ( L ( x ) ) is the occupation ratio; and p ( I ( x ) | L ( x ) ) is local intensity histogram of L ( x ) -th superpixel. Typically, H BE I ( x ) , L ( x ) can be rewritten by
H BE I ( x ) , L ( x ) = H CON I ( x ) | L ( x ) + H OCC L ( x ) ,
where H CON I ( x ) | L ( x ) and H OCC L ( x ) measure local intensity statistic and spatial occupation respectively. Both terms involve the complexity of a visual task reflected by similarity and proximity Gestalt grouping rules and we propose H BE I ( x ) , L ( x ) which is the sum of both terms to integrally measure both aspects with the same significance weight.

4. Results

In this section, we will first introduce the parameter settings of CASP then describe the evaluation criteria; and next, two baseline methods will be introduced for comparison. At last, we will report the experimental results w.r.t. superpixel segmentation performance and encoding efficiency.

4.1. Parameter Settings

In all experiments, we initialize the parameters of CASP as follows:
  • Maximal clustering iteration limit D p : D p determines the maximal depth of the superpixel hierarchy, which is 7 to acquire superpixels in 6 levels.
  • Maximal cluster size Mx c : We set M x c to 100. During each similarity clustering stage, only the group of input pixels which size exceeds 100 is divided into subgroups.
  • Minimal cluster size Mn c : We set M x c to 4. During each merging stage, a cluster with a size smaller than 4 is merged into the most similar neighbor.
  • neighbourhood range N e : We initialized N e to 8 to retrieve 8 neighbors of a pixel.

4.2. Evaluation Criteria

Under-segmentation error ( U S E ), boundary recall ( B R ) and achievable segmatiaon accuracy ( A S A ) are the most commonly used criteria to evaluate superpixel segmentation performance. The U S E , A S A and B R close to 0, 1 and 1 respectively are desired. Denote a ground-truth segmentation of an image as G = g 1 , , g i , , g K and a superpixel partition as S = s 1 , , s j , , s K .
  • USE : Given a ground-truth segmentation g i , U S E measures the fraction of pixels that leak across its boundaries caused by the overlapping superpixels s j [29],
    U S E = s j S : A r e a ( s j g i ) > B j A r e a ( s j ) - A r e a ( g i ) A r e a ( g i ) ,
    where A r e a ( · ) counts the pixel amount of an image region. We set B j to 25 percent of A r e a ( s j ) because CASP may generate the smallest superpixel with a size of 4. We use the average of all segments in G .
  • BR : Given the boundaries of ground-truth segmentations, denoted by δ G , B R measures the percentage of the ground-truth boundaries recovered by superpixels, denoted by δ S . We compute B R by
    B R = p δ G l l ( min q δ S p - q < ϵ ) | δ G | ,
    where p and q denote the location of pixels belong to δ G and δ S respectively. We set ϵ to 1 so that count a pixel at q that falls within a 1 pixel range of q, which operation is denoted by l l .
  • ASA : Given all ground-truth segmentations G , A S A gives the largest overlapping area between superpixels and the ground truth segments [30]
    A S A = s j S max θ ( g i G ) { A r e a ( s j ) A r e a ( g i ) ) } g i G A r e a ( g i ) .
    In general, A S A is the highest achievable accuracy for object segmentation that utilizes superpixels as units.

4.3. Baseline Methods

We compare CASP with two baseline methods, SLIC and LSC, which need to be initialized by the total number of superpixels. Because CASP heuristically generates the total numbers in 6 levels, we specify them to the baseline methods, which will be listed in Section 4.4.
  • SLIC employs a balance factor to weigh the importance of color and coordinates [21]. We initialize the balance factor 40 and merged tiny image regions that radius is smaller than 1.
  • LSC approximates the coherence metric using a kernel function and maps color and coordinates to a high dimensional feature space [22]. We set the compactness factor to 0 . 066 .

4.4. Evaluation Datasets

  • BSD300 [31] is a natural image database, which suits conventional methods working in chrominance space. However, the labels are manually placed, which are not precise, shown in Figure 4a. We hence do not use it for quantitative evaluation. To initialize LSC and SLIC, the superpixel numbers in 6 levels are [ 24 , 532 , 1121 , 2253 , 4390 , 6115 ] .
  • MRBrainS [32] contains 3T T1-weighted magnetic resonance (MR) brain scans. We employ 4 subjects from the training dataset, each of which contains 48 slices. The labels are carefully placed, including gray matter (GM), white matter (WM) and cerebrospinal fluid (CSF). The background image regions are removed out and filled with an intensity of 0. As the MR images are in the UINT16 encoding format, which cannot be used by the LSC, we transform the intensity range to UINT8 and duplicate each slice to 3-channels. To initialize LSC and SLIC, the superpixels number of 6 levels are [ 2 , 15 , 136 , 392 , 595 , 615 ] .
    And next, for testing the segmentation performance under noise corruption, we add spatially varying Rician noise to MR magnitudes. To achieve this, we firstly simulate phase maps according to the literature [33]; and next, we generated both real and imaginary components and added spatially varying Gaussian noise to each component separately. The noise level follows the 2D Gaussian distribution in the same manner as the literature [34]. Based on the noisy components, we eventually generate the noisy magnitude with Rician noise using the simple sum of square (SoS) image reconstruction manner [35]. More specifically, we added 4 levels of spatially varying Gaussian noise to the complex components, with maximal noise levels N max = { 10 , 20 , 30 , 50 } . The noisy images and residual noise maps are shown in Figure 3.

4.5. Segmentation Performance

Figure 4a shows the ground-truth segments of the eagle image, including the superpixel boundaries and a label map for clear graphic-illustration. We can find ground-truth segments do not well adhere to true boundaries. We hence do not quantitatively evaluate superpixels on BSD300. Our qualitative results are shown in Figure 4b. Prominently, we can find CASP partitioned the image from coarse to fine, image details are gradually popped out.
Figure 5a shows an MR brain slice from MRBrainS. The ground truth segments with 4 labels are denoted by 4 different colors in the label map. We can see from Figure 5b, in the row of iteration 1, SLIC fails to segment the intracranial region, while LSC and CASP successfully achieve this. More apparently, CASP aggregates the background as one unit and extracts more details in the intracranial regions.
Figure 6a–c demonstrate the quantitative superpixel segmentation performance w.r.t. U S E , B R , and A S A respectively. The solid curves denote the mean across average values of 4 subjects, while the shades denote the STD. We can find the CASP outperforms others.

4.6. Encoding Efficiency

We first synthesize two groups of toy images as shown in Figure 7a,b respectively. Each group contains three images with the same pixel set but in 3 different patterns. From the left column to the right, the superpixel number increases significantly.
In Figure 7a, each of the three synthetic images only has 256 pixels, with a size of 16 × 16 . Every pixel is unique in intensity, which value involves the UINT8 range from 0 to 255. The synthetic images along with label maps were used to generate the H OCC L ( x ) , H CON I ( x ) | L ( x ) , and H BE I ( x ) , L ( x ) , attached below each label map. We can find H BE I ( x ) , L ( x ) degenerates to H IM ( I ( x ) ) , which equals to 8 all the times. From the left column to the right, with the increase of superpixel number, H OCC L ( x ) increases, and H CON I ( x ) | L ( x ) decreases synchronously, demonstrating their complementary role. When H CON I ( x ) | L ( x ) = 0 , H BE I ( x ) , L ( x ) degenerates to H OCC L ( x ) . The degeneration indicates that the uncertainty contribution will only exist in the group of noticed objects if a visual system has focused on the constant objects. Typically, such uncertainty among objects may equal to the one in their exterior physical appearance.
In Figure 7b, the synthetic images had a size of 256 × 256 . Pixels are roughly in 4 intensity levels, each of which is biased by small-level Gaussian noise. These pixel intensities may overlap. In the third column, we organize the image so that the same pixels belong to one superpixel. The intensity of different superpixels may be the same. We can find the bilateral entropy decreases with the increase of local coherence. Typically, it is interesting to find even when the total number of superpixels is far more than the one in the middle column, and the bilateral entropy has been decreased.
The curves and shades in Figure 8a–c demonstrate the mean and STD values across 4 subjects of H CON I ( x ) | L ( x ) , H OCC L ( x ) , and H BE I ( x ) , L ( x ) , respectively. The descending curve slope in Figure 8a indicates the enhancement of local coherence within superpixels. While as shown in Figure 8b, the increasing amount of superpixels drives ascending curve slope. The integral function of both terms is illustrated by H BE I ( x ) , L ( x ) , shown in Figure 8c. Most prominently, we can see CASP achieves the smallest values for all three cases. In addition, CASP increases H OCC and H BE much less, outperforming LSC and SLIC significantly, although the total number of superpixels of the three methods are roughly identical.

4.7. Segmentation Robustness and Local Image Coherence

Since the human visual system is widely believed to be robust to arbitrary interference, we further evaluate CASP on the MR data, which are manually added spatially varying Rician noise. Accordingly, we have a large number of MR data, both noise free and noisy ones. Ultimately, we evaluate the correlation between local coherence, H CON I ( x ) | L ( x ) and segmentation performance, including U S E , B R and A S A , respectively.
Figure 9 shows both superpixel segmentation, which is in the left column, and encoding efficiency performance, which is in the right column. Different methods are marked by curve colors, while the performance under different noise conditions are denoted by the line types. In general, the higher the noise level, the worse the segmentation performance, and the robustness of a method behaves through, besides better performance, the shape consistency of the group of curves demonstrating under different noise levels. We can find in Figure 9 that, CASP demonstrates superiority on the two aspects, which curves not only locate at places indicating better performance but also demonstrate higher-level consistency in curve shape. By contrast, the curves of SLIC demonstrate arbitrary shapes indicating vulnerability to noise.
Comparing the two sub-figures in the first row of Figure 9, it is interesting to find the H CON curves of SLIC seem to correlate with its U S E curves. To find out the relationship between local image coherence and superpixel segmentation performance, we employ Pearson’s correlation coefficient (PCC).
Figure 10 illustrate the correlation between H CON and U S E , B R , and A S A respectively. Each blue circle denotes a subject (4 in total) that is represented by superpixels in a specific iteration (6 in total) and noise level (5 in total, including noise free). There are 120 samples used to generate PCC in total. No matter from the shape of scattered circles or PCC values, our results support that local image coherence and superpixel segmentation has a strong correlation.

5. Discussions and Conclusions

In this paper, we proposed a superpixel clustering algorithm, named context-aware superpixel, which aggregated pixels explicitly following two of Gestalt grouping rules and top-down hierarchical principle. Next, we proposed the conditional intensity entropy H CON I ( x ) | L ( x ) , spatial occupation entropy H OCC L ( x ) , and bilateral entropy H BE I ( x ) , L ( x ) in the context of superpixel segmentation to test the encoding efficiency of local image coherence, as an alternative of the complex brain.
In the superpixel segmentation task, as shown in Figure 4 and Figure 5, CASP generated superpixels that fitted image contents more precisely than SLIC and LSC. The superiority of CASP attributed to the hierarchical image representation. First, CASP avoided to manually specify the total number of superpixels, like the hierarchical model enables the learning of prior knowledge [36,37]; and second, CASP discerned fine details based on a coarse aggregation, which took advantage of the information in coarse-level [26,38]. As a result, the superpixels adopted by CASP in the fine-level reserved the boundaries of superpixels in the coarse-level. By contrast, the well-segmented boundaries of SLIC and LSC in coarse-level did not present in the fine-level. They can not generate the hierarchical representations although they were initialized by a list of small to large numbers of superpixels. As shown in Figure 6 and Figure 9, quantitative results consistently supported the superiority of CASP.
Theoretically speaking, superpixel methods simulate how the human visual system reduces redundancy. Luck and Vogel (1997) claims that what maintains in visual working memory are the integrated objects rather than individual features [39]; besides, the working memory is only capable of maintaining a limited number of objects at the same time [40]. Therefore, discarding trivial information that does not affect recognition improves the visual efficiency [41].
Meanwhile, Gestalt perceptual rules play a crucial role in both visual working memory and superpixel methods, and they manifest the visual efficiency of the human visual system. Most prominently, contrasting to Barlow’s (1961) redundancy reduction [41], Gestalt grouping rules, or superpixels that did not discard any information but did structure pixels, demonstrated improvement on encoding efficiency. To test the encoding efficiency of superpixels, we proposed three entropy criteria. From Equation (6), we can find the H CON I ( x ) | L ( x ) and H OCC were the two aspects of H BE that concurrently measured local image coherence and spatial occupation. From Figure 8 and Figure 9, we can see CASP achieved less entropy than LSC and SLIC in all three criteria. Specifically, the local image coherence strongly correlates with superpixel segmentation performance, as supported by Pearson’s correlation analysis shown in Figure 10.
To sum up, partitioning image according to image coherence, although without any loss of detail information, induced less entropy. Such results support Gestalt minimal principle and might imply an underlying principle, that is, the minimum entropy principle that a visual system reduces the uncertainty existing in visual stimuli efficiently if the stimuli have a minimal entropy.
Compared to Edwin Jaynes’ (1957) maximum entropy principle, our conclusion seems to contradict. We denote different two aspects. For Edwin Jaynes’, the probability distribution with maximum entropy is the a priori criterion used in an inference process when little or even no information is available [42]. During the inference iterations, this prior distribution changes to have a sharp shape with newly available information [43]. By contrast, for ours, the minimum entropy principle involved the statistics within the data, and it was a measure of uncertainty of the image itself, with reference to local image coherence and spatially occupation. Although they were both named entropy, their roles are different.
More specifically, it is worth noting that the proposed entropy criteria involve the image structure in a hierarchy captured by CASP following the top-down manner. If considering such a hierarchy as the inner structure of visual awareness, each superpixel may involve one visual object. Accordingly, the hierarchical structure is the information that is transmitted from raw stimuli to the visual system. Denote the mutual information
I m u I ( x ) , L ( x ) = H IM I ( x ) - H CON I ( x ) | L ( x ) ,
since H IM I ( x ) must be a constant, I m u I ( x ) , L ( x ) is determined by the H CON I ( x ) | L ( x ) . Thus, maximizing I m u I ( x ) , L ( x ) equals to minimizing H CON I ( x ) | L ( x ) . Because a smaller H CON I ( x ) | L ( x ) indicates a higher-level of intensity coherence of a superpixel, we can summarize that the way to represent image coherence minimizes the conditional entropy H CON I ( x ) | L ( x ) and maximizes mutual information I m u I ( x ) , L ( x ) , maximizes the information transmission, indicating visual efficiency. On the other side, the statistics of local image coherence reflect the complexity within the data. Thus, the local image coherence that is reflected by H CON I ( x ) | L ( x ) can be used to depict the complexity within the data, and it is obvious that an image with a simple structure is easily be recognized by human and be represented by CASP. We, therefore, suggest that the coherence-based image structure might be one type of the data structure of unlabelled data, anticipated by Hinton (2018) [44], which can be fully exploited. Furthermore, it is the coherence rather than the individual features that improve the encoding efficiency, and our results may respond to the query whether feature invariance is always an optimal strategy for biological and artificial vision [38].
As far as we know, this work is the first investigation to test Gestalt minimum hypothesis utilizing a superpixel algorithm (i.e., CASP) and entropy criteria (i.e., H CON I ( x ) | L ( x ) , H OCC L ( x ) and H CON I ( x ) | L ( x ) ). However, there are limitations worth to note. First, the artificial vision system is too simple to understand the true underlying visual mechanisms, and typically, the adopted entropy value might is unable to indicate the true number of firing neurons. Second, the bilateral entropy can only weigh the encoding efficiency of patterns at the very early perceptual stage, viz., the image coherence. Even so, the strong correlation between local image coherence and superpixel segmentation performance suggests the potential to exploit the proposed entropy criteria as constraints in designing a loss function of the segmentation tasks. Besides the intensity coherence over spatial, as claimed by Attneave (1956), visual patterns also demonstrate many other types, such as shape, angle, curvature, and so forth [13]. Simultaneously, Gestalt Berlin school claims a whole-part relationship that the whole is different from the sum of the parts [2]. Both reasons let us know that the proposed entropy criteria may fail to respond to more complex patterns. Therefore, we suggest that it should design specific entropy criterion to measure the encoding efficiency of complex patterns corresponding to specific vision level, by which human vision and computer vision researchers could test and refine their hypotheses theoretically and quantitatively.

Author Contributions

F.L. conceived the idea for this study, performed the research, wrote the codes, analyzed data, validated the methodology, wrote and revised the manuscript. X.Z. and H.W. smoothed and revised the manuscript. J.F. conceived the idea for this study, revised the manuscript, acquired funding, supervised the project. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Key Research and Development Program of China under Grant 2017YFB1002504, and the Xi’an Science and Technology Project funded by the Xi’an Science and Technology Bureau under Grant 201805060ZD11CG44.

Acknowledgments

The authors greatly appreciate the constructive suggestions and insightful guidance from anonymous reviewers. Besides, many thanks for the encouragement from Vincent Shang, we greatly appreciate his productive and professional works.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
CASPContext-Aware SuperPixel
SLICSimple Linear Iterative Clustering
LSCLinear Spectral Clustering
USEUnder-Segmentation Error
BRBoundary Recall
ASAAchievable Segmatiaon Accuracy
MRMagnetic Resonance
STDStandard Deviation
SoSSum of Square
PCCPearson’s Correlation Coefficient

References

  1. Shi, J. Normalized cuts and image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2000, 22, 888–905. [Google Scholar]
  2. Wagemans, J.; Elder, J.H.; Kubovy, M.; Palmer, S.E.; Peterson, M.A.; Singh, M.; von der Heydt, R. A century of Gestalt psychology in visual perception: I. Perceptual grouping and figure–ground organization. Psychol. Bull. 2012, 138, 1172–1217. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Liu, Y.J.; Yu, M.; Li, B.J.; He, Y. Intrinsic manifold SLIC: A simple and efficient method for computing content-sensitive superpixels. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 653–666. [Google Scholar] [CrossRef] [PubMed]
  4. Liu, F.; Feng, J.; Su, W.; Lv, Z.; Xiao, F.; Qiu, S. Normalized Euclidean Super-Pixels for Medical Image Segmentation. In Proceedings of the International Conference on Intelligent Computing, Liverpool, UK, 7–10 August 2017; pp. 586–597. [Google Scholar]
  5. Huo, J.; Wu, J.; Cao, J.; Wang, G. Supervoxel based method for multi-atlas segmentation of brain MR images. NeuroImage 2018, 175, 201–214. [Google Scholar] [CrossRef] [PubMed]
  6. Kaufhold, J.; Collins, R.; Hoogs, A.; Rondot, P. Recognition and segmentation of scene content using region-based classification. In Proceedings of the IEEE International Conference on Pattern Recognition (ICPR), Hong Kong, China, 20–24 August 2006; Volume 1, pp. 755–760. [Google Scholar]
  7. Fulkerson, B.; Vedaldi, A.; Soatto, S. Class segmentation and object localization with superpixel neighborhoods. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Kyoto, Japan, 27 September–4 October 2009; pp. 670–677. [Google Scholar]
  8. Wang, S.; Lu, H.; Yang, F.; Yang, M.H. Superpixel tracking. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Barcelona, Spain, 6–13 November 2011; pp. 1323–1330. [Google Scholar]
  9. Gao, T.; Gao, Z.; Li, J.; Sun, Z.; Shen, M. The perceptual root of object-based storage: An interactive model of perception and visual working memory. J. Exp. Psychol. Hum. Percept. Perform. 2011, 37, 1803–1823. [Google Scholar] [CrossRef] [Green Version]
  10. Peterson, D.J.; Berryhill, M.E. The Gestalt principle of similarity benefits visual working memory. Psychon. Bull. Rev. 2013, 20, 1282–1289. [Google Scholar] [CrossRef] [Green Version]
  11. Gao, Z.; Gao, Q.; Tang, N.; Shui, R.; Shen, M. Organization principles in visual working memory: Evidence from sequential stimulus display. Cognition 2016, 146, 277–288. [Google Scholar] [CrossRef]
  12. Attneave, F. Some informational aspects of visual perception. Psychol. Rev. 1954, 61, 183. [Google Scholar] [CrossRef]
  13. Attneave, F.; Arnoult, M.D. The quantitative study of shape and pattern perception. Psychol. Bull. 1956, 53, 452. [Google Scholar] [CrossRef]
  14. Marblestone, A.H.; Wayne, G.; Kording, K.P. Toward an integration of deep learning and neuroscience. Front. Comput. Neurosci. 2016, 10, 94. [Google Scholar] [CrossRef]
  15. Rajalingham, R.; Issa, E.B.; Bashivan, P.; Kar, K.; Schmidt, K.; DiCarlo, J.J. Large-scale, high-resolution comparison of the core visual object recognition behavior of humans, monkeys, and state-of-the-art deep artificial neural networks. J. Neurosci. 2018, 38, 7255–7269. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Gangopadhyay, P.; Das, J. Do Primates and Deep Artificial Neural Networks Perform Object Categorization in a Similar Manner? J. Neurosci. 2019, 39, 946–948. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Bronstein, M.M.; Bruna, J.; LeCun, Y.; Szlam, A.; Vandergheynst, P. Geometric deep learning: Going beyond euclidean data. IEEE Signal Process. Mag. 2017, 34, 18–42. [Google Scholar] [CrossRef] [Green Version]
  18. Wang, H.; Feng, J.; Bu, Q.; Liu, F.; Zhang, M.; Ren, Y.; Lv, Y. Breast mass detection in digital mammogram based on Gestalt psychology. J. Healthc. Eng. 2018, 2018. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  19. Wang, H.; Feng, J.; Zhang, Z.; Su, H.; Cui, L.; He, H.; Liu, L. Breast mass classification via deeply integrating the contextual information from multi-view data. Pattern Recognit. 2018, 80, 42–52. [Google Scholar] [CrossRef]
  20. Liu, F.; Feng, J.; Chen, G.; Wu, Y.; Hong, Y.; Yap, P.T.; Shen, D. DeepBundle: Fiber Bundle Parcellation with Graph Convolution Neural Networks. arXiv 2019, arXiv:1906.03051. [Google Scholar]
  21. Achanta, R.; Shaji, A.; Smith, K.; Lucchi, A.; Fua, P.; Süsstrunk, S. SLIC superpixels compared to state-of-the-art superpixel methods. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 34, 2274–2282. [Google Scholar] [CrossRef] [Green Version]
  22. Li, Z.; Chen, J. Superpixel segmentation using linear spectral clustering. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 1356–1363. [Google Scholar]
  23. González, R.C.; Woods, R.E.; Eddins, S.L. Digital Image Processing Using MALAB; Pearson: London, UK, 2009. [Google Scholar]
  24. Ungerleider, L.G.; Haxby, J.V. ‘What’and ‘where’in the human brain. Curr. Opin. Neurobiol. 1994, 4, 157–165. [Google Scholar] [CrossRef]
  25. Koffka, K. Perception: An introduction to the Gestalt-Theorie. Psychol. Bull. 1922, 19, 531–585. [Google Scholar] [CrossRef] [Green Version]
  26. Bar, M. Visual objects in context. Nat. Rev. Neurosci. 2004, 5, 617–629. [Google Scholar] [CrossRef]
  27. Lisman, J. The challenge of understanding the brain: Where we stand in 2015. Neuron 2015, 86, 864–882. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  28. Do, C.B.; Batzoglou, S. What is the expectation maximization algorithm? Nat. Biotechnol. 2008, 26, 897–899. [Google Scholar] [CrossRef] [PubMed]
  29. Levinshtein, A.; Stere, A.; Kutulakos, K.N.; Fleet, D.J.; Dickinson, S.J.; Siddiqi, K. Turbopixels: Fast superpixels using geometric flows. IEEE Trans. Pattern Anal. Mach. Intell. 2009, 31, 2290–2297. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  30. Liu, M.Y.; Tuzel, O.; Ramalingam, S.; Chellappa, R. Entropy rate superpixel segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Springs, CO, USA, 20–25 June 2011; pp. 2097–2104. [Google Scholar]
  31. Martin, D.; Fowlkes, C.; Tal, D.; Malik, J. A Database of Human Segmented Natural Images and its Application to Evaluating Segmentation Algorithms and Measuring Ecological Statistics. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Vancouver, BC, Canada, 7–14 July 2001; pp. 416–423. [Google Scholar]
  32. Mendrik, A.M.; Vincken, K.L.; Kuijf, H.J.; Breeuwer, M.; Bouvy, W.H.; De Bresser, J.; Alansary, A.; De Bruijne, M.; Carass, A.; El-Baz, A.; et al. MRBrainS challenge: Online evaluation framework for brain image segmentation in 3T MRI scans. Comput. Intell. Neurosci. 2015, 2015, 813696. [Google Scholar] [CrossRef] [Green Version]
  33. Pizzolato, M.; Fick, R.; Boutelier, T.; Deriche, R. Noise floor removal via phase correction of complex diffusion-weighted images: Influence on DTI and Q-space metrics. In International Conference on Medical Image Computing and Computer-Assisted Intervention; Springer: Cham, Switzerland, 2016; pp. 21–34. [Google Scholar]
  34. Liu, F.; Feng, J.; Yap, P.T.; Shen, D. Multi-Kernel Filtering: An Extension of Bilateral Filtering Using Image Context. arXiv 2019, arXiv:1908.06307. [Google Scholar]
  35. Chen, G.; Dong, B.; Zhang, Y.; Lin, W.; Shen, D.; Yap, P.T. Denoising of Infant Diffusion MRI Data via Graph Framelet Matching in x-q Space. IEEE Trans. Med. Imaging 2019. [Google Scholar] [CrossRef]
  36. Friston, K. Learning and inference in the brain. Neural Netw. 2003, 16, 1325–1352. [Google Scholar] [CrossRef]
  37. Badcock, P.B.; Friston, K.J.; Ramstead, M.J.; Ploeger, A.; Hohwy, J. The hierarchically mechanistic mind: An evolutionary systems theory of the human brain, cognition, and behavior. Cogn. Affect. Behav. Neurosci. 2019, 19, 1319–1351. [Google Scholar] [CrossRef]
  38. Lu, Y.; Yin, J.; Chen, Z.; Gong, H.; Liu, Y.; Qian, L.; Li, X.; Liu, R.; Andolina, I.M.; Wang, W. Revealing detail along the visual hierarchy: Neural clustering preserves acuity from V1 to V4. Neuron 2018, 98, 417–428. [Google Scholar] [CrossRef]
  39. Luck, S.J.; Vogel, E.K. The capacity of visual working memory for features and conjunctions. Nature 1997, 390, 279–281. [Google Scholar] [CrossRef]
  40. Baddeley, A. Working memory. Science 1992, 255, 556–559. [Google Scholar] [CrossRef] [PubMed]
  41. Barlow, H.B. Possible principles underlying the transformation of sensory messages. Sens. Commun. 1961, 1, 217–234. [Google Scholar]
  42. Jaynes, E.T. Information theory and statistical mechanics. Phys. Rev. 1957, 106, 620–630. [Google Scholar] [CrossRef]
  43. Jaynes, E.T. Information theory and statistical mechanics. II. Phys. Rev. 1957, 108, 171–190. [Google Scholar] [CrossRef]
  44. Hinton, G. Deep learning—A technology with the potential to transform health care. J. Am. Med. Assoc. 2018, 320, 1101–1102. [Google Scholar] [CrossRef]
Figure 1. Conventional image entropy is ineffective at responding to pixel patterns over spatial. Although the image may show significantly different patterns, it must derive identical image entropy if they have the same set of pixels. (a) The eagle image is selected from BSD300, with a background of sky and the foreground of one eagle standing on a tree. Its pixels at the background, foreground, and the whole regions are rearranged respectively so that generate 3 different images, as shown in (bd). Their entropy equals 6.3884 although the image contents are different.
Figure 1. Conventional image entropy is ineffective at responding to pixel patterns over spatial. Although the image may show significantly different patterns, it must derive identical image entropy if they have the same set of pixels. (a) The eagle image is selected from BSD300, with a background of sky and the foreground of one eagle standing on a tree. Its pixels at the background, foreground, and the whole regions are rearranged respectively so that generate 3 different images, as shown in (bd). Their entropy equals 6.3884 although the image contents are different.
Entropy 22 00020 g001
Figure 2. The clustering workflow of context-aware superpixel (CASP). Different colors in label maps denote the adopted cluster labels. We can see the whole yellow background of the outcome of 1st similarity clustering is split into several connected parts by the 1st proximity clustering. The outcome of proximity clustering is not only filled into the next iteration of clustering but also used to generate superpixels.
Figure 2. The clustering workflow of context-aware superpixel (CASP). Different colors in label maps denote the adopted cluster labels. We can see the whole yellow background of the outcome of 1st similarity clustering is split into several connected parts by the 1st proximity clustering. The outcome of proximity clustering is not only filled into the next iteration of clustering but also used to generate superpixels.
Entropy 22 00020 g002
Figure 3. After added Rician noise to the ground truth data, we can see in the first row that the image is corrupted by the noise in different levels, denoted by N max . After subtracting the noisy image from the ground truth, we derived residual maps of Rician noise shown in the second row. Because the Rician noise is signal dependent, we can see apparent content textures in the residual noise maps.
Figure 3. After added Rician noise to the ground truth data, we can see in the first row that the image is corrupted by the noise in different levels, denoted by N max . After subtracting the noisy image from the ground truth, we derived residual maps of Rician noise shown in the second row. Because the Rician noise is signal dependent, we can see apparent content textures in the residual noise maps.
Entropy 22 00020 g003
Figure 4. Coarse to fine superpixels on the eagle image. Owing to the hierarchically clustering, CASP can even capture halo in a circular pattern which is neither a salient feature of human vision nor of baseline methods. (a) The ground truth segmentation of the eagle image; (b) superpixel segmentation results of three methods in 6 levels.
Figure 4. Coarse to fine superpixels on the eagle image. Owing to the hierarchically clustering, CASP can even capture halo in a circular pattern which is neither a salient feature of human vision nor of baseline methods. (a) The ground truth segmentation of the eagle image; (b) superpixel segmentation results of three methods in 6 levels.
Entropy 22 00020 g004
Figure 5. Coarse to fine superpixels on the magnetic resonance (MR) brain image. Due to manually specifying the total number of superpixels, simple linear iterative clustering (SLIC) and linear spectral clustering (LSC) evenly generate superpixels in the constant background image region. By contrast, CASP only generates one superpixel for representing background, indicating high-level adaptivity. (a) The ground truth segmentation of a MR slice; (b) superpixel segmentation results of three methods in 6 levels.
Figure 5. Coarse to fine superpixels on the magnetic resonance (MR) brain image. Due to manually specifying the total number of superpixels, simple linear iterative clustering (SLIC) and linear spectral clustering (LSC) evenly generate superpixels in the constant background image region. By contrast, CASP only generates one superpixel for representing background, indicating high-level adaptivity. (a) The ground truth segmentation of a MR slice; (b) superpixel segmentation results of three methods in 6 levels.
Entropy 22 00020 g005
Figure 6. CASP achieves better segmentation performance than baseline methods, with respect to under-segmentation error (a), boundary recall (b), and (c) achievable segmentation accuracy.
Figure 6. CASP achieves better segmentation performance than baseline methods, with respect to under-segmentation error (a), boundary recall (b), and (c) achievable segmentation accuracy.
Entropy 22 00020 g006
Figure 7. Entropy values of synthetic images. Denote a superpixel as one object. The complementary role of H CON and H OCC indicates that uncertainty is transferred from that within objects to that among objects. (a) H BE degenerates to H IM when H CON = 0 and H OCC = H IM . (b) The increasing object amount elevates H BE , while we can find a proper arrangement of pixels may decrease the H BE .
Figure 7. Entropy values of synthetic images. Denote a superpixel as one object. The complementary role of H CON and H OCC indicates that uncertainty is transferred from that within objects to that among objects. (a) H BE degenerates to H IM when H CON = 0 and H OCC = H IM . (b) The increasing object amount elevates H BE , while we can find a proper arrangement of pixels may decrease the H BE .
Entropy 22 00020 g007
Figure 8. CASP derives less entropy values than based line methods. (a) CASP generates superpixels with less H C O N , indicating a higher level of local intensity coherence; (b) Also, H O C C derived from CASP is smaller than other baseline methods; (c) Accordingly, H B E derived from CAPS is the smallest.
Figure 8. CASP derives less entropy values than based line methods. (a) CASP generates superpixels with less H C O N , indicating a higher level of local intensity coherence; (b) Also, H O C C derived from CASP is smaller than other baseline methods; (c) Accordingly, H B E derived from CAPS is the smallest.
Entropy 22 00020 g008
Figure 9. Under image corruption that is caused by no matter slight or severe noise, CASP demonstrates better performance than baseline methods. (Denote noise level N max )
Figure 9. Under image corruption that is caused by no matter slight or severe noise, CASP demonstrates better performance than baseline methods. (Denote noise level N max )
Entropy 22 00020 g009
Figure 10. Pearson’s correlation coefficient between local image coherence, H CON and superpixel segmentation performance w.r.t. U S E , B R , and A S A . U S E has a positive correlation with H CON , while B R and A S A negatively correlate with H CON .
Figure 10. Pearson’s correlation coefficient between local image coherence, H CON and superpixel segmentation performance w.r.t. U S E , B R , and A S A . U S E has a positive correlation with H CON , while B R and A S A negatively correlate with H CON .
Entropy 22 00020 g010

Share and Cite

MDPI and ACS Style

Liu, F.; Zhang, X.; Wang, H.; Feng, J. Context-Aware Superpixel and Bilateral Entropy—Image Coherence Induces Less Entropy. Entropy 2020, 22, 20. https://doi.org/10.3390/e22010020

AMA Style

Liu F, Zhang X, Wang H, Feng J. Context-Aware Superpixel and Bilateral Entropy—Image Coherence Induces Less Entropy. Entropy. 2020; 22(1):20. https://doi.org/10.3390/e22010020

Chicago/Turabian Style

Liu, Feihong, Xiao Zhang, Hongyu Wang, and Jun Feng. 2020. "Context-Aware Superpixel and Bilateral Entropy—Image Coherence Induces Less Entropy" Entropy 22, no. 1: 20. https://doi.org/10.3390/e22010020

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop