Next Article in Journal
Relationship between the Growing Season Maximum Enhanced Vegetation Index and Climatic Factors on the Tibetan Plateau
Next Article in Special Issue
Hierarchical Segmentation Framework for Identifying Natural Vegetation: A Case Study of the Tehachapi Mountains, California
Previous Article in Journal
Predictive Mapping of Dwarf Shrub Vegetation in an Arid High Mountain Ecosystem Using Remote Sensing and Random Forests
Previous Article in Special Issue
Geographic Object-Based Image Analysis Using Optical Satellite Imagery and GIS Data for the Detection of Mining Sites in the Democratic Republic of the Congo
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Efficient Learning of Spatial Patterns with Multi-Scale Conditional Random Fields for Region-Based Classification

by
Mitchel Alioscha-Perez
1 and
Hichem Sahli
1,2,*
1
Electronics and Informatics Department (ETRO), Vrije Universiteit Brussel (VUB), Pleinlaan 2,1050 Brussels, Belgium
2
Interuniveristy Microelectronics Center (IMEC), Kapeldreef 75, BE-3001 Leuven, Belgium
*
Author to whom correspondence should be addressed.
Remote Sens. 2014, 6(8), 6727-6764; https://doi.org/10.3390/rs6086727
Submission received: 19 March 2014 / Revised: 25 June 2014 / Accepted: 30 June 2014 / Published: 24 July 2014
(This article belongs to the Special Issue Advances in Geographic Object-Based Image Analysis (GEOBIA))

Abstract

: Automatic image classification is of major importance for a wide range of applications and is supported by a complex process that usually requires the identification of individual regions and spatial patterns (contextual information) among neighboring regions within images. Hierarchical conditional random fields (CRF) consider both multi-scale and contextual information in a unified discriminative probabilistic framework, yet they suffer from two main drawbacks. On the one hand, their current classification performance still leaves space for improvement, mostly due to the use of very simple or inappropriate pairwise energy expressions to model complex spatial patterns; on the other hand, their training remains complex, particularly for multi-class problems. In this work, we investigated alternative pairwise energy expressions to better account for class transitions and developed an efficient parameters learning strategy for the resultant expression. We propose: (i) a multi-scale CRF model with novel energies that involves information related to the multi-scale image structure; and (ii) an efficient maximum margin parameters learning procedure where the complex learning problem is decomposed into simpler individual multi-class sub-problems. During experiments conducted on several well-known satellite image data sets, the suggested multi-scale CRF exhibited between a 1% and 15% accuracy improvement compared to other works. We also found that, on different multi-scale decompositions, the total number of regions and their average size have a direct impact on the classification results.

1. Introduction

In this work, we consider the following problem: after segmenting an image into several regions, classify each region into one of the predefined classes. We refer to this process as region-based image labeling. The result is both a segmentation of the image and labeling of each region (segment) as a given object class. Automatic (region-based) image labeling is of importance for several applications, such as scene understanding, image data base querying, etc.

In natural scenes, different classes are dependent across similar pixels or regions and also on their co-occurrence. It is not surprising to find regions of certain classes (i.e., river and tree) located next to each other, but it is very unlikely to find some aberrant co-occurrences, such as a building in the middle of a lake. Spatial patterns provide valuable contextual information for achieving consistent labeling. Moreover, since segmentation could provide very small (or too big) ambiguous regions, multi-scale analysis allows one to determine the objects’ class more accurately [1,2] in the most appropriate scale within a range (multi-scale) [3].

One of the popular choices to address image classification, including contextual and multi-scale information, has been Markov random fields (MRF) [4,5]. Recently, a discriminative probabilistic framework, named conditional random fields (CRF) [6,7], has been proposed for the same purpose [1,8,9]. However, despite the undeniable improvements of CRF with respect to MRF, two main issues are still common in most of the CRF state-of-the-art models.

The first issue is that classification results, in several cases, are still very low. We associate this problem with an inappropriate pairwise energy expression that fails to model complex spatial patterns (i.e., class transition detections or co-occurrence patterns). This problem can be alleviated by defining proper features for the energies and by defining the appropriate framework to allow more complex energy expressions.

The work of [10] suggests the use of non-linear (exponential) potentials. Similarly, in this work, we model each potential as an unknown function ψ𝒞 : 𝒴 → ℝ (of unknown distribution) in a set of possible functions ψC, where is a Hilbertian space endorsed with an inner product 〈·, ·〉. We also introduce a set of additional weighting parameters (variables) in order to control the potential combination.

The work of Gould et al. [11] discusses the design of pairwise energies and augments these with relative location information, but their segmentation strategy does not allow the use of multi-scale analysis. The authors in [9] use multi-scale information to express pairwise potentials, despite their local dependencies (intra-scale) involving image structure information (i.e., number of neighbors); their long-range dependencies (multi-scale or inter-scale) only use color distances. We also make use of spectral feature differences in the form of distances, and we go further by including image hierarchical structure (multi-scale) information, such as the scale of analysis, lowest and highest spectral (color) distances to neighboring regions, the number of siblings (the regions sharing the same parent region on the next coarser scale) and the region’s lifetime. We follow a strategy more close to [12], where the region locations are determined by the multi-scale information (inside, tangential, non-tangential, etc.), but based on hierarchical definitions (parents, siblings, neighbors), since our segmentation strategy guarantees a non-overlapping region’s hierarchical decomposition.

The first contribution of this work is a general energy expression that considers:

  • Possible non-linear potential functions ψc(·;θ) = 〈θ, fФ (inner product) that differ from the usual ψc(·;θ) = θT f (dot product), where f is a feature vector;

  • A weighted mixture of potentials EW (·;θ, d) = ∑c dc ψc(·;θ), dc ≥ 0;

  • Simpler (in number of features), yet rich and expressive pairwise energies involving contextual and multi-scale information.

The second issue, present in most CRF models, is related to the complexity of the parameter learning stage. The conditional maximum likelihood estimation (CMLE) approach requires approximations to estimate the partition function [8,13,14] and achieves relatively simple and accurate parameter learning using gradient descendant optimization methods [15]. Maximum margin confidence approaches do not involve the partition function on their formulation, but poses a challenging quadratic program (QP) for the parameter learning, usually solved with (online) sub-gradient [10,16] or cutting-plane methods [1719].

When using any of them in the context of CRF parameter learning, another challenge is related to the necessity of solving an inference problem at each iteration during the optimization loop, which increases the computational burden.

Other training strategies, such as the piecewise training framework [20] and its variants [21], proposed simpler formulations to the parameter learning problem and have been shown to be very competitive [19,21]. They have been extensively used in the context of CMLE [11,19,2123].

An example is the work in [24], where the authors start from a decomposition on the dual formulation of a max-margin QP [25] and introduce a piecewise pseudo-likelihood (PWPL) estimator [21] in order to reduce the (per iteration) inference complexity during training.

In this paper, we consider a particular set of values for the energy weighting coefficients (dc), where the final energy form involves only single sites and their corresponding pairwise factors, while in [24], the single-site factors of neighboring sites are involved, too. Then, in order to learn the parameters, we decoupled the associated QP into simpler individual multi-class sub-problems that can be efficiently solved dually using the Crammer and Singer fixed-point method [26].

The second contribution of this work is a more general energy expression considering contextual and multi-scale information, in a max-margin parameter learning approach within the piecewise framework, which provides a novel strategy to tackle the resultant parameter learning problem. The proposed energy expression connects to and generalizes other works, reaching the same expression as in [22,24] for different weighting schemes (values for the weights).

The remainder is organized as follows. The multi-scale CRF formulation is provided in Section 2. The proposed training strategy is detailed in Section 3, while the inference process is described in Section 4. Experimental results, using remote sensing images, are reported and discussed in Section 5, and conclusions are given in Section 6.

2. Multi-Scale Conditional Random Fields for Image Classification

The CRF formulation model directly the labels conditioned on observations, decomposing the energy expression according to the different clique order. The clique decomposition theorem defines the overall form of a general CRF model as follows:

P ( x | y , θ ) = 1 Z ( y , θ ) exp ( c 𝒞 ψ c ( x c , y , θ ) )
where Z(y,θ) is a normalization function named the partition function.

The above model is known to follow an exponential distribution [4], while the potentials ψc are functions defined over cliques c𝒞, usually assumed to be linear and expressed as an inner product between features y and model parameters θ [1,7,8,22]; xc are the labels limited to factors of order c𝒞.

This general CRF model can be specified for (defined on) the multi-scale region-based image labeling context. This includes an appropriate potential definition, features for the energy expression, among others. For the region-based image labeling, there are some other stages involved in the overall process, such as the image pre-processing (filtering), segmentation, parameter learning and classification. The general framework is illustrated in Figure 1.

2.1. PDE-Based Smoothing

In order to avoid blurring and the delocalization of the image features, an image adaptive scale-space filter is used. In this work, we opted for a method that guides the filtering process in such a way that intra-region smoothing is preferred over inter-region smoothing and edges are gradually enhanced. The employed filter [27], belongs to the class of nonlinear anisotropic diffusion filters. It is a combination of the Catté et al. [28] regularized Perona and Malik filter [29].

Let I = {I(1), I(2), …, I()} be a vector-valued image defined on a finite domain Ω. The multiscale tower (u) of I is governed by the following system of coupled parabolic partial differential equations (PDEs):

{ t u ( r ) = div [ g ( | u σ | ) u ( r ) | u ( r ) | ] r = 1 , 2 , , u ( t = 0 ) = I n u = 0 on δ Ω
where u(r) represents the rth image band, t is the continuous scale parameter, δΩ is the image boundary (with n denoting the normal direction to it) and σ is a regularization parameter, which ensures the well-posedness of the above system; g is the Lorentzian edge stopping function [30], which is formulated as:
g ( | u σ | ) = 1 1 + | u σ | 2 k 2

The so-called contrast parameter k in Equation (3) separates backward from forward diffusion and is estimated using the cumulative histogram of the regularized gradient magnitude (|∇uσ|) [27]. In this work, a discrete version of the multiscale tower u, denoted as U = {u0,, un, , uN }, is obtained by applying the natural scale-space sampling method [31]. From the scale space tower, we select a scale, referred to as the localization scale, at which the diffusion filter has removed the noise without affecting or dislocating the edges of the salient features in the image, and we apply to it the multi-scale region adjacency graph (MSRAG) generation process described in Section 2.2. The localization scale is determined using a maximum correlation criterion proposed in [32].

The maximum correlation criterion selects the scale that maximize the correlation between the original image and the diffused image:

s 0 = arg max t j [ 𝒞 m ρ ( t j ) ] arg max t j [ σ [ u ( t ) ] σ [ n out ( 0 ) ] σ [ n ( 0 ) ] σ [ n out ( t ) ] ]
where σ[u(t)] and σ [u(0)] are the standard deviation of both the diffused image at the t − th scale and the noisy image, respectively, and σ [nout(t)] and σ [nout(0)] are the standard deviation of the outlier noise of both the diffused image at the t − th scale and the noisy image, respectively. The values of σ [nout(t)] and σ [nout(0)] are estimated by using the median absolute deviation of the gradient of the diffused and original images (see [33] for details).

2.2. Multi-Scale Region Adjacency Graph (MSRAG)

In this work, the image classification is treated as a hierarchical labeling problem applied to the nodes S of a multi-scale region adjacency graph (MSRAG). The MSRAG represents a nested hierarchy of partitions, 𝒮 n = { r 1 n , r 2 n , , r m n n }; n = 1, …, N ( r j n being disjoint regions), which preserves the inclusion relationship 𝒮n𝒮n−1, implying that each atom of the set 𝒮n is a disjoint union of atoms from the set 𝒮n−1. In other words, a partition at a coarse level is obtained by merging regions of the finer partition. The MSRAG is defined as a graph 𝒢 = (S, E), with the set of nodes S being the partitions at each scale, S = S1S2 ∪· · · ∪SN, and the edges E are the set of common boundaries between the regions at each scale (intra-scale) and the inter-scale (parent-child) links; see Figure 2.

Hierarchical feature representation through multi-scale segmentation offers new possibilities in object-oriented and multi-scale image analysis [34]. State-of-the art, object-based and object-oriented segmentation algorithms are often based on region growing and merging, or linear and non-linear scale-space.

In the proposed framework, starting from the localization scale s0 (obtained using the PDE of Section 2.1), we follow the approach of [35], where the input multi-spectral image, u0, is first segmented using the watershed transform. Then, the waterfall algorithm [36] is used for producing a nested hierarchy of partitions: the multi-scale region adjacency graph (MSRAG); as illustrated in Figure 2. Such a hierarchy preserves the topology of the initial watershed lines and extracts homogeneous objects of a larger scale. The waterfall algorithm removes from the current partition (hierarchical level) all of the boundaries completely surrounded by higher boundaries. Thus, the saliency of a boundary is measured with respect to its neighborhood. The iteration of the waterfall algorithm ends with a partition of only one region. To apply the watershed, the gradient of the multi-spectral image is obtained by combining, using the approach of [37], the gradients of the texture (orientations) and the DiZenzo [38] multi-spectral gradient. For producing the nested hierarchy, we use the approach proposed in [35], where the saliency measure of a boundary is based on a collection of energy functions used to characterize desired single-segment properties and pair-wise segment properties. The single segment properties includes area, convexity, compactness and color variances within the segment. The pair-wise properties includes color mean differences between two segments and edge strength. In [35], the saliency measure, E( = rirj|ri, rj), of a boundary between two neighboring segments ri and rj (being the cost of merging the regions ri and rj), has been defined as:

E ( r ˜ = r i r j | r i , r j ) = E ( r ˜ ) + E ( r i , r j )
The single segment properties, E(), is expressed as:
E ( r ˜ ) = E area ( r ) 1 E hom ( r ˜ ) c E var c ( r ˜ ) . ( 1 + | E conv ( r ˜ ) | ) sign ( E conv ( r ˜ ) ) . ( 1 + | E comp ( r ˜ ) | ) sign ( E comp ( r ˜ ) )
and the pair-wise property as:
E ( r i , r j ) = E edge ( r i , r j ) E CMDif ( r i , r j )

The different energies associated with segment properties are given in Appendix (C), and additional details can be found in [35].

2.3. Hierarchical Labeling for Region Classification

The labeling is performed using a finite set = {1, …, L} of interpretation labels. We consider a couple of random fields (X, Y) on 𝒢, with X = {Xn, ∀n ∈ [1, …, N]} and X n = { X s n , 𝒮 S n } being the label field and Y = {Yn, ∀n ∈ [1, …, N]} and Y n = { Y s n , 𝒮 S n } the observations field. In our current implementation Y s n represents the region properties at certain site s and scale n. A similar notation holds for the realizations: x = { x s n , n [ 1 , , N ] s S n }, x s n and y = { y s n , n [ 1 , , N ] 𝒮 S n }, y s n p × , with p the number of pixels in the region r s n and the number of spectral bands of the image.

We define the general form of the proposed multi-scale CRF for determining the optimal labeling as:

P ( x | y , θ ) = 1 Z ( y , θ ) exp ( n = 1 N s S n d 1 ψ A ( x s n , y , ω ) + + n = 1 N 1 s S n d 2 ψ L ( x s n , x s ¯ n + 1 , y , β L ) + n = 1 N i S n j η ( i ) d 3 ψ I ( x i n , x j n , y , β I )
with the parent of a node s, η (i) the set of neighbors of site i at the considered level and the parameters θ = (ω, βI,βL)T defined in order to differentiate intra-scale and inter-scale spatial patterns; the weighting parameters dm ∈ ℝ+ regulate each potential individual contribution to the overall energy expression. The pairwise potentials ψI(·; β I) and ψL(·; βL) encode contextual information between neighbors within the same scale and between parent/child regions at any two consecutive scales, respectively; while ψA(·;ω) is the association potential, which provides local evidence.

The considered model interactions are approximately represented in Figure 3. In order to make the representation simpler to follow, we illustrate only parent-child and four neighbors, while in real segmentation, they usually have any amount of neighbors. This structure falls within a research line of higher order potential design [39,40].

The partition function Z(y,θ) is defined as the sum of per-clique model energy for every possible realization x, Z(y, θ) = ∑xΠc𝒞 ψc(xc,y, θ), where 𝒞 account for up to second order cliques. The number of possible values combinations for x is equal to the number of different labels exponential to the number of sites (|||S|); due to this, the exact computation of Z(y,θ) becomes very easily intractable, even for small graphs with a low number of labels.

For the labeling, we consider inter-scale interactions directionally from parents to child, aiming to collect classification results at the finest scale (see Figure 3), which provides the final classification map.

2.4. Potential Definition

The different potentials are defined as follows:

ψ A ( x s , y , ω ) = l 1 ( x s = l ) ω ( l ; f y )
ψ L ( x s , x s ¯ , y , β L ) = l , k 1 ( x s = l ) 1 ( x s ¯ = k ) 1 ( l = k ) β L ( l , k ; h y L )
ψ I ( x s , x j , y , β I ) = l , k 1 ( x i = l ) 1 ( x j = k ) 1 ( l = k ) β I ( l , k ; h y I )
where fy is a unary feature vector, while h y L and h y I are pairwise feature vectors. The function 1(a = b) returns one if the condition (a = b) is met and zero otherwise. The separation among the different classes is determined as the following inner product:
ω ( l ; f y ) = ω l , f y , Φ 𝒣
β L ( l , k ; h y L ) = β 1 ( l = k ) L , h y , Φ L 𝒣
β I ( l , k ; h y I ) = β 1 ( l = k ) I , h y , Φ I 𝒣
being wl a total of l unknown parameters (l), while β 1 ( l = k ) L and β 1 ( l = k ) I are two and two unknown parameters, respectively (four in total with 1(l = k) ∈ {0, 1}).

The (mapped) features fy, h y , Φ L and h y , Φ l are the feature vectors (fy, h y L and h y I) mapped by a function Ф : 𝒴. In practice, neither the mapping function Ф (·) nor the space needs to be explicitly defined; it is sufficient to define a symmetric and positive definite kernel function K : 𝒴 × 𝒴 → ℝ, to inherit both the hypothesis space as its unique reproducing kernel Hilbert space (the Moore–Aronszajn theorem [41]) and the mapping function Ф(·) = K(y, ·).

Note that the distribution of the potentials ψc(·;θ) is assumed to be unknown. For the particular case of using linear kernel functions, our work will recover the usual linear potential functions, and for any other kernel functions, we will recover more complex potential functions. A bias value can be easily introduced in Equations (12)(14) (if wanted or necessary) by making use of algebraic methods.

In this work, we consider class transition detections for pairwise potential modeling {1(xa = l) 1 (xb = k)1(l = k), ∀l, k×}, which can be seen as two binary models: a binary model (transition occur: yes or no) for the intra-scale potential and a binary model (correct region merging: yes or no) for the inter-scale potential.

However, the model has been designed for a broader multi-class pairwise potential definition {1(xa = l) 1 (xb = k), ∀l, k × } as two multi-class co-occurrence models: a multi-class model of L × L classes (the transition of class a to b occurs: yes or no?) for the intra-scale potential and a multi-class model of L × L classes (correct region merging of class a to b: yes or no) for the inter-scale potential.

2.4.1. Unary Local Features

The association potential, ψA(·;ω), is defined as the probability of a site to acquire a certain label considering the site observations independently. The single-site observation feature vector fy is obtained as a concatenation of different features making use of an observation y. The following features have been considered in the current implementation:

  • Spectral: the spectral signature of a region y is characterized for each channel by its mean, standard deviation, kurtosis and skewness.

  • Textural: the texture of a region is characterized using the local binary pattern [42], from which we compute the region’s mean value, repeating the process using eight different radius (neighborhood) values.

  • Morphological: for the shape information, we compute the following morphological features: elongation, area to length ratio and extent.

  • Scale: the actual scale of the region in question.

Considering a color RGB image, the resultant feature vectors will contain 24 components.

2.4.2. Inter-Scale Pairwise Features

The inter-scale potential,ψL(·; βL), aims at modeling the probability distribution of a good region merging having taken place (detecting errors on segmentation) during the segmentation process. This potential function is based on the observations of the region in question (named the actual region) and other additional information, such as:

  • its parent region: a good merging is supposed to keep spectral homogeneity between the actual region and its parent;

  • all of its siblings: a region that is merged with several siblings is supposed to keep spectral homogeneity with each one of them;

  • the number of siblings: the greater the number of regions that are merged (the number of siblings), the more likely it is for a bad merging to occur;

  • the scale of the actual region: in hierarchical segmentation, higher levels are more likely to exhibit regions that combine multiple objects (bad merging);

  • the lifetime of the actual region: whenever a region exhibits a high lifetime (strong edges), it is because it can be well differentiated (in terms of spectral homogeneity) from its neighboring regions; therefore, the merging of regions with a high lifetime implies a bad merging detection.

For the inter-scale pairwise energy expression, we combine all of the previous information in a five-dimensional vector as follows:
h y L = [ D ( y s , y s ¯ ) , j ς ( s ) D ( y s , y j ) , | ς ( s ) | , n , lifetime ( s ) ] T
ϛ(s) being all of the siblings (which have a parent region in common) of the region in question; lifetime(s) is the number of scales at which a region remains the same (without being merged) according to the MSRAG structure; n denotes the level of the actual region in the MSRAG; |ϛ(s)| denotes the number of siblings of the region in question; and D(·, ·) is the Bhattacharyya distance between any two regions’ histogram.

2.4.3. Intra-Scale Pairwise Features

The intra-scale potential, ψI(·; β I), aims at modeling the probability distribution of two regions sharing the same label. This potential function is based on the observations of the two regions in question, but also benefits from additional information, such as:

  • the corresponding neighbor: a region that shares the same label with a neighbor is supposed to keep spectral homogeneity;

  • lowest neighbor distance: the lowest spectral distance from all neighbors provides a measurement of how homogeneous the surrounding area that contains the region in question is;

  • highest neighbor distance: the highest spectral distance from all neighbors provides a measurement of how homogeneous the surrounding area that contains the region in question is;

  • the scale of the actual region: in hierarchical segmentation, higher levels are more likely to exhibit regions that combine multiple objects (bad merging).

For the intra-scale pairwise energy expression, we concatenate all of the previous information in a four-dimensional vector as follows:

h y I = [ D ( y s , y s ¯ ) , min j η ( s ) D ( y s , y j ) , min j η ( s ) D ( y s , y j ) , n ] T
n being the actual processing scale of site s in the MSRAG and D(·, ·) the Bhattacharyya distance between the two regions’ histogram.

3. Parameter Learning

3.1. Piecewise Training Framework

The piecewise training framework has been proposed by [20] in order to reduce the complexity of the partition function estimation in CMLE approaches, consisting of dividing the full model into pieces 𝒫𝒯, according to a set of piece types 𝒯 [21]. An illustration of the standard piecewise framework (𝒫 = 𝒞2) can be seen in Figure 4b.

From the partitions 𝒫, a new family of upper bounds log Z𝒫 (y, θ) can be used to provide a simpler (tractable) estimator [20] for the minimization of the log-partition function log Z(y, θ) ∝ ∑𝒫𝒯log Z𝒫 (y, θ). This allows one to re-write the log-likelihood of Equation (1) as follows:

log P ( x | y , θ ) = log 1 Z ( y , θ ) exp ( c 𝒞 ψ c ( x c , y , θ ) ) = 𝒫 𝒯 c 𝒫 ψ c ( x c , y , θ ) 𝒫 𝒯 log Z 𝒫 ( y , θ ) = 𝒫 𝒯 ( c 𝒫 ψ c ( x c , y , θ ) log Z 𝒫 ( y , θ ) )
where ψ c(·; θ) consists of the per clique and per piece-type energy given θ and Z𝒫(y, θ) the local partition function estimator per model piece-type 𝒫 [21]. Based on Equation (18), gradient descendant techniques provide a simple way to learn θ .

3.2. Max-Margin Parameter Learning

Within the max-margin parameter learning framework, the problem is posed differently. The parameter learning problem aims to find a θ, such that, for the observed labels and observations y in training samples, P ( | y, θ) has to be the highest possible, among all possible values of x:

P ( x ^ | y , θ ) P ( x | y , θ ) x x ^

This will guarantee that for any possible value x, arg maxx P (x | y, θ) will attain the maximum when x = . Considering the clique decomposition theorem (Equation (1)), the previous inequality takes the following form:

1 Z ( y , θ ) exp ( 𝒫 𝒯 𝒫 P ψ c ( x ^ c , y , θ ) ) 1 Z ( y , θ ) exp ( 𝒫 𝒯 𝒫 P ψ c ( x c , y , θ ) )
therefore:
𝒫 𝒯 c 𝒫 ψ c ( x ^ c , y , θ ) 𝒫 𝒯 c 𝒫 ψ c ( x c , y , θ ) x x ^
where the partition Z(y, θ) function is no longer necessary to be computed.

In the max-margin learning approach, we want that the previous in-equivalence Equation (20) not only holds, but also has the maximum margin. Assuming the following notations:

E 𝒫 = c 𝒫 ψ c ( x c , y , θ )
E ^ 𝒫 = c 𝒫 ψ c ( x ^ c , y , θ )
and introducing a generic loss function (x, ), the parameter learning can be formulated as the following optimization problem:
max λ > 0 , θ λ s . t . 𝒫 𝒯 E 𝒫 𝒫 𝒯 E ^ 𝒫 λ ( x , x ^ ) θ 2 1 x x ^
which is similar to the max-margin parameter learning problems posed in [18] without considering noisy data (hard margin). The application of cutting-planes or sub-gradient methods are a common choice to solve this problem, requiring a full inference on every iteration step [16,18]. This is the first drawback associated with the max-margin parameter learning. The second drawback is related to the number of constraints included by x in Equation (23), which is an exponential value in the number of sites equal to |||S| − 1.

3.3. Proposed Efficient Max-Margin Parameter Learning

In our proposed learning strategy, we avoided using sub-gradient and cutting-plane methods by decomposing the original problem into several individual multi-class sub-problems that can be efficiently solved dually, and we significantly reduced the number of constraints included by x. Starting from Equation (23), a usual assumption [43] is that, likewise the per-clique energy functions ψc(·;θ), the loss function (x, ) is decomposable according to the different cliques’ order. Using the notation 𝒫 = 𝒫(x𝒫, 𝒫), we can state that:

( x , x ^ ) = 𝒫 𝒯 𝒫 ( x 𝒫 , x ^ 𝒫 )
= 𝒫 𝒯 𝒫

Then, for any model partition (as simple or as complex as wanted) under 𝒯, we propose to decouple our formulation of Equation (23), decomposing it into the following set of individual multi-class sub-problems (see Appendix A):

𝒫 𝒯 { min θ 𝒫 1 2 θ 𝒫 2 2 + C i ξ i s . t . E 𝒫 E ^ 𝒫 𝒫 ξ i , i = 1 N 𝒫 x i Ω 𝒫
where Ω𝒫 is the set of label combinations for a 𝒫-order clique (see Appendix A) and N𝒫 the total number of related factors [21]; xi ∈ Ω𝒫 represents the i-th factor (i = {1, …, N𝒫}) possible labels, ξi are slack variables and i along with yi the corresponding (i-th) ground-truth and observations, respectively, present in the training samples.

Note that, since 𝒯 defines the relations of each site with its neighborhood in a Markov blanket (the set of its parent, children and brothers), it will determine the number of constraints per each i-th site, as well as the energy form related to each individual site i. Since they have a direct impact on the model complexity, simple model partitions, such as piecewise (PW) or piecewise-pseudolikelihood (PWPL), are often preferred, but any structure can be defined for 𝒯 .

The per-site energy expression, usually defined as E(·; θ) = ∑ψc(·; θ), can take different forms and vary according to the site’s relations with its neighbors in the Markov blanket. It is known that such simple energy expressions are sometimes inappropriate, and the use of extra weighting terms is necessary to adjust the potential combination [9,18,22,39].

We propose a more general formulation EW (·; θ, d) for the per-site energy expression, which involves a weighted mixture of potential functions, resulting in a weighted combination of local, contextual and multi-scale information:

E W ( x i , y ; θ , d ) = d 1 ψ A ( x i , y , θ ) + neighbour | j η ( i ) j C ( d 4 ψ A I ( x j , y , θ ) + d 2 ψ I ( x j , x i , y , θ ) ) + parent / child | j η ( i ) j C ( d 5 ψ A L ( x j , y , θ ) + d 3 ψ L ( x j , x i , y , θ ) )

In order to establish connections with other energy expressions, we defined the model structure 𝒢 according to some MSRAG and considered, without loss of generality, some model partition, such as the PWPL proposed in [21] and assumed in [24].

As can be seen, ψ A I(·) and ψ A L(·) are unknown potential functions involving single factors of neighbors within the same scale and parent/child scales, respectively. The weighting terms dm, m = 1, , 5 control the mixture of the different potential information.

The proposed energy expression, EW, generalizes the one used in the works of Alahari et al. [24], referred to as EPWPL. Note that assuming a fixed set of weights d = (1, 1, 0, 1, 0), both energy expressions EW and EPWPL become equivalents: EWEPWPL. It also generalizes the one used in [22] within the context of CMLE and referred to as EPW, since, for the particular case of d = (1, 1, 0, 0, 0), both energy expressions EW and EPW become equivalents: EWEPW .

Then, we introduce the weighted potential mixture into our learning strategy to obtain the following set of QP problems:

𝒫 𝒯 { min θ 𝒫 1 2 θ 𝒫 2 2 + C i ξ i s . t . E 𝒫 W E ^ 𝒫 W 𝒫 ξ i , i = 1 N 𝒫 , x i Ω 𝒫
where:
E 𝒫 W ( x i , y , θ ) = c 𝒫 d c ψ c ( x i , y , θ )

We considered values for d, such that m = 1 5 d m + with d4 = d5 = 0 and solved the parameter learning in a max-margin approach, where each resultant problem of Equation (28) takes the form of a standard multiclass support vector machine (SVM) model, very similar to the work of Crammer and Singer [26], that can be efficiently solved dually using the fixed-point method.

Similar to [39] for max-margin and to [22] for CMLE, we assume a fixed value for d during the training stage, and we teach them afterword on a validation set.

Since each individual problem is solved dually, the obtained set of optimal dual variables (α*-coefficients) are related, under a strong duality assumption (see [25,26]), to the primal variables (ω, β L and β I) by the Karush–Khun–Tucker (KKT) conditions. Therefore, the optimal ωl is the following:

l ( ω l ) * = i | x ^ i = l α i l * y i
where α i , l * are the (already optimal) ith dual variables (also known as support vectors) for the class l and yi the observation of the i − th sample; the same for β I, β L. For more details, we refer to [26].

The following algorithm details the proposed training strategy:

1:function LearnParams(y, , 𝒢, 𝒯)
2:  for (𝒫𝒯 do
3:    y𝒫←StructuredObservations(y, 𝒢, 𝒫)▹ Defined y n s in Section 2.3
4:    𝒫←StructuredSamples(, 𝒢, 𝒫)  ▹ Detailed in Section 3.4
5:    α𝒫MulticlassSVM(y𝒫, 𝒫)  ▹ Dual of Equation (28)
6:    ω𝒫←ToPrimal (α𝒫, y𝒫, 𝒫)   ▹ Using Equation (30)
7:  end for
8:  θ*=(ω𝒫1,…w𝒫N)T, N = |𝒯|
9:  return θ*
10:end function

3.4. Structuring the Training Samples According to the MSRAG

Regarding the observations, since the training data is a limited set of labeled pixels (regions) from the original image, it is structureless and has no related hierarchical information. Therefore, in our model, they are represented by the observations on the first (finest) scale y s 1 only.

In order to obtain the training samples at the other (higher) scales, we propagate this information to the next coarser scales (see Figure 5), considering that a parent region will have the same label as its child, only if all of the children share the same label. When children have different labels, we select the most represented class, by counting the number of pixels related to each class and taking the highest one:

arg max l i | i ¯ = s | y i n | 1 x ^ i n = l
1 x ^ i n = l = 1 if the region r s i n has been labeled with a label l on the ground-truth or zero otherwise, and | y i n | is the number of pixels inside the region in question. Then, a region at a higher scale will inherit this class according to the following condition:
x ^ s n + 1 = { l if 1 i | i ¯ = s | y i n | 1 x ^ i n = l | y s n + 1 | th 0 otherwise

The above condition guarantees (for the training stage only) that a parent region will receive the same label of the class most represented on its children regions, only if the total number of known pixels on its children covers more than a certain percent of the parent region. Such a percent is given by the threshold th; otherwise, it will remain as an unknown (non-labeled, represented by x s n + 1) region, and therefore, it will not be part of the training sample set. A similar threshold has been used in [39] for the same purpose. We usually keep this value as th = 1 L, where L is the total number of available classes.

4. Inference

Once the optimal parameters θ* haven been learned, we can use them to label new samples, by finding the most likely label configuration given the observations and parameters: arg maxxX P (x | y, θ*), where:

P ( x | y , θ * ) = exp ( n = 1 N s S n d 1 ψ A ( x s n , y , ω * ) + + n = 1 N 1 s S n d 2 ψ L ( x s n , x s ¯ n + 1 , y , β L * ) + n = 1 N i S n j η ( i ) d 3 ψ I ( x i n , x j n , y , β I * )

The inference process can be performed using several methods in the case of multiclass problems structured as a general graph with loops. Among them, we can find: message passing algorithms, such as loopy believe propagation (LBP) [44] or tree re-weighted LBP (TR-LBP) [45], and move making algorithms (graph-cut), such as α/βswap [46] or α-expansion [46].

We selected the LBP method for the inference process, which approximate the marginalized posterior by a value named believes, computed using a message passing algorithm [44].

Initially, each site belief is equal to the association potential. Then, the neighbors and parent of each region start to iteratively influence the site belief value, until all of the belief values settle, providing the final belief value as an approximation of the marginalized posterior.

Note that different classification maps are obtained from the inference process on different scales, and coarser scale results have an influence on finer classification maps during inference (see Figure 6). This influence consists in the enforcement of long-range dependencies from coarser scales to finer scales.

The convergence properties of LBP have been associated with the properties of Bethe free energy and converge to stationary points of this free energy approximation; see [44] for further details. Recalling Equation (33), it might be possible that the terms of each independent potential are correlated. This can lead to over-counting problems during the inference. Once the parameter θ* has been found for a fixed weighting scheme, likewise in [9,18,22,39], we learn the weights d1, d2, d3 via cross-validation.

5. Experimental Results and Discussion

For the evaluation of the proposed model, we selected two well-known remote sensing datasets, namely the Pavia City Center and Pavia University (both images available online at http://www.ehu.es/ccwintco/index.php/Hyperspectral Remote Sensing Scenes), and a high resolution aerial RGB image. The Pavia images are used to compare the proposed method to state-of-art ones, and the high resolution image is used to analyze the effectiveness of our multi-scale CRF approach.

5.1. Pavia University Dataset

For the case of the Pavia University dataset (Figure 7a), both the training samples (Figure 7b) and the ground-truth (Figure 7c) were available and used in the work of Aksoy [47], where a pixel-based and a region-based classification method have been compared. The details of the training sample pixel distribution can be found in [47].

The segmentation scheme described in Section 2.2 has been applied to Pavia University using three spectral bands (Bands 68, 30 and 2), out of the 103 available bands. We further used the five first obtained scales in order to define the different potentials, learn the parameters and perform the classification. It has to be noted that state-of-the-art hierarchical methods use a different number of scales: two in [39], three in [40] and four in [1]; however, we found that scales higher than five usually do not provide complementary information, due to the limited ground-truth.

Results using the the proposed multi-scale CRF model, the pixel-level Bayesian (P-L-B) model of [47] and the region-based Bayesian (R-B-B) model of [47] are given in Tables 1 and 2. It can be observed that the proposed approach outperforms both the previous results, as well as other region-based approaches, despite the fact that our proposed model was trained using only three spectral bands, out of the 103 available bands used in [47].

5.2. Pavia Center Dataset

In the case of the Pavia Center dataset, the ground-truth was available, but not the training samples. As consequence, we decided to take 10% of the ground-truth as the training samples. We used a random uniform distribution for determining: (1) squares of a random size; and (2) to locate these squares in random positions within the image. During the random training selection, we also granted that all of the classes were roughly represented in a balanced way, resulting in 10% of each class selection conforming to 10% of the ground-truth. The randomly obtained training samples are detailed in Table 3 and are illustrated in Figure 8b. As such, compared to [47], we use less training pixels for the following five out of the considered nine classes: trees, meadows, bricks, bare soil and shadow.

Results using the the proposed multi-scale CRF model, the pixel-level Bayesian (P-L-B) model of [47] and the region-based Bayesian (R-B-B) model of [47] are given in Table 3.

From Tables 2 and 3, it is apparent and confirmed according to the measured accuracy considering the available ground-truth that the multi-scale CRF-based strategy is generally more effective for classification compared to state-of-art Bayesian (region and pixel) approaches.

5.3. Airborne RGB Color Image

In order to asses the effectiveness of the proposed multi-scale CRF-based classification, we considered an areal color (RGB) image, as illustrated in Figure 9a, with nine classes, namely: shadow, dark roads and paths, light roads and paths, grass land, blue rooftop, red rooftop, green rooftop and arid land. We conducted an assessment process consisting of: (i) evaluating the importance of the MSRAG structure, i.e., hierarchical segmentation; and (ii) analyzing the discriminative power of the different models along with the impact of the potentials on the result. For all of the experiments, we used the same training data of Figure 9a.

5.3.1. Segmentation and MSRAG Generation Analysis

The proposed multi-scale CRF-based classification can be applied to any hierarchical decomposition of the image. In this section, we evaluate the classification results using different hierarchical decomposition approaches. Namely, the proposed MSRAG approach of Section 2.2 and a scale-space-based hierarchy, denoted MRAT (multiscale region adjacency tree), as described in [3].

The creation of the MRAT is based on the creation of the multi-scale tower using the PDE approach of Section 2.1. Then, at a manually selected localization scale, u0, a gradient watershed transformation is performed to detect a set of regions with well-localized contours. To create the hierarchy, the regions at the localization scale are tracked across the multi-scale tower, and a parent-child linking [27] process is made. This is achieved by spatially projecting the regional minima at each scale into the coarser one. Using the duality between the regional minima of the gradient and the catchment basins of the watershed, each regional minimum residing in a given scale is linked with a regional minimum in the coarser scale, if and only if its projection falls in the catchment basin of that regional minimum (for more details, refer to [27]). It has to be noted that, in [3], the MRAT has the form of a truncated tree, with each node (apart from the ones at the highest scale) having a unique predecessor (its parent) and not considering intra-scale links (edges denoting the region adjacency at a given scale).

An illustration of the different hierarchical segmentations is provided in Figures 10 and 11 along with the respective classification results. The multi-scale CRF classification applied to the considered multi-scale region adjacency graphs can be compared under visual inspection in Figure 12, and their per-class accuracy is detailed in Table 4.

In our experiments, we considered the following multi-scale region adjacency graphs:

  • A-MRAT (augmented MRAT): this being the MRAT produced using the approach of [3], where the localization scale was selected manually, and we added the intra-scale edges/links to conform to our definition of MSRAG (see Figure 2);

  • MSRAG A–D: four multi-scale region adjacency graphs obtained starting from the same localization scale used in the A-MRAT and applying the hierarchical segmentation scheme of Section 2.2 with an extra step of small regions merging before the waterfall process, taking into account an expected granularity during the waterfall, with different parameters;

  • MSRAG E: this multi-scale region adjacency graph has been obtained by applying (on the original image) the PDE filtering and localization scale selection of Section 2.1 followed by the hierarchical segmentation scheme of Section 2.2.

The resultant classification accuracy is given in Table 5, along with the regions and average region size, per scale and per multi-scale region adjacency graph.

As can be seen, multi-scale region adjacency graphs with a lesser number of regions seem to be better for the proposed multi-scale CRF model, as far as the reduction of the number of regions not incurring the addition of excessive region merging errors on the initial scales. Additionally, there seems to be an influence of the average region’s size over the accuracy of the proposed multi- scale CRF model classification results.

It is worth noting that the optimal region’s size depends on the size of objects present in the image and naturally varies from one image to another. Moreover, to better characterize the spectral, textural and morphological features of a region, the larger the region is, the better the characterization is. In summary, the experimental results suggest a direct influence of the total number of regions and their average size on the classification results. The obtained results also suggest that the proposed segmentation scheme using the PDE smoothing and localization scale selection of Section 2.1 and the multi-scale segmentation scheme of Section 2.2 produces satisfactory results.

5.3.2. Different Classification Models

For the valuation of the discriminative power of the proposed multi-scale CRF potentials, which involves both intra-scale and inter-scale potentials trained using the above-described max-margin training strategy, we compared it with: (i) a region-based support vector machine (SVM) model; (ii) the MRF classification model of [3]; and (iii) a CRF model using standard potentials.

Classification accuracy results are detailed in Table 6, using the ground-truth of Figure 13b. The classification maps of the MRF and the proposed CRF, using the A-MRAT hierarchy, are illustrated in Figure 13c,d, respectively.

The SVM is a discriminative unstructured model, which has been widely used in remote sensing applications [48]. The related SVM classification is parameterized by a penalty factor C and the kernel parameters. For our tests, we used the LibSVM (available at http://www.csie.ntu.edu.tw/cjlin/libsvm/) package, a linear kernel (K(x, x) = 〈x, x〉) and a hyper-parameter C = 0.1. Both training and classification were performed over the finest scale of the hierarchy; each region was described by its color, texture and shape features.

For the MRF, we used the implementation of the multiscale region-based classification of Katartzis et al. [3], where the authors proposed a causal Markovian model, defined on the hierarchy of a multiscale region adjacency tree (MRAT). As mentioned above, their proposed MRAT graph has the form of a truncated tree, with each node (apart from the ones at the highest scale) having a unique predecessor (its parent). It has to be noted that this model does not consider intra-scale information. In [3], the observation model is a unary region property that represents the spectral signature of the region at the original image. The parameters were learned using a maximum likelihood approach (using expectation-maximization) and using an efficient exact inference procedure thanks to the tree-like structure of the model.

As standard CRF, we used the Undirected Graphical Models (UGM) matlab code [49]. The model is based on Equation (1) using linear unary and pairwise energies E(·; θ) = θT f. We adapted the implementation to the proposed multi-scale structure and used the max-margin parameters learning strategy of Section 3.3, however, keeping the original standard energies.

As can be seen from Figure 13, the MRF model of [3] does not exhibit the salt and pepper effect. This is due to the use of the region-based approach and the long-range dependencies enforced using the multi-scale information. However, it is unable to properly enforce spatial consistency, as it does not model intra-scale dependencies.

From Table 6, the higher performance of the proposed multi-scale CRF model compared to the MRF of [3] has to be related to: (i) the CRF’s discriminative nature [7] compared to the generative nature [6,7] of MRF, which is also confirmed when comparing to the SVM results; (ii) the design of the unary and pairwise potentials; and (iii) the type of MSARG.

The highest per-class performance has been underlined in Table 6. It is worth noting that, using the MRAT as defined in [3], MRF is capable of achieving slightly better performance in one of the eight classes compared to the CRF models and in five of eight classes compared to the CRF applied on the A-MRAT. However, the accuracy improvements of the MRF compared to the CRF models is in the order of 1.5% in the highest case, while the proposed CRF, using MSRAG E, is capable of improving the classification results in the order 22% of the accuracy value for certain classes (i.e., C3). The low overall accuracy of the standard CRF can be due to the parameter settings in our implementation.

5.3.3. The Impact of the Intra- and Inter-Scale Potentials

The above comparison between the MRF and CRF models does not use the same features to characterize the classes. Indeed, for MRF, only the spectral signature of the regions has been used, while, for the CRF, each region was characterized by its spectral, texture and shape information. Therefore, in order to measure the real impact of each individual potential (intra- and inter-scale) in the classification results, we considered different sets of values for the potential combination of the energy expression Equation (33). The obtained results are detailed in Table 7 using the considered areal color image.

The results indicate that for the proposed model, for the considered image and model structure (MSRAG), the multi-scale information is valuable. This can be seen comparing the results that involve multi-scale information (d2 and d3 used, with the second column as Yes) with the others that do not involve it (only d1 used, with the second column as No).

From the potentials’ influence point of view, it is clear that their different combinations on the energy expression of Equation (33) lead to different classification results. However, each potential contribution will vary depending on the image nature and on the training samples.

For instance, if certain images exhibit clear class transition patterns and these patterns are present in the training samples, then we expect the intra- scale potential to be influential on the energy expression in order to improve the classification results. If, on the contrary, no clear class transition patterns can be distinguished in the image or if they are not present in the training samples, then we should not expect the intra-scale potential to provide valuable information on the energy expression. A similar analysis can be done for the inter-scale potential.

5.3.4. Computational Complexity

All of the reported experiments of Section 5.3.2. have been made on an Intel Core i7 at 3.40 Ghz PC with 8 GB of RAM. Table 8 summarizes the computational costs of the evaluated approaches, considering the same hierarchical segmentation as the input.

For all evaluated models, we assume that they share the same computational complexity for the hierarchical segmentation stage, which can be more or less efficient (it can vary depending on the used segmentation scheme). As can be seen, SVM applied on the localization scale is the most computational efficient, since it is simple, but cannot benefit from any available contextual information (hierarchy). For the MRF and CRF models, the major difference in computational complexity is due to the parameter learning and inference process for multiclass problems. The computational complexity is dependent on: (i) the parameter learning approach (maximum likelihood or max-margin); and (ii) the inference process according to the type of input structure of the hierarchy (with or without loops). From the theoretical point of view, likewise in [24], the proposed parameter learning is much more efficient, mainly due to the reduction of the number of constraints (∀x ≠ x̂ in Equation (23)) to a much less number of constraints (∀x𝒫 ≠ x̂𝒫, ∀𝒫𝒯 in Equation (38) in Appendix A); but it is also due to the proposed problem decomposition into multiple individual problems that can be solved using the fixed-point method (see [26] for the details on the method’s efficiency). The execution time values detailed in Table 8 confirm the suitability of the proposed max-margin parameter learning strategy. The proposed CRF and the standard CRF models, both involving loopy structures and trained using max-margin approach, perform similar to the MRF with a loopy-free, tree-like structure. The proposed CRF also largely outperforms the standard CRF model when trained in a traditional maximum likelihood fashion. For the CRF models, the reported execution time for the inference process accounts for a fixed set of values for d1, …, d5.

6. Conclusions

The contributions of this paper are a multi-scale CRF model with novel energies along with an efficient parameter learning strategy, where the complex learning problem is decomposed into simpler, individual multi-class sub-problems.

The paper described the main principles of our method and illustrated classification results on a set of remote sensing images, together with qualitative and quantitative comparisons with pixel-based techniques, region-based approaches that follow a Bayesian–Markovian framework, either on hierarchical structures (MRF) or the original image lattice, region-based SVM and standard CRF models. The extensive experimental results validated the suitability of the model for region-based image labeling.

We found that there is a relationship between the obtained segmentation and the classification process. The performed experiments suggested that this influence is related (although not exclusively) to the created hierarchy (parent-child link), the total number of regions and their average size, as well as the features used.

In the future, we will be investigating how to pose the per-clique weight (d) and potential (θ) parameter learning as one joint optimization problem and how to solve it efficiently. We are also planning to compare the model accuracy and performance with respect to other (more recent) state-of-the-art models in other (not necessarily aerial) type of images (e.g., [2]).

Author Contributions

Mitchel Alioscha-Perez designed and implemented the methodology, conducted the experiments and analysis with guidance from Hichem Sahli, who defined the energy terms. The manuscript was written and revised by Mitchel Alioscha-Perez. Hichem Sahli was responsible for the research design and contributed to editing and review of the manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix

A. Decomposition of QP

Starting from the problem (23), recalling from Equations (21) and (22) that E𝒫 and Ê𝒫 are dependant on θ, and considering Equation (25), we obtain the following optimization problem:

max λ > 0 , θ λ
s . t . 𝒫 𝒯 ( E 𝒫 E ^ 𝒫 ) 𝒫 𝒯 λ 𝒫 x x ^ θ 2 1

From Corollary 1 (see Appendix B), under the maximum margin criteria, we can redefine constraints (35) as a set of 𝒫 constraints E𝒫Ê𝒫λℓ𝒫, ∀𝒫𝒯, as follows:

max λ > 0 , θ λ
s . t . E 𝒫 E ^ 𝒫 λ 𝒫 x x ^ , 𝒫 𝒯 θ 2 1

Considering a model partition (i.e., Piecewise)) [20,21], labels can be structured as ∀x ∝ ∀x𝒫 ≠ x̂𝒫, ∀𝒫𝒯 ; then:

max λ > 0 , θ λ s . t . E 𝒫 E ^ 𝒫 λ 𝒫 x 𝒫 x ^ 𝒫 , 𝒫 𝒯 θ 2 1

Transforming to eliminate λ [50], we obtain the following QP:

min θ 1 2 θ 2 2 s . t . E 𝒫 E ^ 𝒫 𝒫 x 𝒫 x ^ 𝒫 , 𝒫 𝒯

Taking into account the per-potential type (𝒫) block structure of θ, the equivalence k θ 2 2 = Σ 𝒫 θ 𝒫 2 2 [51] leads us to:

min θ 1 2 𝒫 θ 𝒫 2 2
s . t . E 𝒫 E ^ 𝒫 𝒫 x 𝒫 x ^ 𝒫 , 𝒫 𝒯
where the problem decouples as:
𝒫 𝒯 { min θ 𝒫 1 2 θ 𝒫 2 2 s . t . E 𝒫 E ^ 𝒫 𝒫 , x 𝒫

Note that for the case of x𝒫 = 𝒫, Constraint (41) is met, since both sides of the inequality (E𝒫Ê𝒫 and 𝒫) become zero; therefore, we considered ∀x𝒫 instead of ∀x𝒫 ≠ x̂𝒫.

Let us define Ω 𝒫 as the set of label combinations for any clique-order 𝒫 and N𝒫 = |𝒫| the total number of (related) factors [21]; in our problem, Ω𝒫 = site ∈ {1, …, L}, Ω 𝒫 = neighbor ∈ {0, 1} and Ω 𝒫 = parent/child ∈ {0, 1}. For each i-th factor (i = {1, ., N𝒫}), we associate a possible label xi ∈ Ω 𝒫 with the corresponding observation yi and the corresponding true label i from the training data:

𝒫 𝒯 { min θ 𝒫 1 2 θ 𝒫 2 2 s . t . E 𝒫 E ^ 𝒫 𝒫 , i = 1 N 𝒫 , x i Ω 𝒫

The introduction of slack variables ξ will allow one to assume errors on the obtained separating hyperplane (soft margin):

𝒫 𝒯 { min θ 𝒫 , ξ i 1 2 θ 𝒫 2 2 + C i ξ i s . t . E 𝒫 E ^ 𝒫 𝒫 ξ i , i = 1 N 𝒫 , x i Ω 𝒫

B. Necessary and Sufficient Conditions for Maximum Margin

Proposition 1: Constraints xnlnn are sufficient (although not necessary) conditions for the constraint

xn ≥ ∑ln to be satisfied.

Proof: This is a property of inequalities in algebra. Let { x n , l n } n = 1 N be a set of arbitrary real numbers satisfying that xnlnn:

x 1 l 1 x 1 + l 2 l 1 + l 2 x 2 l 2 x 1 + x 2 l 1 + l 2 x N l N x 1 + + x N l 1 + + l N x n l n
Proposition 2: The constraint ∑xn ≥ ∑ln admits a representation of the form ∑uU xu, where U = {n | xn < ln} are indexes related to unsatisfied constraints, UA, A = {1,, N} and is a slack variable.

Proof: Let S = A \ U be the relative complement of U in A. Let us consider the residual (xplp), where xp = ∑sS xs and lp = ∑sS ls, part of the following slack definition:

l ˜ = u U l u ( x p l p )
then:
u U x u l ˜ u U x u u U l u s s x s + s S l s due to Equation ( 46 ) n A x n n A l n since A = S U
Theorem 1: Let x ∈ ℝN = {xn | ∀nA} be a solution of a problem P satisfying the constraint ∑xn∑ln, ln ∈ ℝ, where A = {1, , N}. Let M : ℝN → ℝ, M(x) = ∑xn∑ln be a solution margin of P associated with x, U = {n | xn < ln} and S = A \ U be the relative complement of U in A. Any solution margin MU≠∅ of P, such that U ≠ ∅, will be bounded above by another solution margin MU = ∅ of P, where U = ∅.

Proof:

M = s S ( x s l s ) + u U ( x u l u ) where ( x s l s ) 0 by definition of S if U = then u U ( x u l u ) vanishes ( zero ) if U x u l u < 0 u U then u U ( x u l u ) < 0
therefore, MU≠∅ < MU = ∅.

Corollary 1: For any candidate solution x of a problem P, the constraints xnlnn are: (i) sufficient conditions to satisfy the constraint ∑xn ≥ ∑ln; and (ii) necessary and sufficient conditions to maximize the margin ∑(xnln).

Proof:

Proposition 1 is the proof of (i), while Proposition 2 confirms the possible existence of solutions x that do not necessarily satisfy xnlnn. However, a consequence of Theorem 1 is that only solutions that satisfy xnlnn (U = ∅) will guarantee the maximum margin M = ∑(xnln), which proves (ii).

C. Single and Pairwise Energy of Segments

- Area

E area ( r i ) = 0.002 N M Area ( r i )
where N × M is the image size and Area(ri) is the area of the region. The 0.002 term means that regions with a 0.2% area of the whole image have energy 1.0.

- Convexity

E conv ( r i ) = Area ( ConvexHull ( r i ) ) Area ( r i ) 1.25
represents the region convexity energy. We assume that regions with convexity larger than 1.25 are not preferred, and those with convexity energy smaller than 1.25 are desired. Therefore, the offset for the convexity energy is set to −1.25.

Compactness

E comp ( r i ) = Perimeter ( r i ) 2 4 π Area ( r i ) = 1.25
represents the region compactness energy.

The compactness energy Perimeter ( r i ) 2 4 π Area ( r i ) is always greater than or equal to one (one for a circle, 4/π for a square). We assume that regions with compactness larger than 1.25 are not preferred. Again, the offset for compactness energy is set to −1.25.

Homogeneity

E hom ( r i ) = 1 σ ( r i ) V ( r i )
represents the region’s intensity (I) homogeneity. Homogeneity is largely related to the local information extracted from an image and reflects how uniform a region is.

Color variance

E var c ( r i ) = 1 15 σ c ( r i )
represents the color homogeneity of a region, with σc(ri) the standard deviation of the color c ∈ {L, a, b} within region ri. The normalization factor for color variances is derived from statistical analysis of the color variance results on the image data base [52].

Color contrast
E CMDif ( r i , r j ) = ( Δ μ L ( r i , r j ) ) 2 + ( Δ μ a ( r i , r j ) ) 2 + ( Δ μ b ( r i , r j ) ) 2 T d

represents the normalized mean color difference. The normalization factor Td (for color merging) is estimated as Td = µdσv, with µd the mean of the color differences Dis, and σ u = 1 / n i = 1 n ( D i μ d ) 2 the standard deviation of the n = k ( k 1 ) 2 color differences between the k regions.

Edgeness

Eedge(ri, rj) represents the dynamics of the contour of the edge between ri and rj.

References

  1. Reynolds, J.; Murphy, K.P. Figure-ground segmentation using a hierarchical conditional random field. Proceedings of the 2007 Fourth Canadian Conference on Computer and Robot Vision, Montreal, QC, Canada, 28–30 May 2007; pp. 175–182.
  2. Farabet, C.; Couprie, C.; Najman, L.; LeCun, Y. Learning hierarchical features for scene labeling. IEEE Trans. Patt. Anal. Mach. Intell 2013, 35, 1915–1929. [Google Scholar]
  3. Katartzis, A.; Vanhamel, I.; Sahli, H. A hierarchical Markovian model for multiscale region-based classification of vector-valued images. IEEE Trans. Geosci. Remote Sens 2005, 43, 548–558. [Google Scholar]
  4. Besag, J. On the statistical analysis of dirty pictures. J. R. Stat. Soc. Ser. B 1986, 48, 259–302. [Google Scholar]
  5. Geman, S.; Geman, D. Stochastic relaxation, Gibbs distributions and the bayesian restoration of images. J. Appl. Stat 1993, 20, 25–62. [Google Scholar]
  6. Lafferty, J.; McCallum, A.; Pereira, F. Conditional random fields: Probabilistic models for segmenting and labeling sequence data. In Proceedings of the Eighteenth International Conference on Machine Learning; Morgan Kaufmann Publishers Inc: San Francisco, CA, USA, 2001; pp. 282–289. [Google Scholar]
  7. Kumar, S.; Hebert, M. Discriminative random fields: A discriminative framework for contextual interaction in classification. Proceedings of the 2003 Ninth IEEE International Conference on Computer Vision, Nice, France, 13–16 October 2003; pp. 1150–1157.
  8. He, X.; Zemel, R.S.; Carreira Perpinan, M.A. Multiscale conditional random fields for image labeling. In Proceedings of the 2004. In IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Washington, DC, USA, 27 June–2 July 2004; pp. 695–702.
  9. Yang, M.Y.; Förstner, W.; Drauschke, M. A Hierarchical conditional random field model for labeling and classifying images of man-made scenes. Proceedings of the 2011 IEEE International Conference on Computer Vision Workshops, Barcelona, Spain, 6–13 November 2011; pp. 196–203.
  10. Munoz, D.; Bagnell, J.A.; Vandapel, N.; Hebert, M. Contextual classification with functional max-margin markov networks. Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 975–982.
  11. Gould, S.; Rodgers, J.; Cohen, D.; Elidan, G.; Koller, D. Multi-class segmentation with relative location prior. Int. J. Comput. Vis 2008, 80, 300–316. [Google Scholar]
  12. Su, X.; He, C.; Feng, Q.; Deng, X.; Sun, H. A supervised classification method based on conditional random fields with multiscale region connection calculus model for SAR image. IEEE Geosci. Remote Sens. Lett 2011, 8, 497–501. [Google Scholar]
  13. Martinis, S.; Twele, A. A hierarchical spatio-temporal Markov model for improved flood mapping using multi-temporal X-band SAR data. Remote Sens 2010, 2, 2240–2258. [Google Scholar]
  14. Kasetkasem, T.; Rakwatin, P.; Sirisommai, R.; Eiumnoh, A. A joint land cover mapping and image registration algorithm based on a Markov random field model. Remote Sens 2013, 5, 5089–5121. [Google Scholar]
  15. Vishwanathan, S.V.N.; Schraudolph, N.N.; Schmidt, M.W.; Murphy, K.P. Accelerated training of conditional random fields with stochastic gradient methods. Proceedings of the 23rd International Conference on Machine Learning, Pittsburgh, PA, USA, 25–29 June 2006; pp. 969–976.
  16. Ratliff, N.D.; Bagnell, J.A.; Zinkevich, M.A. Subgradient methods for structured prediction. Proceedings of the Eleventh International Conference on Artificial Intelligence and Statistics (AIStats), San Juan, Puerto Rico, 21–24 March 2007; pp. 380–387.
  17. Tsochantaridis, I.; Hofmann, T. Large margin methods for structured and interdependent output variables. J. Mach. Learn. Res 2005, 6, 1453–1484. [Google Scholar]
  18. Szummer, M.; Kohli, P.; Hoiem, D. Learning CRFs using graph cuts. Proceedings of the 10th European Conference on Computer Vision: Part II, Marseille, France, 12–18 October 2008; pp. 582–595.
  19. Nowozin, S.; Gehler, P.; Lampert, C. On parameter learning in CRF-based approaches to object class image segmentation. Proceedings of the 11th European Conference on Computer Vision: Part VI, Heraklion, Crete, Greece, 5–11 September 2010; pp. 98–111.
  20. Sutton, C.; Mccallum, A. Piecewise training for undirected models. Proceedings of the Twenty-First Conference Annual Conference on Uncertainty in Artificial Intelligence (UAI-05), Edinburgh, Scotland, 26–29 July 2005; pp. 568–575.
  21. Sutton, C.; Mccallum, A. Piecewise pseudolikelihood for efficient training of conditional random fields. Proceedings of the 24th International Conference on Machine learning, Corvallis, OR, USA, 20–24 June 2007; pp. 863–870.
  22. Zhong, P.; Wang, R. Learning conditional random fields for classification of hyperspectral images. IEEE Trans. Image Process 2010, 19, 1890–1907. [Google Scholar]
  23. Koltun, V. Efficient inference in fully connected CRFs with gaussian edge potentials. Adv. Neural Inf. Process. Syst. 2011. arXiv:1210.5644. [Google Scholar]
  24. Alahari, K.; Russell, C.; Torr, P.H.S. Efficient piecewise learning for conditional random fields. Proceedings of the 2010 IEEE Conference on Computer Vision and Pattern Recognition, San Francisco, CA, USA, 13–18 June 2010; pp. 895–901.
  25. Kumar, M.P.; Zisserman, A.; Torr, P.H.S. Efficient discriminative learning of parts-based models. Proceedings of the 2009 IEEE 12th Conference on Computer Vision and Pattern Recognition, Kyoto, Japan, 29 September–2 October 2009; pp. 552–559.
  26. Crammer, K.; Singer, Y. On the algorithmic implementation of multiclass kernel-based vector machines. J. Mach. Learn. Res 2001, 2, 265–292. [Google Scholar]
  27. Vanhamel, I.; Pratikakis, I.; Sahli, H. Multiscale gradient watersheds of color images. IEEE Trans. Image Process 2003, 12, 617–626. [Google Scholar]
  28. Catté, F.; Lions, P.L.; Morel, J.M.; Coll, T. Image selective smoothing and edge detection by nonlinear diffusion. SIAM J. Numer. Anal 1992, 29, 182–193. [Google Scholar]
  29. Perona, P.; Malik, J. Scale-space and edge detection using anisotropic diffusion. IEEE Trans. Pattern Anal. Mach. Intell 1990, 12, 629–639. [Google Scholar]
  30. Black, M.; Sapiro, G.; Marimont, D.; Heeger, D. Robust anisotropic diffusion. IEEE Trans. Image Process 1998, 7, 421–432. [Google Scholar]
  31. Koenderink, J. The structure of images. Biol. Cybernet 1984, 50, 363–370. [Google Scholar]
  32. Vanhamel, I.; Alrefaya, M.; Sahli, H. Multi scale representation for remotely sensed images using fast anisotropic diffusion filtering. Proceedings of the 2010 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Honolulu, HI, USA, 25–30 July 2010; pp. 2222–2225.
  33. Vanhamel, I.; Mihai, C.; Sahli, H.; Katartzis, A.; Pratikakis, I. Scale selection for compact scale-space representation of vector-valued images. Int. J. Comput. Vis 2008, 84, 194–204. [Google Scholar]
  34. Blaschke, T. Object based image analysis for remote sensing. J. Photogramm. Remote Sens 2010, 65, 2–16. [Google Scholar]
  35. Geerinck, T.; Sahli, H.; Henderickx, D.; Vanhamel, I.; Enescu, V. Modeling attention and perceptual grouping to salient objects. In Attention in Cognitive Systems; Paletta, L., Tsotsos, J.K., Eds.; Springer-Verlag: Berlin, Germany, 2009; pp. 166–182. [Google Scholar]
  36. Marcotegui, B.B.S. Fast implementation of waterfall based on graphs. In 40 Years On Mathematical Morphology; Springer-Verlag: Berlin, Germany, 2005; pp. 177–186. [Google Scholar]
  37. O’Callaghan, R.J.; Bull, D.R. Combined morphological-spectral unsupervised image segmentation. Image Process 2005, 14, 49–62. [Google Scholar]
  38. DiZenzo, S. A note on the gradient of a multi-image. Comput. Vis. Graph. Image Process 1986, 33, 116–125. [Google Scholar]
  39. Ladicky, L.; Russell, C.; Kohli, P.; Torr, P.H.S. Associative hierarchical CRFs for object class image segmentation. Proceedings of the 2009 IEEE 12th International Conference on Computer Vision, Kyoto, Japan, 29 September 2009–2 October 2009; pp. 739–746.
  40. Kohli, P.; Torr, P. Robust higher order potentials for enforcing label consistency. Int. J. Comput. Vis 2009, 82, 302–324. [Google Scholar]
  41. Aronszajn, N. Theory of reproducing kernels. Trans. Am. Math. Soc 1950, 68, 337–404. [Google Scholar]
  42. Zhou, H.; Wang, R.; Wang, C. A novel extended local-binary-pattern operator for texture analysis. Inf. Sci 2008, 178, 4314–4325. [Google Scholar]
  43. Komodakis, N. Efficient training for pairwise or higher order CRFs via dual decomposition. In proceedings of the 2011 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 20–25 June 2011; pp. 1841–1848.
  44. Yedidia, J.; Freeman, W. Understanding belief propagation and its generalizations. In Exploring Artificial Intelligence in the New Millennium; Morgan Kaufmann Publishers Inc: San Francisco, CA, USA, 2003; pp. 236–269. [Google Scholar]
  45. Wainwright, M. Tree-reweighted belief propagation algorithms and approximate ML estimation by pseudo-moment matching. Workshop Artif. Intell. Stat 2003, 21, 97–104. [Google Scholar]
  46. Boykov, Y.; Kolmogorov, V. An experimental comparison of min-cut/max-flow algorithms for energy minimization in vision. Science 2004, 26, 1124–1137. [Google Scholar]
  47. Aksoy, S. Spatial techniques for image classification. In Signal and Image Processing for Remote Sensing; CRC Press: Boca Raton, FL, USA, 2006; pp. 491–513. [Google Scholar]
  48. Petropoulos, G.P.; Kalaitzidis, C.; Vadrevu, K.P. Support vector machines and object-based classification for obtaining land-use/cover cartography from Hyperion hyperspectral imagery. Comput. Geosci 2012, 41, 99–107. [Google Scholar]
  49. Schmidt, M. UGM: Matlab Code for Undirected Graphical Models. Available online: http://www.cs.ubc.ca/~schmidtm/Software/UGM.html/ (accessed on 20 April 2011).
  50. Taskar, B.; Guestrin, C.; Koller, D. Max-Margin Markov networks. Adv. Neural. Inf. Process. Syst 2004, 16, 25–32. [Google Scholar]
  51. Bach, F.R.; Lanckriet, G.R.G.; Jordan, M.I. Multiple kernel learning, conic duality, and the SMO algorithm. Proceedings of the Twenty-First International Conference on Machine Learning-ICML ’04, Banff, AB, Canada, 4–8 July 2004; pp. 6–13.
  52. Luo, J.; Guo, C. Perceptual grouping of segmented regions in color images. Patt. Recogn 2003, 36, 2781–2792. [Google Scholar]
Figure 1. Stages involved in the proposed multi-scale region-based image labeling process.
Figure 1. Stages involved in the proposed multi-scale region-based image labeling process.
Remotesensing 06 06727f1 1024
Figure 2. The multi-scale region adjacency graph (MSRAG) defines the hierarchical structure of an image. (a) Nested hierarchy of partitions; (b) multi-scale region adjacency graph (MSRAG).
Figure 2. The multi-scale region adjacency graph (MSRAG) defines the hierarchical structure of an image. (a) Nested hierarchy of partitions; (b) multi-scale region adjacency graph (MSRAG).
Remotesensing 06 06727f2 1024
Figure 3. In the considered structure, during classification, the information flows from coarser (higher) to finer (lower) scales to regulate long-range labels consistency and flows within the same scale to capture local interactions between neighboring regions.
Figure 3. In the considered structure, during classification, the information flows from coarser (higher) to finer (lower) scales to regulate long-range labels consistency and flows within the same scale to capture local interactions between neighboring regions.
Remotesensing 06 06727f3 1024
Figure 4. Piecewise pairwise factors of a Markovian model are considered individual pieces of the same original model. (a) Original model. (b) Standard piecewise training of the same model (pairwise factors).
Figure 4. Piecewise pairwise factors of a Markovian model are considered individual pieces of the same original model. (a) Original model. (b) Standard piecewise training of the same model (pairwise factors).
Remotesensing 06 06727f4 1024
Figure 5. Information flows from the finest to coarsest scales in order to conform the training sample set structured according to the MSRAG. (a) The provided training samples. (b) Training at Scale 1. (c) Training at Scale 5.
Figure 5. Information flows from the finest to coarsest scales in order to conform the training sample set structured according to the MSRAG. (a) The provided training samples. (b) Training at Scale 1. (c) Training at Scale 5.
Remotesensing 06 06727f5 1024
Figure 6. Inference at coarser (higher) scales will influence inference on finer (lower) scales, enforcing long-range dependencies. (a) Inference result at Scale 1. (b) Inference result at Scale 9.
Figure 6. Inference at coarser (higher) scales will influence inference on finer (lower) scales, enforcing long-range dependencies. (a) Inference result at Scale 1. (b) Inference result at Scale 9.
Remotesensing 06 06727f6 1024
Figure 7. The Pavia University dataset (45°11′ 39.84″ N 9°08′06″ E). (a) Image. (b) Training samples. (c) Testing samples.
Figure 7. The Pavia University dataset (45°11′ 39.84″ N 9°08′06″ E). (a) Image. (b) Training samples. (c) Testing samples.
Remotesensing 06 06727f7 1024
Figure 8. Pavia Center dataset (45°11′0.96″ N 9 08′25.44″E). (a) Image. (b) Training samples. (c) Testing samples.
Figure 8. Pavia Center dataset (45°11′0.96″ N 9 08′25.44″E). (a) Image. (b) Training samples. (c) Testing samples.
Remotesensing 06 06727f8 1024
Figure 9. The considered areal color (RGB) image. (a) Training samples. (b) Testing samples.
Figure 9. The considered areal color (RGB) image. (a) Training samples. (b) Testing samples.
Remotesensing 06 06727f9 1024
Figure 10. Scales and classification maps of the augmented multiscale region adjacency tree (A-MRAT), MSRAG A and MSRAG B. The first two columns provide regions at Scales 1 and 9, respectively, while the last column provides the classification map. (a) A-MRAT Scale 1. (b) A-MRAT Scale 9. (c) A-MRAT class-map. (d) MSRAG A Scale 1. (e) MSRAG A Scale 9. (f) MSRAG A class-map. (g) MSRAG B Scale 1. (h) MSRAG B Scale 9. (i) MSRAG B class-map.
Figure 10. Scales and classification maps of the augmented multiscale region adjacency tree (A-MRAT), MSRAG A and MSRAG B. The first two columns provide regions at Scales 1 and 9, respectively, while the last column provides the classification map. (a) A-MRAT Scale 1. (b) A-MRAT Scale 9. (c) A-MRAT class-map. (d) MSRAG A Scale 1. (e) MSRAG A Scale 9. (f) MSRAG A class-map. (g) MSRAG B Scale 1. (h) MSRAG B Scale 9. (i) MSRAG B class-map.
Remotesensing 06 06727f10 1024
Figure 11. Scales and classification map of the MSRAG C, MSRAG D and MSRAG E. The first two columns provide regions at Scales 1 and 9, respectively, while the last column provides the classification map. (a) MSRAG C Scale 1. (b) MSRAG C Scale 9. (c) MSRAG C class-map. (d) MSRAG D Scale 1. (e) MSRAG D Scale 9. (f) MSRAG D class-map. (g) MSRAG E Scale 1. (h) MSRAG E Scale 9. (i) MSRAG E class-map.
Figure 11. Scales and classification map of the MSRAG C, MSRAG D and MSRAG E. The first two columns provide regions at Scales 1 and 9, respectively, while the last column provides the classification map. (a) MSRAG C Scale 1. (b) MSRAG C Scale 9. (c) MSRAG C class-map. (d) MSRAG D Scale 1. (e) MSRAG D Scale 9. (f) MSRAG D class-map. (g) MSRAG E Scale 1. (h) MSRAG E Scale 9. (i) MSRAG E class-map.
Remotesensing 06 06727f11 1024
Figure 12. Different MSRAG generations will provide different classification results. (a) Training samples. (b) Testing samples. (c) A-MRAT. (d) MSRAG A. (e) MSRAG B. (f) MSRAG C. (g) MSRAG D. (h) MSRAG E.
Figure 12. Different MSRAG generations will provide different classification results. (a) Training samples. (b) Testing samples. (c) A-MRAT. (d) MSRAG A. (e) MSRAG B. (f) MSRAG C. (g) MSRAG D. (h) MSRAG E.
Remotesensing 06 06727f12 1024
Figure 13. The proposed multi-scale CRF exhibits superior classification results compared with MRF for the airborne image. (a) Training samples. (b) Testing samples. (c) Multi-scale MRF (MRAT) [3]. (d) Proposed multi-scale CRF (A-MRAT).
Figure 13. The proposed multi-scale CRF exhibits superior classification results compared with MRF for the airborne image. (a) Training samples. (b) Testing samples. (c) Multi-scale MRF (MRAT) [3]. (d) Proposed multi-scale CRF (A-MRAT).
Remotesensing 06 06727f13 1024
Table 1. Pavia University dataset: performance evaluation of the proposed approach in comparison with previous results.
Table 1. Pavia University dataset: performance evaluation of the proposed approach in comparison with previous results.
Model# of Used Spectral BandsAccuracy (%)
Pixel-Level Bayesian [47]10 (PCA from 103)74.92
Quadratic Gaussian [47]10 (PCA from 103)81.27
Region-Level Bayesian [47]10 (PCA from 103)84.44
Proposed CRF model3 (68, 30 and 2)87.16
Table 2. Pavia University dataset: performance evaluation of the proposed approach in comparison with previous results; per-class accuracy (%) and overall accuracy (%). P-L-B, pixel-level Bayesian; R-B-B, region-based Bayesian.
Table 2. Pavia University dataset: performance evaluation of the proposed approach in comparison with previous results; per-class accuracy (%) and overall accuracy (%). P-L-B, pixel-level Bayesian; R-B-B, region-based Bayesian.
ClassProposed CRFModels in [47]
P-L-BR-B-B
C1-Asphalt86.3861.0069.67
C2-Meadows95.9178.8792.48
C3-Gravel57.2269.8464.79
C4-Trees83.2995.5395.99
C5-Metal Sheets90.0499.7099.93
C6-Bare Soil76.0974.4779.82
C7-Bitumen67.5162.3377.67
C8-Bricks83.6556.0671.56
C9-Shadow95.3597.4798.10
Overall Accuracy87.1674.9284.44
Table 3. Pavia Center dataset: description of the randomly selected samples, and performance evaluation (as accuracy %) of the proposed approach.
Table 3. Pavia Center dataset: description of the randomly selected samples, and performance evaluation (as accuracy %) of the proposed approach.
ClassTraining PixelsTotal Pixels(%)Proposed CRFModels in [47]
P-L-BR-B-B
C1-Water441065,97106.6899.9899.8599.74
C2-Trees713759809.3889.0784.581.71
C3-Meadows346309011.2069.1987.9695.21
C4-Bricks174268506.4876.1683.3583.69
C5-Bare Soil800658412.1595.3778.7793.24
C6-Asphalt1058924811.4490.2785.3993.74
C7-Bitumen1202728716.5095.183.1889.28
C8-Tiles438642,82610.2499.7297.6798.96
C9-Shadow268286309.3689.3182.8286.20

Overall13,357148,15209.0297.2094.8996.76
Table 4. Per-class accuracy (%) and overall accuracy (%) of the multi-scale CRF applied to the considered multi-scale region adjacency graphs.
Table 4. Per-class accuracy (%) and overall accuracy (%) of the multi-scale CRF applied to the considered multi-scale region adjacency graphs.
ClassA-MRATMSRAG AMSRAG BMSRAG CMSRAG DMSRAG E
C1-Shadow83.9885.4789.4992.1590.5297.30
C2-Roads/Path (Dark)79.6577.5275.5587.8785.3687.49
C3-Grass land82.8879.9080.1968.0580.1883.22
C4-Rooftop 1 (Blue)47.0046.1648.0549.3052.6049.58
C5-Rooftop 2 (Red)77.7679.8379.8879.6777.4379.23
C6-Rooftop 3 (Green)96.0696.0696.0697.7598.2597.25
C7-Arid land61.7364.6365.6281.6077.8473.86
C8-Roads/Path (White)71.1971.8677.8074.9969.4974.55

Overall Accuracy75.4974.9275.3978.7680.5281.45
Table 5. The different hierarchical segmentation strategies provide different numbers of regions, as well as average region sizes (between brackets) per scale, which influence the classification results of the CRF model.
Table 5. The different hierarchical segmentation strategies provide different numbers of regions, as well as average region sizes (between brackets) per scale, which influence the classification results of the CRF model.
ScaleA-MRATMSRAG AMSRAG BMSRAG CMSRAG DMSRAG E
114,870 (43)14,870 (43)14,870 (43)3224 (199)3224 (199)3081 (208)
212,997 (49)11,781 (54)6776 (94)2499 (256)1475 (434)1417 (452)
311,481 (56)9941 (64)4253 (150)2073 (309)911 (703)886 (722)
410,155 (63)8680 (74)3150 (203)1767 (362)659 (971)647 (989)
58953 (71)7773 (82)2523 (254)1381 (414)526 (1217)522 (1226)
67926 (81)7106 (90)2170 (295)1241 (463)440 (1455)442 (1448)
76990 (92)6557 (98)1933 (331)1130 (516)387 (1654)383 (1671)
86145 (104)6096 (105)1761 (363)1035 (566)348 (1839)340 (1882)
95397 (119)5710 (112)1628 (393)963 (618)316 (2025)303 (2112)

Total84,91478,51439,07415,31382868021

Accuracy75.4974.9275.3978.4980.5281.45
Table 6. Per-class accuracy (%) and overall accuracy (%) of the proposed CRF model considering different MSRAG structures and the region-based multi-scale Markov random fields (MRF) of [3]; the highest per-class accuracy has been underlined. Std. CRF refers to a CRF model with standard energies/potentials.
Table 6. Per-class accuracy (%) and overall accuracy (%) of the proposed CRF model considering different MSRAG structures and the region-based multi-scale Markov random fields (MRF) of [3]; the highest per-class accuracy has been underlined. Std. CRF refers to a CRF model with standard energies/potentials.
ClassSVMMRF [3]Std. CRFProposed CRFProposed CRF
A-MRATA-MRATMSRAG E
C1-Shadow84.4985.4283.4583.9897.30
C2-Roads/Path (Dark)82.9880.8064.3179.6587.49
C3-Grass land76.9460.2342.5482.8883.22
C4-Rooftop 1 (Blue)48.4740.3248.3447.0049.58
C5-Rooftop 2 (Red)80.7779.9376.0877.7679.23
C6-Rooftop 3 (Green)00.0098.7400.0096.0697.25
C7-Arid land60.9956.3154.3561.7373.86
C8-Roads/Path (White)62.2163.2668.5171.1974.55

Overall Accuracy72.7866.8855.7175.4981.45
Table 7. The potentials mixture weights (dm, m = 1, …, 3 in Equation (33)) establish the final performance of the multi-scale CRF classifier.
Table 7. The potentials mixture weights (dm, m = 1, …, 3 in Equation (33)) establish the final performance of the multi-scale CRF classifier.
Solution #Multi-Scale InformationUnary (d1)PairwiseCRF Acc
Intra-Scale (d2)Inter-Scale (d3)
1No1.00.0-78.30
2Yes1.00.00.080.30
3Yes1.00.01.080.30
4Yes1.00.11.080.52
5No1.00.3-78.87
6No6.00.3-79.17
7Yes6.00.30.081.45
8Yes6.00.311.081.45
Table 8. Execution time for the parameter learning and inference stage of different models: SVM as Support Vector Machines; Prop. CRF as the proposed CRF model; Std. CRF (Max-Margin) as CRF model with standard energies/potentials trained using max-margin approach; Std. CRF (Max-Likelihood) as CRF model with standard energies/potentials trained using max-likelihood approach.
Table 8. Execution time for the parameter learning and inference stage of different models: SVM as Support Vector Machines; Prop. CRF as the proposed CRF model; Std. CRF (Max-Margin) as CRF model with standard energies/potentials trained using max-margin approach; Std. CRF (Max-Likelihood) as CRF model with standard energies/potentials trained using max-likelihood approach.
StageSVMMRFProp. CRFStd. CRF (Max-Margin)Std. CRF (Max-Likelihood)
Params Learning--3 s36 s>4000 s *
Inference--340 s716 s757 s

Total<1 s789 s343 s752 s> 4500 s

*Ended before convergence after 150 iterations using l-BFGS descent (Broyden–Fletcher–Goldfarb–Shanno).

Share and Cite

MDPI and ACS Style

Alioscha-Perez, M.; Sahli, H. Efficient Learning of Spatial Patterns with Multi-Scale Conditional Random Fields for Region-Based Classification. Remote Sens. 2014, 6, 6727-6764. https://doi.org/10.3390/rs6086727

AMA Style

Alioscha-Perez M, Sahli H. Efficient Learning of Spatial Patterns with Multi-Scale Conditional Random Fields for Region-Based Classification. Remote Sensing. 2014; 6(8):6727-6764. https://doi.org/10.3390/rs6086727

Chicago/Turabian Style

Alioscha-Perez, Mitchel, and Hichem Sahli. 2014. "Efficient Learning of Spatial Patterns with Multi-Scale Conditional Random Fields for Region-Based Classification" Remote Sensing 6, no. 8: 6727-6764. https://doi.org/10.3390/rs6086727

APA Style

Alioscha-Perez, M., & Sahli, H. (2014). Efficient Learning of Spatial Patterns with Multi-Scale Conditional Random Fields for Region-Based Classification. Remote Sensing, 6(8), 6727-6764. https://doi.org/10.3390/rs6086727

Article Metrics

Back to TopTop