Next Article in Journal
Application of UAV Remote Sensing for a Population Census of Large Wild Herbivores—Taking the Headwater Region of the Yellow River as an Example
Next Article in Special Issue
Band Priority Index: A Feature Selection Framework for Hyperspectral Imagery
Previous Article in Journal
Teleconnections and Interannual Transitions as Observed in African Vegetation: 2015–2017
Previous Article in Special Issue
A Component-Based Multi-Layer Parallel Network for Airplane Detection in SAR Imagery
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Level Set Method for Infrared Image Segmentation Using Global and Local Information

1
School of Electronic and Optical Engineering, Nanjing University of Science and Technology, Nanjing 210094, China
2
Department of Electrical and Computer Engineering, Computer Vision and Systems Laboratory, Laval University, 1065 av. de la Médecine, Quebec City, QC G1V 0A6, Canada
3
State Key Laboratory of Networking and Switching Technology, Beijing University of Posts and Telecommunications, Beijing 100876, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2018, 10(7), 1039; https://doi.org/10.3390/rs10071039
Submission received: 1 June 2018 / Revised: 28 June 2018 / Accepted: 29 June 2018 / Published: 2 July 2018
(This article belongs to the Special Issue Pattern Analysis and Recognition in Remote Sensing)

Abstract

:
Infrared image segmentation plays a significant role in many burgeoning applications of remote sensing, such as environmental monitoring, traffic surveillance, air navigation and so on. However, the precision is limited due to the blurred edge, low contrast and intensity inhomogeneity caused by infrared imaging. To overcome these challenges, a level set method using global and local information is proposed in this paper. In our method, a hybrid signed pressure function is constructed by fusing a global term and a local term adaptively. The global term is represented by the global average intensity, which effectively accelerates the evolution when the evolving curve is far away from the object. The local term is represented by a multi-feature-based signed driving force, which accurately guides the curve to approach the real boundary when it is near the object. Then, the two terms are integrated via an adaptive weight matrix calculated based on the range value of each pixel. Under the framework of geodesic active contour model, a new level set formula is obtained by substituting the proposed signed pressure function for the edge stopping function. In addition, a Gaussian convolution is applied to regularize the level set function for the purpose of avoiding the computationally expensive re-initialization. By iteration, the object of interest can be segmented when the level set function converges. Both qualitative and quantitative experiments verify that our method outperforms other state-of-the-art level set methods in terms of accuracy and robustness with the initial contour being set randomly.

Graphical Abstract

1. Introduction

Infrared (IR) imaging system that passively receives IR radiation (760 nm–1 mm) and converts the invisible rays into images has been extensively applied in remote sensing [1,2]. Infrared (IR) image segmentation is one of the most important techniques in IR systems and plays a fundamental role in many modern remote sensing applications, e.g., unmanned aerial vehicle navigation, pedestrian surveillance, space warning and oil leakage detection [3]. However, the accuracy and robustness of IR image segmentation is hard to be guaranteed due to the inherent drawbacks of IR imaging itself, including the blurred edge, low target/background contrast, local inhomogeneity and so on [4]. As a result, it is of great necessity for us to further investigate a robust and precise segmentation method for IR image.
During the past decades, extensive studies about image segmentation have been made and various techniques have been exploited [5,6], among which level set method, also called active contour model (ACM), is one of the most influential approaches. Its main idea is to evolve a curve with certain constraints to extract the target of interest in an image [7]. According to the different types of constraints, the existing ACMs can be classified into three categories: edge-based, region-based and hybrid ACMs [8]. Edge-based ACMs attracts the potential contours towards the boundary of target by virtue of gradient information. Geodesic active contour (GAC) [9] is a typical example that uses an edge-based stopping term to decide the stopping position and a balloon force term to expand or shrink the contour [10]. Later, Melonakos et al. [11] proposed a Finsler ACM by adding directionality in the GAC framework, which proves to be more effective in partitioning roads, blood vessels, neural tracts and other bar-like objects. What is more, a shape-driven variational framework for knowledge-based segmentation was proposed by Paragios et al. [12], in which domain specific knowledge, visual information with shape constraints and user-specific knowledge are integrated. However, the main drawbacks of edge-based ACMs are that they are quite sensitive to noise and the segmentation results are highly dependent on the initialization of contour [13]. More seriously, the resulting boundary is always incomplete when the edge of target is weak or fuzzy, which is the so-called ‘boundary leakage’ problem [14].
Different from the evolutionary mechanism of edge-based ACMs, statistical information is introduced by region-based ACMs to guide the evolving curve to approach the real boundary [8]. Chen-Vese (CV) model [15] is one of the most popular region-based ACMs, which is designed on the basis of Mumford–Shah segmentation techniques [16]. It can achieve satisfactory performances in binary phase segmentation with the assumption that the intensities inside and outside the object boundary are two constants [17]. In other words, CV model is easy to fail when dealing with inhomogeneous images. To overcome this defect, piece-wise smooth (PS) models [18,19] were proposed under the framework of minimizing the Mumford–Shah function, but this kind of methods always suffer from the low running speed caused by the complex procedures involved. Besides, Zhang et al. [20] proposed a selective binary and Gaussian filtering regularized level set (SBGFRLS) model by means of fusing GAC model and CV model. In this model, a region-based signed pressure function (SPF) is constructed and it is utilized to replace the edge stopping function of GAC so that the advantages of GAC and CV can be maintained. However, SBGFRLS does not work well with images with local inhomogeneity and weak boundaries due to the fact that the SPF proposed is only composed of global intensity information. Furthermore, Li et al. [21] proposed a local binary fitting (LBF) model where the neighboring pixels are predicated by the intensity of the current pixel, and it is able to address images with intensity inhomogeneity. Considering the importance of the zero-crossings of image Laplacian for edge detection, Zhang et al. [22] developed an ACM based on image Laplacian fitting energy (ILFE). Aimed to resist the interference of noise and enlarge the capture range, the total variation of image Laplacian is also utilized in the energy function of ILFE. Motivated by clustering theories, Wang et al. [23] applied a local correntropy-based K-means (LCK) clustering into level set image segmentation. Owning to the LCK algorithm, this method is robust to unknown complex noise; as a local method, it also has desirable performance for image with intensity in homogeneity.
Recently, hybrid ACMs, which take full advantages of both global and local information to form the level set formula, have attracted much attention. A representative one is the local and global intensity fitting energy model (LGIF) [24] proposed by Wang et al. LGIF is robust to noise and contour initialization, mainly because it integrates CV model and LBF model via a weighting parameter. Inspired by LGIF, Dong et al. [25] designed an adaptive weighting parameter to fuse the global and local terms more efficiently. Based on the framework of SBGFRLS, Cao et al. [8] added a local term generated by the difference between the averaging filtered image and the original image to the SPF. Under the circumstances, the ACM can be extended to suit the condition where local inhomogeneity exists. On account of the capability of coping with image inhomogeneity and weak boundaries, hybrid ACMs do have advantages in IR image segmentation. Nevertheless, there are still two pivotal aspects to be further considered: (i) the representation of local information; (ii) the weight coefficient that combines the global and local terms.
Despite the efficiency of level set methods, the fully automatic segmentation of the object from the background can hardly be precise without high level knowledge of interest object, especially when the texture information in the image is very complex [26]. Therefore, semi-automatic segmentation algorithms involving user interactions, which is called interactive image segmentation, have been developed and are becoming more and more popular. The objective of interactive image segmentation is to classify the image pixels into fore-and-background classes under the circumstances that some foreground and background markers are provided [27]. Ning et al. [28] proposed a maximal similarity-based region merging (MSRM) mechanism where the initial image is roughly segmented by mean shift algorithm and the region merging process is guided with the help of makers. Inspired by Veksler’s [29] star-convexity prior, Gulshan et al. [30] introduced a multiple stars and Geodesic path-based shape constraint for interactive image segmentation so that the space of possible segmentations to a smaller subset can be restricted. What is more, a promising segmentation framework using synthetic graph coordinates was presented by Panagiotakis et al. [27]. In this method, a min–max Bayesian criterion is minimized and the interactive segmentation problem is solved in two steps considering visual information, proximity distances as well as the given markers, without any requirement of training.
Since most of the conventional level set methods only draw upon the intensity feature, it is easy to get incomplete contours and large quantities of false targets when they are directly applied to process IR images with remarkable intensity inhomogeneity and blurred edges. To solve these challenging problems mentioned above, a new level set method combing global and local information is presented in this paper. First of all, we propose to construct a novel SPF by integrating a global term and a local term together. Specifically speaking, the global term is calculated by the global average intensities inside and outside of the evolving curve. In addition, the local term is constructed by the signed driving force that is associated with four local statistical features: gradient, entropy, standard deviation and filtered difference. Please note that both the direction and magnitude of the driving force are determined by the distances between the feature vector and the two mean multi-feature vectors inside and outside the contour. Then, we use an adaptive weight matrix calculated based on the range value of each pixel to fuse the global and local terms as a hybrid SPF. Next, the level set formula is formed under the framework of GAC model by substituting the edge stopping function with our proposed SPF. Since the presented level set formula takes both global intensity and local multi-features into account, the curve can evolve fast when it is far away from the real boundary meanwhile the evolution is strictly controlled when it is near the object of interest. Besides, a Gaussian convolution is introduced to regularize the level set function so that the re-initialization with high-computational amount can be avoided. By iteration, the curve will evolve to the target boundary until the level set function converges. Figure 1 shows the complete flow chart of our work.
In conclusion, the main contributions and advantages of this study can be summarized as the following aspects:
  • a new SPF integrating both global-intensity-based information and local-multi-feature-based information via an adaptive weight matrix is developed;
  • four statistical features are dynamically weighted by range ratio to form the driving force which is utilized to form the local term of SPF;
  • a level set formula which is able to cope with the weak edge and intensity inhomogeneity is constructed using the SPF proposed;
  • the segmentation results are not obviously influenced by the initialization of contour.
The remainder of this paper is organized as follows: in Section 2, the related work in regard to the GAC framework is introduced; in Section 3, the detailed theories of the proposed level set method are depicted; in Section 4, comparisons with other state-of-the-art methods are conducted to verify the superiority of our method; finally, a conclusion is drawn in Section 5.

2. Related Work about Geodesic Active Contour (GAC) Model

Suppose that Ω 2 is a two-dimensional image domain and I : Ω + is a given single channel IR image. Let C ( p ) = ( x ( p ) , y ( p ) ) ,   p [ 0 , 1 ] be a differentiable parameterized curve in Ω . The GAC model is aimed to minimize the following energy functional:
E GAC ( C ( p ) ) = 0 1 g ( | I ( C ( p ) ) | ) | C ( p ) | d p
where, is a gradient operator; g ( ) is the edge stopping function that can be understood as an edge detection function, and its definition is given as
g ( | I | ) = 1 1 + | G σ I | 2
where, | C ( P ) | represents the arc length of the curve C; G σ I means convolving the IR image I using a two-dimensional Gaussian kernel G σ = 1 2 π σ 2 exp ( x 2 + y 2 2 σ 2 ) whose standard deviation is σ . Obviously, lim x g ( x ) = 0 satisfies the requirement that the edge stopping function gets the minimum when the curve stops on the boundary; on the contrary, Equation (2) tends to be 1 when the curve is on non-edge regions.
By virtue of the variational theory [31], the Euler-Langrange equation of Equation (1) can be obtained as Equation (3).
C t = g ( | I | ) κ N ( g N ) N
where, C t means the derived function of C; κ and N represent the curvature of the contour and the inward normal to the curve, respectively. In addition, a constant velocity term α is employed to increase the propagation speed and Equation (3) can be re-written as:
C t = g ( | I | ) ( κ + α ) N ( g N ) N
Furthermore, by representing the closed curve C with the zero level set of the level set function ϕ , the corresponding level set formula can be deduced as follows:
ϕ t = g | ϕ | ( div ( ϕ | ϕ | ) + α ) + g ϕ
where, α is also called as the balloon force and it controls the contour expanding or shrinking; div ( ϕ | ϕ | ) stands for the curvature term.
Please note that the contour evolution only depends on the edge information, indicating that the segmentation result of GAC model is easy to be affected by noise and fall into a local minimum when the initial contour is far from the desired object [10]. Motivated by this key point, we consider substituting its edge stopping function with a specific term that contains the global and local information simultaneously. By this means, the new level set model is able to be robust to the adverse interferences existing in IR images.

3. Theory

In this section, the theories regarding the proposed level set method are illustrated at length. In Section 3.1, the way of constructing the SPF is introduced; in Section 3.2, we further depict the implement of this method; at last, a summary of the whole procedure is revealed in Section 3.3.

3.1. Signed Pressure Fucntion Integrating Global and Local Information

As is mentioned in Section 2, one of the key techniques in this study is to design a new SPF containing a global term and a local term to replace the edge stopping function under the GAC framework. According to the previous works about SPF [7,8,32], there are two basic design principles to be noticed: (i) the defined SPF should be in the range [ 1 , 1 ] ; (ii) the essence of SPF is to modulate the signs of the pressure forces inside and outside the image region. By the guidance of SPF, the contour expands when it is inside the target of interest and shrinks when it is outside the target to be extracted.

3.1.1. Design of the Global Term

The global term of SPF is supposed to make sense when the evolving curve is far away from the object and it provides a powerful force to drive the curve to approach the real boundary with a fast speed. Based on this consideration, the global term needs to focus on the average intensities inside and outside of the curve [7]. Thus, the following definition is adopted:
s p f G ( x , y ) = I ( x , y ) c 1 + c 2 2 max ( | I ( x , y ) c 1 + c 2 2 | ) , ( x , y ) Ω
where, c 1 and c 2 are two parameters related to the level set function ϕ and they stand for the average intensities inside and outside the contour. Equation (6) modulates the signs of the pressure forces inside and outside the region of interest so that the contour shrinks when outside the object and expands when inside the object. The mathematical expressions of c 1 and c 2 are shown as Equations (7) and (8).
c 1 ( ϕ ) = Ω I ( x , y ) H ( ϕ ) dxdy Ω H ( ϕ ) dxdy
c 2 ( ϕ ) = Ω I ( x , y ) ( 1 H ( ϕ ) ) dxdy Ω ( 1 H ( ϕ ) ) dxdy
where, H ( ) is the Heaviside function whose regularized version is selected as:
H ( x ) = 1 2 ( 1 + 2 π arctan ( x ε 1 ) )
where, arctan ( ) is the inverse tangent function and ε 1 is a control factor. H(x) can be understood as an approximation of the condition: H ( x ) = 1 , if x > 0 ; H ( x ) = 0 , if x < 0 .

3.1.2. Design of the Local Term

To provide a more precise evolution direction when the curve is in the regions with intensity inhomogeneity or weak edge, local features need to be taken into consideration in the level set method. In this paper, four statistical features: gradient υ 1 ( x , y ) , local entropy υ 2 ( x , y ) , local standard deviation υ 3 ( x , y ) and filtered difference υ 4 ( x , y ) , that can reflect the local property of each pixel are selected to compose a feature vector υ ( x , y ) = ( υ 1 ( x , y ) , υ 2 ( x , y ) , υ 3 ( x , y ) , υ 4 ( x , y ) ) T . In addition, each feature is defined as Equations (10)–(13).
υ 1 ( x , y ) = ( I ( x , y ) I ( x + 1 , y ) ) 2 + ( I ( x , y ) I ( x , y 1 ) ) 2
υ 2 ( x , y ) = W m ( x , y ) t = 0 L 1 p ( t ) log 2 p ( t )
υ 3 ( x , y ) = 1 n 1 ( s , t ) W m ( x , y ) ( I ( s , t ) μ ) 2
υ 4 ( x , y ) = | m I ( x , y ) I ( x , y ) |  
where, W m ( x , y ) is an m × m sized local window set around the center pixel ( x , y ) ; L is the maximal gray level of the input IR image ( L = 256 for an 8-bit image); p ( t ) = n t n means the probability of gray level t in W m ( x , y ) , where n t is the number of pixels whose gray level is t and n = m × m is the total pixel number in the local window; μ = 1 n ( s , t ) W m ( x , y ) I ( s , t ) represents the average intensity in W m ( x , y ) ; m is an m × m sized mean filter and m I ( x , y ) means a mean filtering convolution. Here, a further explanation of Equations (10)-(13) is given: υ 1 ( x , y ) is the vector sum of horizontal and vertical gradients, and it reflects the comprehensive gradient of (x, y); υ 2 ( x , y ) and υ 3 ( x , y ) stand for the local entropy and local standard deviation of (x, y) respectively, and they are utilized to measure the local texture of each pixel; υ 4 ( x , y ) is the absolute difference between the original intensity and the intensity processed by mean filtering, and this filtered difference is commonly applied to reflect the local roughness [8].
Next, we argue that the four features in each feature vector should possess different weights according to the degree of inhomogeneity when they are further utilized to calculate the driving force. In light of this consideration, we employ a metric called range ratio ρ i [33] to be the adaptive coefficient, and Equation (14) below shows the definition.
ρ i ( x , y ) = ( υ i max ( x , y ) υ i min ( x , y ) ) / max ( υ i ) ρ max ( x , y ) , i = 1 , 2 , 3 , 4
where, υ i max ( x , y ) and υ i min ( x , y ) denote the maximal and minimal feature values in the local window W m ( x , y ) ; max ( υ i ) is the maximal pixel value of the whole feature map; ρ max ( x , y ) = max ( ρ 1 ( x , y ) , ρ 2 ( x , y ) , ρ 3 ( x , y ) , ρ 4 ( x , y ) ) stands for the maximal range ratio among the four features and it is utilized for normalization. Thus, a feature weight vector ρ ( x , y ) = ( ρ 1 ( x , y ) ρ 2 ( x , y ) ρ 3 ( x , y ) ρ 4 ( x , y ) ) can be obtained for each pixel.
Then, let us focus on studying the construction of driving force which is closely related to the afore-proposed local multi-features. We notice that in LBF model [21], the local intensity information inside and outside the contour is embedded via a Gaussian kernel function used in the energy functional, which is revealed as Equations (15) and (16).
f in ( x , y ) = K σ e ( x , y ) [ M in ( ϕ ( x , y ) ) I ( x , y ) ] K σ e ( x , y ) M in ( ϕ ( x , y ) )
f out ( x , y ) = K σ e ( x , y ) [ M out ( ϕ ( x , y ) ) I ( x , y ) ] K σ e ( x , y ) M out ( ϕ ( x , y ) )
where, f in ( x , y ) and f out ( x , y ) are two values that predict the average intensities inside and outside the contour; K σ e is a Gaussian kernel function whose standard deviation is σ e ; M in ( ϕ ) = 1 H ( ϕ ) and M out ( ϕ ) = H ( ϕ ) , where H ( ) is the Heaviside function introduced in Equation (9). For a further understanding of Equations (15,16), f in ( x , y ) and f out ( x , y ) are weighted averages of the intensities in a neighborhood of (x,y), whose size is proportional to the scale parameter σ e [21].
Referring to LBF model, we also embed gradient, local entropy, local standard deviation and filtered difference to our model with a Gaussian kernel function K σ e [34,35]. For further emphasizing the local property, the mean operator m ( ) that computes the mean value in a m × m sized local window W m ( x , y ) is also used to smooth the embedded local feature for each pixel. By this means, each pixel has two mean multi-feature vectors inside and outside the contour. This procedure is mathematically expressed by Equations (17) and (18) and is intuitively depicted by Figure 2.
ψ in ( x , y ) = ( ψ i n 1 ( x , y ) ψ i n 2 ( x , y ) ψ i n 3 ( x , y ) ψ i n 4 ( x , y ) ) = ( m ( K σ e ( x , y ) [ M i n ( ϕ ( x , y ) ) υ 1 ( x , y ) ] K σ e ( x , y ) M i n ( ϕ ( x , y ) ) ) m ( K σ e ( x , y ) [ M i n ( ϕ ( x , y ) ) υ 2 ( x , y ) ] K σ e ( x , y ) M i n ( ϕ ( x , y ) ) ) m ( K σ e ( x , y ) [ M i n ( ϕ ( x , y ) ) υ 3 ( x , y ) ] K σ e ( x , y ) M i n ( ϕ ( x , y ) ) ) m ( K σ e ( x , y ) [ M i n ( ϕ ( x , y ) ) υ 4 ( x , y ) ] K σ e ( x , y ) M i n ( ϕ ( x , y ) ) ) )
ψ out ( x , y ) = ( ψ o u t 1 ( x , y ) ψ o u t 2 ( x , y ) ψ o u t 3 ( x , y ) ψ o u t 4 ( x , y ) ) = ( m ( K σ e ( x , y ) [ M out ( ϕ ( x , y ) ) υ 1 ( x , y ) ] K σ e ( x , y ) M out ( ϕ ( x , y ) ) ) m ( K σ e ( x , y ) [ M out ( ϕ ( x , y ) ) υ 2 ( x , y ) ] K σ e ( x , y ) M out ( ϕ ( x , y ) ) ) m ( K σ e ( x , y ) [ M out ( ϕ ( x , y ) ) υ 3 ( x , y ) ] K σ e ( x , y ) M out ( ϕ ( x , y ) ) ) m ( K σ e ( x , y ) [ M out ( ϕ ( x , y ) ) υ 4 ( x , y ) ] K σ e ( x , y ) M out ( ϕ ( x , y ) ) ) )
where, ψ in ( x , y ) and ψ out ( x , y ) are the two mean multi-feature vectors inside and outside the contour, and they are composed of four single-feature values denoted as ψ i n i ( x , y ) and ψ o u t i ( x , y )   ( i = 1 , 2 , 3 , 4 ) respectively.
Furthermore, the distances between υ ( x , y ) , ψ in ( x , y ) and ψ out ( x , y ) that measure the similarities are computed using the feature weight vector ρ ( x , y ) as Equations (19) and (20).
L in ( x , y ) = ρ 1 ( x , y ) | υ 1 ( x , y ) ψ i n 1 ( x , y ) | + ρ 2 ( x , y ) | υ 2 ( x , y ) ψ i n 2 ( x , y ) |      + ρ 3 ( x , y ) | υ 3 ( x , y ) ψ i n 3 ( x , y ) | + ρ 4 ( x , y ) | υ 4 ( x , y ) ψ i n 4 ( x , y ) |
L out ( x , y ) = ρ 1 ( x , y ) | υ 1 ( x , y ) ψ o u t 1 ( x , y ) | + ρ 2 ( x , y ) | υ 2 ( x , y ) ψ o u t 2 ( x , y ) |      + ρ 3 ( x , y ) | υ 3 ( x , y ) ψ o u t 3 ( x , y ) | + ρ 4 ( x , y ) | υ 4 ( x , y ) ψ o u t 4 ( x , y ) |
where, | | represents calculating the absolute value. In essence, L in ( x , y ) and L out ( x , y ) are the weighted sums of the four feature distances, and they indicate the evolving direction of the current pixel ( x , y ) . Accordingly, the magnitudes of the driving forces from the two directions can be determined as follows:
F in ( x , y ) = 1 1 + η L in ( x , y )
F out ( x , y ) = 1 1 + η L out ( x , y )
where, F in ( x , y ) , F out ( x , y ) ( 0 , 1 ] denote the internal and external driven forces, respectively; η is a constant parameter.
By comparing L in ( x , y ) and L out ( x , y ) , the driving force s p f L ( x , y ) , which is also regarded as the local term of SPF, are computed as Equation (23).
s p f L ( x , y ) = { + F in ( x , y ) , if   L in ( x , y ) L out ( x , y ) F out ( x , y ) ,   otherwise
Please note that the driving force indicates both the evolution direction and magnitude for each pixel on the curve.

3.1.3. Combination of Global and Local Terms

After computing the global and local terms using the procedures introduced in Section 3.1.1 and Section 3.1.2, we need to fuse the two terms adaptively and construct a complete SPF. That is to say, the SPF can be written as:
spf ( x , y ) = ( 1 ω ( x , y ) ) × s p f G ( x , y ) + ω ( x , y ) × s p f L ( x , y )
where, ω ( x , y ) is the adaptive weighted coefficient of each pixel, and we call ω as the adaptive weight matrix in this paper. Equation (24) indicates that whether the SPF in our method is dominated by the global information or the local information is determined by ω .
As is emphasized above, the intensities in the regions far away from the object vary slowly. As a result, it is unsuitable and unnecessary to use local features to describe these image blocks, i.e., the weight of s p f G ( x , y ) should increase in these regions. On the contrary, the intensities tend to change drastically, and c 1 as well as c 2 computed by Equations (7) and (8) may deviate from the real average intensities inside and outside the contour when the curve is in the regions near the boundary of object. In this case, the weight of s p f L ( x , y ) should increase accordingly. To sum up, the global term plays a dominant role in the smooth regions while the local term becomes extremely significant in the regions with remarkable intensity inhomogeneity.
Similar to the concept of range ratio used in Equation (14), we propose to calculate the range matrix ζ that measures the local roughness as Equation (25).
ζ ( x , y ) = max ( W m ( x , y ) ) min ( W m ( x , y ) )
where, max ( ) and min ( ) means computing the maximum and minimum in W m ( x , y ) . To some degree, ζ ( x , y ) can be understood as the local contrast of (x,y).
Next, we need to find an appropriate mapping function Φ to establish a mapping relationship between the range matrix ζ and the adaptive weight matrix ω . Two basic requirements that the mapping function needs to meet are reported below:
  • it should be a non-negative and monotonically increasing function;
  • lim ζ ζ min Φ ( ζ ) = 0 and lim ζ ζ max Φ ( ζ ) = 1 , where ζ max and ζ min denote the maximum and minimum in the range matrix ζ .
Inspired by the previous work about IR small target enhancement [36], we fortunately find that the standard sigmoid function whose mathematical form is Sg ( x ) = 1 1 + e x can meet the above two points (see Figure 3a). However, we notice that there is a transition interval in Sg(x) that occupies a large percentage of the whole interval. We argue that the width of this transition interval should be strictly controlled, i.e., ζ 1 corresponding to the central point A and ζ 2 corresponding to the lower inflection point B should be carefully selected (see Figure 3b), because once the relatively smooth area with a small ζ is dominated by the local term of SPF, the evolution of curve will be extremely slow and is easy to drop into local minima. Thus, some modifications in regard to the scale and phase terms are necessary. Suppose that the final mapping function based on the standard sigmoid function is expressed as follows:
Φ ( ζ ) = Sg ( a ζ + b ) = 1 1 + e ( a ζ + b )
where, a and b represent the scale and phase parameters, respectively. As is shown in Figure 3b, these two parameters are determined by the two key points A = ( ζ 1 , 1 2 ) and B = ( ζ 2 , ε 2 ) , where ε 2 is a constant close to the lower bound 0. By substituting A and B into Equation (26),
{ a = 1 ζ 2 ζ 1 ln ε 2 1 ε 2 b = ζ 1 ζ 2 ζ 1 ln ε 2 1 ε 2
where, ζ 1 = λ max ( ζ ) and ζ 2 = 0.7 × ζ 1 in this paper; λ ( 0 , 1 ) is a constant which is set by experience.
By this stage, the adaptive weight matrix can be worked out as ω ( x , y ) = Φ ( ζ ( x , y ) ) and the final SPF can be constructed as Equation (24).

3.2. Implementation

Although we have addressed the problem of constructing the SPF fusing global and local information, there are still three aspects that should be noticed in the implementation of the whole algorithm: (i) the initialization of level set function; (ii) the construction of level set formula; (iii) the evolution of level set function.

3.2.1. Initialization of Level Set Function

In our method, the level set function is initialized with a random positive constant using the following formulation:
ϕ ( ( x , y ) , t = 0 ) = { τ , ( x , y ) Ω 0 Ω 0   0 , ( x , y ) Ω 0 τ , ( x , y ) Ω - Ω 0
where, τ is a positive constant which is set randomly; Ω 0 is a subset of the image domain Ω 0 and Ω 0 denotes the boundary of Ω 0 .
With regards to Ω 0 , it can be set randomly when the level set function is partly controlled by a global term [20]. That is to say, the initial location, shape and size of Ω 0 will not influence the curve evolution obviously in our method. For convenience, we set Ω 0 as an s 0 × s 0 sized square which is located in the center of the input image for all the experiments.
In Section 4, we will further demonstrate that the initialization of the level set function does not influence the segmentation results by experiments.

3.2.2. Construction of Level Set Formula

In this paper, the edge stopping function of the GAC model is replaced by the proposed hybrid SPF for the purpose of dealing with intensity inhomogeneity and blurred edge. Substituting the edge stop function g ( ) in Equation (5) with the proposed SPF, our level set formula can be expressed using the following formulation:
ϕ t = spf | ϕ | ( div ( ϕ | ϕ | ) + α ) + s p f ϕ
In Equation (29), the curvature-based term | ϕ | div ( ϕ | ϕ | ) is utilized to regularize the level set function ϕ . Since ϕ is a signed distance function that satisfies the condition | ϕ | = 1 , the regularized term can be expressed by a Laplacian of ϕ . As pointed out in the theory of scale-space [37], the evolution of a function with its Laplacian equals to a Gaussian kernel filtering the initial condition of the function [7]. Hence, a Gaussian filter K σ r whose standard deviation is σ r can be used to regularize ϕ in this study. In this case, the regularize term div ( ϕ | ϕ | ) can be removed from Equation (29). Also, the term spf ϕ is unnecessary because the proposed method uses the statistical information of regions that has a large capture range and capacity of anti-edge leakage [7,8]. Lastly, the level set formula of our method can be written as:
ϕ t = spf α | ϕ | = ( ( 1 ω ) s p f G + ω s p f L ) α | ϕ |

3.2.3. Evolution of Level Set Function

Since the level set formula ϕ t can be regarded as the derivative between ϕ t and ϕ t 1 , the level set function ϕ t in the t-th iteration can be updated as:
ϕ t = ϕ t 1 + ϕ t
After that, let ϕ t = 1 if ϕ t 0 ; otherwise, ϕ t = 1 . According to Zhang’s comment [7], this step is alternative, i.e., it is necessary if we want to selectively segment the desired object; otherwise, it is unnecessary.
Moreover, as introduced in Section 3.2.2, a Gaussian kernel filtering is employed to regularize the new level set function ϕ t as:
ϕ t = ϕ t   K σ r
where, K σ r is the Gaussian kernel utilized whose standard deviation is σ r ; the kernel size is 2 × round (   2 × σ r   )   +   1 , where round ( ) is a rounding operator.
By iteration, the evolution of ϕ t can be regarded as convergent when the current ϕ t meets the following condition:
ϕ t ϕ t 1 F ϕ t 1 F < δ
where, means calculating the Frobenius norm, δ is a convergence threshold. In fact, Equation (33) indicates that the level set function does not change obviously any further.
Finally, the zero-level set ϕ ( ( x , y ) , t ) = 0 is selected as the resulting contour Ω .

3.3. Summary of the Proposed Method

To highlight the complete implementation of our method, the main procedures are summarized in Algorithm 1 below according to all the afore-introduced theories.
Algorithm 1 Level set method using global and local information
Input: an IR image I
1. Initialization: initialize the level set function ϕ to be a binary function using Equation (28).
2. While not convergence do
3. for each pixel do
4. Calculate the global term s p f G of SPF using Equation (6)
5. Calculate the local term s p f L of SPF using Equation (23)
6. Calculate the adaptive weight matrix ζ using Equation (25)
7. Combing s p f G and s p f L to construct SPF using Equation (24)
8. end
9. Construct the level set formulation ϕ t according to Equation (30)
10. Evolve ϕ t = ϕ t 1 + ϕ t according to Equation (31)
11. Let ϕ t = 1 , if ϕ t 0 ; otherwise, ϕ t = 1
12. Regularize ϕ t using a Gaussian kernel function K σ r according to Equation (32)
13. if ϕ t ϕ t 1 F ϕ t 1 F < δ
14. break
15. end if
16. end
Output: The resulting contour Ω : ϕ ( ( x , y ) , t ) = 0 .

4. Experiment and Discussion

In this section, experiments based on real IR test images and relevant discussions are made. In Section 4.1, the value settings of all the parameters utilized in our method are introduced at first; in Section 4.2, comparisons in terms of segmentation results, segmentation precision, running efficiency and influence of contour initialization are made successively. All the experiments are implemented using Matlab 2012b on a PC with a 2.60 GHz INTEL CPU and 4.0 GB installed memory (RAM).

4.1. Parameter Setting

Here, all the parameters involved, as well as the corresponding meanings and their default values are listed in Table 1.

4.1.1. Parameter Description

According to the previous analysis [21], if ε 1 is too small, the energy functional would fall into a local minimum meanwhile the final contour location may also drift from the ground truth if ε 1 is too large. Thus, ε 1 is set as 1.5 which has been proved to be an appropriate choice [7]. The local windows used for computing local entropy, standard deviation, filtered difference, range ratio and the range matrix are uniformly fixed as m × m sized. The window size is usually 5 × 5 to 9 × 9 , and we set m = 5 in the simulations. Besides, the standard deviation of the Gaussian filter used for embedding local features is set as σ e = 3.0 which is referred to the related analysis made by Li et al. [21]. Through their analysis based on the results for different values of σ e , this parameter has little influence on the segmentation result. η and λ are two unique parameters used in our study. It can be seen from Equations (21) and (22) that η affects the magnitude of the driving force to some extent, and through our repeated experiments, η = 1.5 can achieve a robust and satisfactory performance. λ decides the central point of Φ ( x ) and also determines the width of the transition interval of Φ ( x ) indirectly. Strictly speaking, λ varies for different images. In this paper, we set λ = 0.3 as a default value which is selected by experience. More analyses about η and λ are presented in Section 4.1.2. ε 2 is a constant that represents the value of the lower inflection point of Φ ( x ) , and it will not obviously affect the output of Φ ( x ) as long as it is close to 0 (usually smaller than 0.0001) [38]. The balloon force α is a crucial parameter that directly determines the precision of segmentation and the convergence rate of the level set function. A larger α leads to a fast convergence, but it may also generate leakage of boundary. Vice versa. To keep a balance, we set α = 400 in this study. The constant τ used for contour initialization only needs to meet the requirement that τ > 0 , so we set τ = 1 according to Zhang’s choice [7]. s 0 denotes the size of initial contour Ω 0 and it is set as s 0 = 7 in our experiments. As a matter of fact, this parameter has little influence on the segmentation results, which is demonstrated in Section 4.2.4. What is more, the standard deviation σ r of the Gaussian filter used for regularizing level set function is also quite important. On the one hand, if the standard deviation is too small, the presented model will be sensitive to noise; on the other hand, false extraction of object and leakage of boundary are easy to happen if it is too large. The empirical value of σ r is in the range [1.5, 3.0], and it is set as r = 2.5 in our model. Lastly, the convergence threshold δ is used as a stop condition of the curve evolution and Equation (33) indicates that the level set function ϕ t tends to be constant at t-th iteration. Based on this consideration, δ should be a constant close to 0 and we find that the resulting curve has no remarkable changes any more when the rate of change of ϕ is smaller than 3%.

4.1.2. Sensitivity Analysis

In this section, we focus on discussing the sensitivities of two significant parameters η and λ so as to obtain their suitable settings. As mentioned in Section 4.1.2, the default values of other parameter are selected according to the previous studies, so we do not discuss them again.
Here, F value [39,40] is employed as a measuring metric for sensitivity analysis. The define of F value is given as follows:
F = ( 1 + γ 2 ) P R γ 2 P + R   ( if   P = R = 0 ,   F = 0 )
where, γ is a harmonic coefficient and γ = 1 in this paper; P and R denote the precision rate and recall rate respectively, which are defined as:
P = A C A R
R = A C A S
where, A R represents the area extracted by the level set method; A S represents the object region provided by the ground truth; A C represents the correctly segmented region, i.e., A C = A R A S . For an intuitive illustration, Figure 4 provides a schematic diagram to show relationship between A R , A S and A C . Obviously, a larger F value indicates a more accuracy segmentation result, and F is in the range [ 0 , 1 ] .
We consider that an appropriate parameter setting of η or λ need to get higher F values in most of the practical IR scenes. Here, 6 typical IR images shown in Figure 5 are utilized to make the sensitivity analysis, in which we continuously change the values of η or λ while other parameters are fixed with the default values listed in Table 1.
A statistic of F values with different η is implemented. η changes from 0.1 to 5 (the step length is chosen as 0.1) and the presented algorithm is conducted to produce a resulting segmentation image for each η , after which the corresponding F value is calculated. For λ , the same procedure is adopted to obtain the statistic of F values. In our simulation, λ varies from 0.01 to 1 and the step length is set as 0.02. By this means, we can get 50 groups of data for each parameter and the following two cartograms drawn in Figure 6 reveal the sensitivities of η and λ respectively.
As for η , it does influence the segmentation accuracy significantly since the η -sensitivity curves shown in Figure 6a undergo striking fluctuations in most scenes. However, we find that the F values can keep in high levels when η is in the range ( 0.75 , 2 ) and the statistical results are relatively stable. To this end, we adopt η = 1.5 for all the experiments.
It can be seen from Figure 6b that the λ -sensitivity curves increase rapidly before λ 0.1 and level off after λ 0.4 . As far as we are concerned, the level set function is dominated by the local term when λ < 0.1 while it is dominated by the global term when λ > 0.4 . To keep a balance, we set λ = 0.3 in this paper.
We would like to state that the specific values of the parameters involved in this study may still need to be adjusted in different cases so as to get more accuracy segmentation results (especially η , λ and σ r ), and the values provided in Section 4.1 are just default values.

4.2. Comparative Experiment

In this part, 6 state-of-the-arts level set methods: GAC [9], CV [15], SBGFRLS [20], LBF [21], ILFE [22], Cao’s model [8] and an interactive segmentation method: MSRM [28] that all have been discussed in Section 1 are selected to make both qualitative and quantitative comparisons with the method proposed in this paper. In Section 4.2.1, the level-set-based algorithms are tested with a uniform initial contour while MSRM is conducted with the object and background markers being added artificially. Then, two metrics indicating the segmentation precision are employed to evaluate the segmentation results in Section 4.2.2. Next, comparisons in terms of running time and iterations are made in Section 4.2.3. Finally, an additional experiment to test the influence of contour initialization is implemented in Section 4.2.4.

4.2.1. Segmentation Result

In this section, apart from 30 IR images (part of them can be downloaded from Ref. [41]), 5 natural images [42] and 5 remote sensing images [43] are also adopted to be the database in order to verify that the proposed algorithm is general.
Figure 7, Figure 8 and Figure 9 report the segmentation results of IR images, natural images and remote sensing images respectively. The IR images utilized can be classified into two categories: single-object-images and multiple-object-images, and they all contain blurred object boundaries and a degree of local intensity inhomogeneity. In particular, test images, like IR.10 and IR.11 in Figure 7a, are contaminated by noises and the disturbances of false alarm (e.g., edges of the cart and laboratory furniture) are quite obvious. The target/background contrasts of the natural images and remote sensing images are higher and the details, especially the edges and textures, are richer.
For a fair comparison, we use a uniform initial contour to define the binary level set function for each test image, which can be seen from the blue rectangle in the first columns of Figure 7, Figure 8 and Figure 9. Also, the initial object and background markers of MSRM, which are marked in green and blue, are shown in the second columns. In addition, the 3–8 columns correspond to the segmentation results of GAC, CV, SBGFRLS, LBF, ILFE, Cao’s model, MSRM and our method, respectively. Please note that the final evolving curves are marked in red.
As is introduced above, GAC model is highly dependent on the initialization of contour, and it can only get complete segmentation results when the initial contour contains the whole object of interest. In this comparative experiment, the initial contour is set in the image center and most of the objects are partly covered. As a result, the contours evolved by GAC model all converge to local minima, which leads to the seriously incomplete segmentations. More seriously, the curve totally disappears if the initial contour is far away from the object (see IR. (20, 23) in Figure 7c and RS. 3 in Figure 9c).
Chen-Vese (CV) model draws upon the statistical intensity information to evolve the curve and is able to achieve satisfactory performances in binary phase segmentation. Basically speaking, all the objects are extracted by CV model, but boundary leakage occurs in those inhomogeneous areas due to the fact that this model is established with the assumption that image intensities in each region always maintain constant. For example, as can be seen from IR. (1, 7, 11, 14) in Figure 7d, some striking holes on the clothes are mistakenly generated in the final segmentation results. Moreover, CV model is quite sensitive to the disturbance of noise, which can be demonstrated by the fake targets on the cart and ground in IR. (10, 11) in Figure 7d. Besides, we find that the convergence of CV model is quite unstable, e.g., IR. (18, 20, 26, 29) in Figure 7d, which we infer may be associated with the contour initialization.
In regard to SBGFRLS model, it also does well with IR images with homogeneous intensities, e.g., IR. (18, 20, 26, 29) in Figure 7e and natural images, but its performance turns to be bad when dealing with inhomogeneous cases. From our experiments, we can clearly find that incompleteness of segmentation always occurs near the borders between the object and the ground, the cloth parts of humans in IR images (see IR. (3, 5–8, 14) in Figure 7e) and the lands in remote sensing images (see RS. (1, 2, 4) in Figure 9e), because the intensity distributions of these regions are quite uneven, which causes severe interferences for the expansion or shrink of the contour.
Additionally, it is obvious that LBF and ILFE get a large number of false targets when coping with images with rich texture information. On the one hand, many edges belonging to the background are recognized as the real boundaries in IR. (10, 11) in Figure 7f and Figure 7g; on the other hand, the real object is divided into too many blocks, which absolutely cannot satisfy the visual requirement of image segmentation. It should be noticed that the initial contour has an outstanding influence on the results of ILFE, since there is a remarkable false target with a relatively large area around the position of the initial contour in IR. (20, 21, 26, 27) in Figure 7g.
Cao’s model takes full advantages of global and local intensity information, so it can successfully address the weak edges and local intensity inhomogeneities. As a whole, Cao’s method is able to extract the objects in all the test images, although there are still some incompleteness existing in the final contours, especially in the feet and clothes parts, which can be seen from IR. (3, 5–8, 13, 14) in Figure 7h. We notice that its performances in remote sensing images are not satisfactory enough and there are several false segmentations existing in RS. (1, 2) in Figure 9h, while the segmentation results are almost perfect in natural images. From our point of view, this is mainly because that the local term of Cao’ model that only takes intensity information into account is still not robust to complex texture patterns.
Maximal similarity-based region merging (MSRM) method is an interactive image segmentation that guides the region merging with the aid of two kinds of markers. Although the fore-and-background markers are correctly provided, we notice that the object boundaries obtained still deviate from the ground truth in some cases, especially in IR images, such as IR. (1–3, 10–11, 18, 27) in Figure 7i. Intuitively, those inaccurate boundaries are lack of smoothness and many background image patches are mistakenly classified into the foreground class. In our opinion, there are two main reasons that lead to the segmentation errors: (i) the rough segmentation using mean shift or super-pixel is always inaccurate in gray images; (ii) IR images are lack of color and texture information, so the similarities between each region are not convincing.
Fortunately, our method obtains the most complete and precise segmentation results among the 8 algorithms owing to its strategy of integrating the average intensity-based global information and multi-feature-based local information. As can be seen from IR. (1–3, 28) in Figure 7j, the feet are perfectly picked out and the problem of boundary leakage does not occur in our method. Also, the interferences of noise and edge in IR. (10, 11) in Figure 7j do not take any effects on the final results. Thus, we argue that the superiority of the proposed method is proved by this comparative experiment. On the other hand, our method is somewhat weak at dealing with tiny objects, e.g., the ox horn in N. 2 in Figure 8j and the small circle-shape land in RS. 5 in Figure 9j are lost, and the fluorescent lamp in IR. 18 is mistakenly recognized as the object.
Last but not least, all these methods, except GAC model, is capable of partitioning multiple objects in a single image simultaneously, which can be verified by the results shown in IR. (25–30) in Figure 7 and N. (1–3) in Figure 8.

4.2.2. Comparison of Segmentation Accuracy

To further prove the superiority of our method, we would like to compare the precision of segmentation results through a quantitative way. First of all, the segmentation results shown in Figure 7, Figure 8 and Figure 9 are transferred to binary maps according to the corresponding final level set functions. Figure 10, Figure 11 and Figure 12 display the binary maps as well as the ground truth, respectively.
As is introduced in Equation (34), F value is extensively employed to measure the segmentation accuracy in the field of object detection. Here, Table 2 lists the corresponding F values of all the binary segmentation maps in Figure 10, Figure 11 and Figure 12 and the average F value of each algorithm is given in the last row. The top two F values in each group are marked with bold values.
Apart from F values that indicate the overlapping rate, another metric called boundary precision (BP) that evaluates ability of precise boundary locating [27] is also applied to make quantitative comparisons in this section. Below, the definition of BP is given as:
B P = 1 / u min v d ( u , v ) + v min u d ( v , u ) # u + # v
where, u and v are the pixels on the result boundary and ground-truth boundary; # u and # v denote the number of pixels on the corresponding boundaries; for each pixel u, min v d ( u , v ) means the minimal distance between u and the pixels on the ground-truth boundary; similarly, for each pixel v, min u d ( v , u ) means the minimal distance between v and the pixels on the result boundary. Obviously, a higher BP refers to a more accuracy boundary locating.
Figure 13, Figure 14 and Figure 15 reveal the boundary maps extracted by Sobel algorithm [44] and the ground-truth boundaries. Table 3 reports the BP value of each test group clearly.
As is revealed from Table 2, our method achieves the largest F values on average (96.04%) while GAC model always gets the worst F values in all the test images. GAC model is an edge-based level set method, so its segmentation results are poor if the initial contour cannot totally contain the real object. In our case, the initial contour is set as a rectangle located in the image center which partly covers the object. Hence, the segmentation precision is doomed to be very low. LBF model and ILFE model excessively concentrate on the local intensity feature, but they ignore the global intensity instead. Although this strategy can extract the details in a cautious way, it may result in too many segmented blocks in the object region and cause large quantities of false targets in the background. That is the main reason their F values basically maintain around 50–60%. SBGFRLS model is able to partition the objects well, but it cannot do in addressing the details perfectly. When the evolving curve approaches the boundary, the further evolution becomes inaccuracy due to the fact that the regions near the boundary are always inhomogeneous, which does not accord with the assumption of this model. That is to say, the average intensities inside and outside the contour estimated by Equations (7) and (8) are distorted. As a result, it is difficult for SBGFRLS to achieve F values greater than 95%. CV model’s robustness to local intensity inhomogeneities is stronger than SBGFRLS model, but the it cannot avoid causing false segmentations inside the object. This drawback is reflected by those minor holes on the human bodies. Also, the disturbances of noise decrease the F values of CV model to some extent. The F values of Cao’s model are around 92% which are approximately on the same level of SBGFRLS model, but the average value is decreased by its weak performance in a small part of the test images, such as IR. (18, 26) in Figure 10. To further analyze, we consider that it may be the inaccuracy of its weight function used to integrate the global and local terms that results in the striking boundary leakage, indicating that its robustness needs to be improved further. With the help of makers, MSRM achieves relatively high F values, even higher than our method in certain test images. We think it is quite reasonable because the additional makers do provide much useful information for MSRM to distinguish between objects and backgrounds. The average F value of our method is about 4.5% higher than the second place (SBGFRLS), and our method gets the top largest F values in almost all the database. Through observing the binary segmentation results, we find that our method possesses an outstanding ability to overcome the problem of boundary leakage and it is much less sensitive to the noise, which guarantees its high segmentation precision. However, we still need to point out that the segmentation errors of the proposed method are mainly derived from the miss of groove parts (e.g., the inner thigh).
The overall distribution of BP value depicted in Table 3 is almost the same as F value, i.e., our proposed level set method generally outperforms other comparing algorithms in terms of boundary precision (0.8575 on average) while the BPs of GAC model are far smaller than others (only 0.0219 on average). This phenomenon matches the statistical result of F values quite well, proving that the conclusions derived from these two metrics are both convincing. It is interesting that MSRM often gets the top two best BPs among the IR image database, but SBGFRLS models takes the second place on average. This phenomenon indicates that SBGFRLS model is more robust than MSRM method and it can keep in a relatively high boundary locating level as well. CV, LBF and ILFE absolutely do not have strong boundary locating abilities according to Table 3. For CV model, the weak and unstable convergence decreases it overall BPs to a great extent while the large quantities of false boundaries caused by LBF and ILFE models weaken their BPs sharply. The performance of Cao’s model is moderate among all the 8 algorithms. There are no severe segmentation errors, but the BPs cannot be improved further since the leakage of boundary is still not solved.

4.2.3. Comparison of Running Time

As is commonly known, running speed is also a significant factor in evaluating an algorithm. In this section, we test the number of iterations (except MSRM) and the execution time respectively, which are listed in Table 4 and Table 5.
With respect to the number of iterations, it can be clearly observed from Table 4 that our method has the fewest iterations in most of the test images (fewer than 50 times), which demonstrates its best convergence among all the other methods. We consider that its outstanding convergence is mainly owing to the fact that the average intensity-based global term accelerates the evolving speed meanwhile the local multi-feature-based term avoids the energy functional falls into local minima. SBGFRLS, LBF and Cao’s models also achieve relatively satisfactory iteration times which maintain around 50 times on the whole. However, we notice that the convergences of SBGFRLS model and Cao’s method become poor when dealing with IR. 10 that contains several edge disturbances. This phenomenon indicates that these two models would still drop into local minima when facing the complex scenes. When compared with the afore-discussed methods, CV and ILFE models turn to be relatively weak in iterations. Since CV model does not contain any local intensity information, the evolution of curve may slow down in inhomogeneous regions. For ILFE model, the Laplacian fitting term may have an adverse effect on the convergence also. Lastly, GAC model is the worst one whose iterations are more than 10 times than the proposed method. As has been discussed above, its curve evolution is easy to stagnate around the initial contour if the initialization itself is set inappropriately.
When considering the whole execution time, it can be seen from Table 5 that those level set methods that only involve global intensity information, e.g., GAC, CV and ILFE, get relatively fast running speeds. Among them, although GAC and CV suffer from large numbers of iterations, their whole running time is still shorter than SBGFRLS and Cao’s model owing to its high efficiency of the single iteration. The total running time of LBF is also fine, but it is because of its few iterations, rather than the single iteration efficiency. MSRM gets the fastest execution speed among all the 8 algorithms, because this method does not have the problem of convergence and the user’s interaction time is not included in our test. Our method achieves the moderate performance in this comparison, meaning that the running efficiency is not the best. As far as we are concerned, it is the procedures of calculating the multi-features and constructing the driving forces inside and outside the contour that increase the running time. Thus, it is of great necessity for us to adopt the parallel processors, e.g., FPGA and GPU, and optimize the codes further.

4.2.4. Influence of Contour Initialization

As is extensively acknowledged that an excellent level set method needs to be robust to the initialization of contour. Based on this consideration, an additional experiment aimed to test the influence of contour initialization on the final segmentation results is implemented in this section. As is shown in Figure 16 and Figure 17, two groups of IR images (called ‘man’ and ‘plane’) are used and the initial contour is set as four different conditions. The corresponding binary maps and boundary maps are presented in Figure 18, Figure 19, Figure 20 and Figure 21. Also, the F values and BP values are listed in Table 6 and Table 7.
Here, three key points: position, size and shape, are considered by us. For both Figure 16 and Figure 17, the initial positions of contours can be classified as inside the object, outside the object, partly inside the object; the initial sizes of contours can be classified as 7 × 7 and 15 × 15 ; the initial shapes can be classified as square and circle.
Intuitively speaking, the initialization of contour has little effect on SBGFRLS, Cao’s model and the proposed method according to our experiment. As is reported in Table 6 and Table 7, the Fs and BPs of our method and Cao’s model are still superior to other comparing algorithms while those of SBGFRLS are moderate. In general, the utilized metrics of the three methods do not undergo remarkable fluctuations. By contrast, the segmentation results of GAC model are seriously affected by the initial contours. On the one hand, it is highly possible for the contour to converge towards the position where its initial condition locates; besides, it is even possible to be absorbed completely if the initial contour has no overlaps with the object. Although the performances of LBF and ILFE still keep in a low level in terms of the two metrics, we still notice that different shapes of initial curve will result in different evolving curves. As is shown in contour 2, 3 of Figure 17, the head part of ‘man’ image will be segmented into fewer blocks by LBF and a striking false contour around the head part will be generated by ILFE when the initial contour is circle. In addition, as we can see in contour 2, 3 of Figure 16 and contour 2, 4 of Figure 17, the curves of CV model stop evolving if it is initialized in a region suffering from intensity inhomogeneity seriously. Accordingly, its Fs and BPs both occupy wide ranges, varying from 0.34–0.96 and 0.06–0.97 respectively.

5. Conclusions

5.1. Methdology

In this paper, a new level set method fusing average-intensity-based global information and multi-feature-based local information is proposed to segment IR images. Different from the conventional level set methods that only consider single intensity feature, the presented model constructs a hybrid SPF made up of a global term and a local term. The global term is calculated based on the average intensities inside and outside the contour while the local term is represented by the driving force computed by four weighted local features. To keep a balance, the two terms are combined via an adaptive weight matrix. By substituting the edge stopping function in GAC model with the proposed SPF, the level set formula is constructed and the level set function is re-initialized by a Gaussian filtering. By iteration, the final contour of object can be obtained and the object is thus extracted. Both qualitative and quantitative experiments verify that our method outperforms other state-of-the-art level set methods in terms of segmentation accuracy, convergence and robustness to contour initialization.

5.2. Applicability and Limitation to Remote Sensing

IR imaging belongs to one of the significant branches of remote sensing and IR image segmentation is extensively applied in many remote sensing applications, e.g., intelligent urban surveillance, unmanned aerial vehicles, pedestrian detection, etc. Although the proposed level set method is aimed to cope with the blurred boundary, low contrast and intensity inhomogeneity in IR image, it can also be directly applied to process remote sensing images when they are converted to single channel images, which has been demonstrated in Section 4.2.1. However, the color and spectrum information of remote sensing images are not fully used when they are processed using our method, indicating that the inherent advantages are ignored. On the other hand, we find that remote sensing images usually contain rich texture information, but the proposed algorithm is somewhat weak at coping with the tiny objects, especially the small circular holes existing in the land. We infer it is the Gaussian filtering used for regularizing the level set function that removes these details.

5.3. Future Work

In the future work, we plan to further optimize the codes of our algorithm and try to transplant the Matlab codes to parallel hardware, e.g., FPGA and GPU, so that the whole computational time can be reduced greatly and the real-time running can be expected. Besides, we notice that the target of interest in IR image can be seen as a salient object, so its visual saliency may be exploited for further investigation.

Author Contributions

M.W. conceived of, designed and performed the algorithm, analyzed the data and wrote the paper; G.G., J.S. and X.M. are the research supervisors; K.R. helped modify the language; W.Q. and Q.C. also provided technical assistance to the research; The manuscript was discussed by all the co-authors.

Funding

This paper is financially supported by the National Natural Science Foundation of China (61675099), the National Natural Science Foundation of China (61701233), Open Foundation of State Key Laboratory of Networking and Switching Technology (Beijing University of Posts and Telecommunications) (SKLNST-2016-2-07), Canada Research Program, and China Scholarship Council.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Huang, J.; Ma, Y.; Zhang, Y.; Fan, F. Infrared image enhancement algorithm based on adaptive histogram segmentation. Appl. Opt. 2017, 56, 9686–9697. [Google Scholar] [CrossRef] [PubMed]
  2. Zingoni, A.; Diani, M.; Corsini, G. A Flexible Algorithm for Detecting Challenging Moving Objects in Real-Time within IR Video Sequences. Remote Sens. 2017, 9, 1128. [Google Scholar] [CrossRef]
  3. Lei, S.; Zou, Z.; Liu, D.; Xia, Z.; Shi, Z. Sea-land segmentation for infrared remote sensing images based on superpixels and multi-scale features. Infrared Phys. Technol. 2018, 91, 12–17. [Google Scholar] [CrossRef]
  4. Wan, M.; Gu, G.; Qian, W.; Ren, K.; Chen, Q.; Maldague, X. Particle swarm optimization-based local entropy weighted histogram equalization for infrared image enhancement. Infrared Phys. Technol. 2018, 91, 164–181. [Google Scholar] [CrossRef]
  5. Pal, N.R.; Pal, S.K. A review on image segmentation techniques. Pattern Recognit. 1993, 26, 1277–1294. [Google Scholar] [CrossRef]
  6. Lee, L.K.; Liew, S.C.; Thong, W.J. A review of image segmentation methodologies in medical image. In Advanced Computer and Communication Engineering Technology; Springer: New York, NY, USA, 2015; pp. 1069–1080. [Google Scholar]
  7. Niu, S.; Chen, Q.; de Sisternes, L.; Ji, Z.; Zhou, Z.; Rubin, D.L. Robust noise region-based active contour model via local similarity factor for image segmentation. Pattern Recognit. 2017, 61, 104–119. [Google Scholar] [CrossRef]
  8. Cao, J.; Wu, X. A novel level set method for image segmentation by combining local and global information. J. Mod. Opt. 2017, 64, 2399–2412. [Google Scholar] [CrossRef]
  9. Caselles, V.; Kimmel, R.; Sapiro, G. Geodesic active contours. Int. J. Comput. Vis. 1997, 22, 61–79. [Google Scholar] [CrossRef]
  10. Tian, Y.; Duan, F.; Zhou, M.; Wu, Z. Active contour model combining region and edge information. Mach. Vis. Appl. 2013, 24, 47–61. [Google Scholar] [CrossRef]
  11. Melonakos, J.; Pichon, E.; Angenent, S.; Tannenbaum, A. Finsler active contours. IEEE Trans. Pattern Anal. Mach. Intell. 2008, 30, 412–423. [Google Scholar] [CrossRef] [PubMed]
  12. Paragios, N. Variational methods and partial differential equations in cardiac image analysis. In Proceedings of the IEEE International Symposium on Biomedical Imaging: Nano to Macro, Arlington, VA, USA, 15–18 April 2004; pp. 17–20. [Google Scholar]
  13. Wang, D.; Zhang, T.; Yan, L. Fast hybrid fitting energy-based active contour model for target detection. Chin. Opt. Lett. 2011, 9, 071001-071001. [Google Scholar] [CrossRef]
  14. Zhao, W.; Xianze, X.; Zhu, Y.; Xu, F. Active contour model based on local and global Gaussian fitting energy for medical image segmentation. Opt. Int. J. Light Electron. Opt. 2018, 158, 1160–1169. [Google Scholar] [CrossRef]
  15. Chan, T.F.; Vese, L.A. Active contours without edges. IEEE Trans. Image Process. 2011, 10, 266–277. [Google Scholar] [CrossRef] [PubMed]
  16. Mumford, D.; Shah, J. Optimal approximations by piecewise smooth functions and associated variational problems. Commun. Pure Appl. Math. 1989, 42, 577–685. [Google Scholar] [CrossRef] [Green Version]
  17. Wang, Y.; Huang, T.Z.; Wang, H. Region-based active contours with cosine fitting energy for image segmentation. J. Opt. Soc. Am. A Opt. Image Sci. Vis. 2015, 32, 2237–2246. [Google Scholar] [CrossRef] [PubMed]
  18. Tsai, A.; Yezzi, A.; Willsky, A.S. Curve evolution implementation of the Mumford-Shah functional for image segmentation, denoising, interpolation, and magnification. IEEE Trans. Image Process. 2001, 10, 1169–1186. [Google Scholar] [CrossRef] [PubMed]
  19. Vese, L.A.; Chan, T.F. A multiphase level set framework for image segmentation using the Mumford and Shah model. Int. J. Comput. Vis. 2002, 50, 271–293. [Google Scholar] [CrossRef]
  20. Zhang, K.; Zhang, L.; Song, H.; Zhou, W. Active contours with selective local or global segmentation: A new formulation and level set method. Image Vis. Comput. 2010, 28, 668–676. [Google Scholar] [CrossRef]
  21. Li, C.; Kao, C.Y.; Gore, J.C.; Ding, Z. Minimization of region-scalable fitting energy for image segmentation. IEEE Trans. Image Process. 2008, 17, 1940–1949. [Google Scholar] [PubMed]
  22. Zhang, K.; Xu, S.; Zhou, W.; Liu, B. Active contours based on image Laplacian fitting energy. Chin. J. Electron. 2009, 18, 281–284. [Google Scholar]
  23. Wang, L.; Pan, C. Robust level set image segmentation via a local correntropy-based K-means clustering. Pattern Recognit. 2014, 47, 1917–1925. [Google Scholar] [CrossRef]
  24. Wang, L.; Li, C.; Sun, Q.; Xia, D.; Kao, C.Y. Active contours driven by local and global intensity fitting energy with application to brain MR image segmentation. Comput. Med. Imaging Graph. 2009, 33, 520–531. [Google Scholar] [CrossRef] [PubMed]
  25. Dong, F.; Chen, Z.; Wang, J. A new level set method for inhomogeneous image segmentation. Image Vis. Comput. 2013, 31, 809–822. [Google Scholar] [CrossRef]
  26. Zhao, Y.; Nie, X.; Duan, Y.; Huang, Y.; Luo, S. A benchmark for interactive image segmentation algorithms. In Proceedings of the IEEE Workshop on Person-Oriented Vision (POV), Kona, HI, USA, 7 January 2010; pp. 33–38. [Google Scholar]
  27. Panagiotakis, C.; Papadakis, H.; Grinias, E.; Komodakis, N.; Fragopoulou, P.; Tziritas, G. Interactive image segmentation based on synthetic graph coordinates. Pattern Recognit. 2013, 46, 2940–2952. [Google Scholar] [CrossRef]
  28. Ning, J.; Zhang, L.; Zhang, D.; Wu, C. Interactive image segmentation by maximal similarity based region merging. Pattern Recognit. 2010, 43, 445–456. [Google Scholar] [CrossRef]
  29. Veksler, O. Star shape prior for graph-cut image segmentation. In Proceedings of the European Conference on Computer Vision, Marseille, France, 12–18 October 2008; pp. 454–467. [Google Scholar]
  30. Gulshan, V.; Rother, C.; Criminisi, A.; Blake, A.; Zisserman, A. Geodesic star convexity for interactive image segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), San Francisco, CA, USA, 13–18 June 2010; pp. 3129–3136. [Google Scholar]
  31. Zhi, X.H.; Shen, H.B. Saliency driven region-edge-based top down level set evolution reveals the asynchronous focus in image segmentation. Pattern Recognit. 2018, 80, 241–255. [Google Scholar] [CrossRef]
  32. Xu, C.; Yezzi, A.; Prince, J.L. On the relationship between parametric and geometric active contours. In Proceedings of the IEEE Conference Record of the Thirty-Fourth Asilomar Conference on Signals, Systems and Computers, Pacific Grove, CA, USA, 29 October–1 November 2000; pp. 483–489. [Google Scholar]
  33. Yu, Y.; Zhang, C.; Wei, Y.; Li, X. Active contour method combining local fitting energy and global fitting energy dynamically. In Proceedings of the International Conference on Medical Biometrics, Hong Kong, China, 28–30 June 2010; pp. 163–172. [Google Scholar]
  34. Wan, M.; Gu, G.; Qian, W.; Ren, K.; Chen, Q. Hybrid active contour model based on edge gradients and regional multi-features for infrared image segmentation. Opt. Int. J. Light Electron. Opt. 2017, 140, 833–842. [Google Scholar] [CrossRef]
  35. Zhang, T.; Han, J.; Zhang, Y.; Bai, L. An adaptive multi-feature segmentation model for infrared image. Opt. Rev. 2016, 23, 220–230. [Google Scholar] [CrossRef]
  36. Wan, M.; Gu, G.; Qian, W.; Ren, K.; Chen, Q. Infrared small target enhancement: Grey level mapping based on improved sigmoid transformation and saliency histogram. J. Mod. Opt. 2018, 65, 1161–1179. [Google Scholar] [CrossRef]
  37. Perona, P.; Malik, J. Scale-space and edge detection using anisotropic diffusion. IEEE Trans. Pattern Anal. Mach. Intell. 1990, 12, 629–639. [Google Scholar] [CrossRef] [Green Version]
  38. Wan, M.; Gu, G.; Qian, W.; Ren, K.; Chen, Q.; Maldague, X. Infrared Image Enhancement Using Adaptive Histogram Partition and Brightness Correction. Remote Sens. 2018, 10. [Google Scholar] [CrossRef]
  39. Mangale, S.A.; Khambete, M.B. Approach for moving object detection using visible spectrum and thermal infrared imaging. J. Electron. Imaging 2018, 27. [Google Scholar] [CrossRef]
  40. Peng, H.; Li, B.; Ling, H.; Hu, W.; Xiong, W.; Maybank, S.J. Salient object detection via structured matrix decomposition. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 818–832. [Google Scholar] [CrossRef] [PubMed]
  41. Database Collection of Infrared Image. Available online: http://www.dgp.toronto.edu/nmorris/IR/ (accessed on 16 June 2018).
  42. MSRA10K Salient Object Database. Available online: http://mmcheng.net/msra10k/ (accessed on 19 June 2018).
  43. Gu, Y.; Ren, K.; Wang, P.; Gu, G. Polynomial fitting-based shape matching algorithm for multi-sensors remote sensing images. Infrared Phys. Technol. 2016, 76, 386–392. [Google Scholar] [CrossRef]
  44. Gao, W.; Zhang, X.; Yang, L.; Liu, H. An improved Sobel edge detection. In Proceedings of the 2010 3rd IEEE International Conference on Computer Science and Information Technology (ICCSIT), Chengdu, China, 9–11 July 2010; pp. 67–71. [Google Scholar]
Figure 1. Flow chart of the proposed method (IR is the acronym of infrared).
Figure 1. Flow chart of the proposed method (IR is the acronym of infrared).
Remotesensing 10 01039 g001
Figure 2. Schematic diagram of calculating ψ i n ( x , y ) and ψ o u t ( x , y ) , where ψ i n ( x , y ) and ψ o u t ( x , y ) are mean multi-feature vectors inside and outside the contour.
Figure 2. Schematic diagram of calculating ψ i n ( x , y ) and ψ o u t ( x , y ) , where ψ i n ( x , y ) and ψ o u t ( x , y ) are mean multi-feature vectors inside and outside the contour.
Remotesensing 10 01039 g002
Figure 3. Schematic diagram of the mapping function: (a) the standard sigmoid function; (b) the modified sigmoid function.
Figure 3. Schematic diagram of the mapping function: (a) the standard sigmoid function; (b) the modified sigmoid function.
Remotesensing 10 01039 g003
Figure 4. Schematic diagram in terms of the relationship between A R , A S and A C ,where A R , A S and A C represent the extracted region, the ground-truth object region and the correctly segmented region.
Figure 4. Schematic diagram in terms of the relationship between A R , A S and A C ,where A R , A S and A C represent the extracted region, the ground-truth object region and the correctly segmented region.
Remotesensing 10 01039 g004
Figure 5. IR images used to make sensitivity analysis: (ad) humans; (e,f) vehicles.
Figure 5. IR images used to make sensitivity analysis: (ad) humans; (e,f) vehicles.
Remotesensing 10 01039 g005
Figure 6. Sensitivity curves: (a) η -curves; (b) λ -curves.
Figure 6. Sensitivity curves: (a) η -curves; (b) λ -curves.
Remotesensing 10 01039 g006
Figure 7. Segmentation results of IR image database: (a) initial contour of level set methods; (b) initial makers of MSRM; (c) GAC; (d) CV; (e) SBGFRLS; (f) LBF; (g) ILFE; (h) Cao’s model; (i) MSRM; (j) the proposed method.
Figure 7. Segmentation results of IR image database: (a) initial contour of level set methods; (b) initial makers of MSRM; (c) GAC; (d) CV; (e) SBGFRLS; (f) LBF; (g) ILFE; (h) Cao’s model; (i) MSRM; (j) the proposed method.
Remotesensing 10 01039 g007aRemotesensing 10 01039 g007b
Figure 8. Segmentation results of natural image database: (a) initial contour of level set methods; (b) initial makers of MSRM; (c) GAC; (d) CV; (e) SBGFRLS; (f) LBF; (g) ILFE; (h) Cao’s model; (i) MSRM; (j) the proposed method.
Figure 8. Segmentation results of natural image database: (a) initial contour of level set methods; (b) initial makers of MSRM; (c) GAC; (d) CV; (e) SBGFRLS; (f) LBF; (g) ILFE; (h) Cao’s model; (i) MSRM; (j) the proposed method.
Remotesensing 10 01039 g008
Figure 9. Segmentation results of remote sensing image database: (a) initial contour of level set methods; (b) initial makers of MSRM; (c) GAC; (d) CV; (e) SBGFRLS; (f) LBF; (g) ILFE; (h) Cao’s model; (i) MSRM; (j) the proposed method.
Figure 9. Segmentation results of remote sensing image database: (a) initial contour of level set methods; (b) initial makers of MSRM; (c) GAC; (d) CV; (e) SBGFRLS; (f) LBF; (g) ILFE; (h) Cao’s model; (i) MSRM; (j) the proposed method.
Remotesensing 10 01039 g009
Figure 10. Binary segmentation results of IR image database: (a) GAC; (b) CV; (c) SBGFRLS; (d) LBF; (e) ILFE; (f) Cao’s model; (g) MSRM; (h) the proposed method; (i) ground truth.
Figure 10. Binary segmentation results of IR image database: (a) GAC; (b) CV; (c) SBGFRLS; (d) LBF; (e) ILFE; (f) Cao’s model; (g) MSRM; (h) the proposed method; (i) ground truth.
Remotesensing 10 01039 g010aRemotesensing 10 01039 g010b
Figure 11. Binary segmentation results of natural image database: (a) GAC; (b) CV; (c) SBGFRLS; (d) LBF; (e) ILFE; (f) Cao’s model; (g) MSRM; (h) the proposed method; (i) ground truth.
Figure 11. Binary segmentation results of natural image database: (a) GAC; (b) CV; (c) SBGFRLS; (d) LBF; (e) ILFE; (f) Cao’s model; (g) MSRM; (h) the proposed method; (i) ground truth.
Remotesensing 10 01039 g011
Figure 12. Binary segmentation results of remote sensing image database: (a) GAC; (b) CV; (c) SBGFRLS; (d) LBF; (e) ILFE; (f) Cao’s model; (g) MSRM; (h) the proposed method; (i) ground truth.
Figure 12. Binary segmentation results of remote sensing image database: (a) GAC; (b) CV; (c) SBGFRLS; (d) LBF; (e) ILFE; (f) Cao’s model; (g) MSRM; (h) the proposed method; (i) ground truth.
Remotesensing 10 01039 g012
Figure 13. Boundary maps of IR image database: (a) GAC; (b) CV; (c) SBGFRLS; (d) LBF; (e) ILFE; (f) Cao’s model; (g) MSRM; (h) the proposed method; (i) ground truth.
Figure 13. Boundary maps of IR image database: (a) GAC; (b) CV; (c) SBGFRLS; (d) LBF; (e) ILFE; (f) Cao’s model; (g) MSRM; (h) the proposed method; (i) ground truth.
Remotesensing 10 01039 g013aRemotesensing 10 01039 g013b
Figure 14. Boundary maps of natural image database: (a) GAC; (b) CV; (c) SBGFRLS; (d) LBF; (e) ILFE; (f) Cao’s model; (g) MSRM; (h) the proposed method; (i) ground truth.
Figure 14. Boundary maps of natural image database: (a) GAC; (b) CV; (c) SBGFRLS; (d) LBF; (e) ILFE; (f) Cao’s model; (g) MSRM; (h) the proposed method; (i) ground truth.
Remotesensing 10 01039 g014
Figure 15. Boundary maps of remote sensing image database: (a) GAC; (b) CV; (c) SBGFRLS; (d) LBF; (e) ILFE; (f) Cao’s model; (g) MSRM; (h) the proposed method; (i) ground truth.
Figure 15. Boundary maps of remote sensing image database: (a) GAC; (b) CV; (c) SBGFRLS; (d) LBF; (e) ILFE; (f) Cao’s model; (g) MSRM; (h) the proposed method; (i) ground truth.
Remotesensing 10 01039 g015
Figure 16. Segmentation results for ‘man’ image with different initial contours: (a) initial contour; (b) GAC; (c) CV; (d) SBGFRLS; (e) LBF; (f) ILFE; (g) Cao’s model; (h) the proposed method.
Figure 16. Segmentation results for ‘man’ image with different initial contours: (a) initial contour; (b) GAC; (c) CV; (d) SBGFRLS; (e) LBF; (f) ILFE; (g) Cao’s model; (h) the proposed method.
Remotesensing 10 01039 g016
Figure 17. Segmentation results of ‘plane’ image with different initial contours: (a) initial contour; (b) GAC; (c) CV; (d) SBGFRLS; (e) LBF; (f) ILFE; (g) Cao’s model; (h) the proposed method.
Figure 17. Segmentation results of ‘plane’ image with different initial contours: (a) initial contour; (b) GAC; (c) CV; (d) SBGFRLS; (e) LBF; (f) ILFE; (g) Cao’s model; (h) the proposed method.
Remotesensing 10 01039 g017
Figure 18. Binary segmentation results of ‘man’ image: (a) ground truth; (b) GAC; (c) CV; (d) SBGFRLS; (e) LBF; (f) ILFE; (g) Cao’s model; (h) the proposed method.
Figure 18. Binary segmentation results of ‘man’ image: (a) ground truth; (b) GAC; (c) CV; (d) SBGFRLS; (e) LBF; (f) ILFE; (g) Cao’s model; (h) the proposed method.
Remotesensing 10 01039 g018
Figure 19. Binary segmentation results of ‘plane’ image: (a) ground truth; (b) GAC; (c) CV; (d) SBGFRLS; (e) LBF; (f) ILFE; (g) Cao’s model; (h) the proposed method.
Figure 19. Binary segmentation results of ‘plane’ image: (a) ground truth; (b) GAC; (c) CV; (d) SBGFRLS; (e) LBF; (f) ILFE; (g) Cao’s model; (h) the proposed method.
Remotesensing 10 01039 g019
Figure 20. Boundary maps of ‘man’ image: (a) ground truth; (b) GAC; (c) CV; (d) SBGFRLS; (e) LBF; (f) ILFE; (g) Cao’s model; (h) the proposed method.
Figure 20. Boundary maps of ‘man’ image: (a) ground truth; (b) GAC; (c) CV; (d) SBGFRLS; (e) LBF; (f) ILFE; (g) Cao’s model; (h) the proposed method.
Remotesensing 10 01039 g020
Figure 21. Boundary maps of ‘plane’ image: (a) ground truth; (b) GAC; (c) CV; (d) SBGFRLS; (e) LBF; (f) ILFE; (g) Cao’s model; (h) the proposed method.
Figure 21. Boundary maps of ‘plane’ image: (a) ground truth; (b) GAC; (c) CV; (d) SBGFRLS; (e) LBF; (f) ILFE; (g) Cao’s model; (h) the proposed method.
Remotesensing 10 01039 g021
Table 1. Parameter settings.
Table 1. Parameter settings.
ParameterMeaningDefault Value
ε 1 The control factor of Heaviside function1.5
mThe length of local window5
σ e The standard deviation of the Gaussian filter used for embedding local features3.0
η Help to determine the magnitude of driving force1.5
λ Help to determine the central point of Φ ( x ) 0.3
ε 2 The value of the lower inflection point of Φ ( x ) 0.00001
α The balloon force of level set formula400
τ The constant utilized for initializing the contour1
s 0 The side length of initial contour Ω 0 7
σ r The standard deviation of the Gaussian filter used for regularizing level set function2.5
δ The convergence threshold0.03
Table 2. Statistic for F values, and bold values indicate the top two results.
Table 2. Statistic for F values, and bold values indicate the top two results.
GACCVSBGFRLSLBFILFECao’s ModelMSRMOurs
IR. 100.97580.93780.36510.50690.93780.87420.9859
IR. 20.02350.93590.94990.65820.81050.96240.78790.9664
IR. 30.01430.93610.88860.78260.81010.88830.85380.9595
IR. 40.01500.93100.95250.74790.74750.95170.86160.9837
IR. 50.00290.96430.93760.27450.55360.94200.97940.9778
IR. 60.00270.95220.90540.43770.65200.90200.96670.9683
IR. 70.00130.95370.91570.43070.59860.91450.94060.9796
IR. 800.95300.93110.72990.81180.92400.97290.9617
IR. 90.00400.93510.97870.49460.71720.97450.98380.9876
IR. 100.00730.94410.94450.65710.69410.94300.50160.9545
IR. 110.00740.96430.96410.60660.64940.96620.83570.9783
IR. 120.00930.95480.94580.64590.74110.94550.97400.9759
IR. 130.00670.89410.87250.59460.72640.87170.96320.9695
IR. 140.00310.90790.88430.51220.76320.88500.95250.9647
IR. 1500.71260.85980.58190.61970.86110.96600.9790
IR. 160.00700.92710.94620.47110.74250.94680.96210.9679
IR. 170.05560.74980.93060.68560.70730.93550.94150.9731
IR. 180.02330.36380.77320.65720.75260.78610.68570.9296
IR. 190.32220.53990.89200.79510.71140.88200.91910.9179
IR. 2000.09100.85940.76920.78020.84660.90410.9128
IR. 210.15850.79050.89540.60740.88350.89540.91230.9637
IR. 220.00180.95640.96770.87120.93070.96770.95420.9642
IR. 2300.95430.95970.17820.46650.96060.96620.9640
IR. 240.01000.90070.90140.55210.69320.90160.92330.9315
IR. 250.03040.46130.89210.69780.76360.90060.96960.9716
IR. 260.00070.32770.80630.60820.74160.82150.92050.9264
IR. 2700.95520.93470.65100.73940.93720.85570.9699
IR. 280.00320.86890.92370.62640.76550.92430.89900.9668
IR. 290.01570.17010.89960.57130.75290.88440.59530.8500
IR. 300.02580.92570.89240.57820.57840.90400.81500.9262
N. 10.03070.86720.92040.43250.26670.94100.82880.9382
N. 200.83910.95430.12760.14670.94880.91220.9730
N. 30.01110.97650.96230.61920.68130.96430.77270.9809
N. 40.10460.89230.91090.48690.07070.90100.91650.9730
N. 50.00030.84690.89490.27740.10220.88750.74380.9486
RS. 100.92070.94180.63870.66570.94690.64000.9908
RS. 20.02300.95240.97160.64220.60540.97280.90380.9914
RS. 300.98830.99170.02070.07460.99020.92880.9784
RS. 40.03930.87560.89780.19950.14500.88290.91570.9500
RS. 50.02170.90890.97370.18040.03850.95270.83620.9639
Ave.0.02460.8241 0.91910.53660.60520.91880.87590.9604
Table 3. Statistic for BP values, and bold values indicate the top two results.
Table 3. Statistic for BP values, and bold values indicate the top two results.
GACCVSBGFRLSLBFILFECao’s ModelMSRMOurs
IR. 10.01540.61830.52310.14100.15030.43270.17730.5480
IR. 20.01350.51110.43650.17640.09740.37290.10800.4943
IR. 30.01560.70620.30870.18000.12330.24120.17570.5311
IR. 40.01460.24560.71380.15710.12710.70320.15031.7023
IR. 50.00990.12540.09360.08250.07830.10310.53720.7152
IR. 60.01000.20170.13110.14060.10480.12420.65000.3344
IR. 70.00980.10920.12120.18230.10400.11780.09310.8206
IR. 80.00420.00310.00270.00310.00310.00280.00300.0030
IR. 90.01250.26340.86670.11150.08450.48811.15051.4123
IR. 100.00840.24270.32690.04510.02590.36740.01740.1727
IR. 110.00780.16170.21440.05830.04740.23240.08250.5685
IR. 120.01420.43870.41300.14290.11080.44770.58570.9073
IR. 130.00690.12730.08050.06500.04930.08380.35300.4652
IR. 140.00620.12260.10900.08730.06290.11920.38070.4451
IR. 150.01260.06640.08530.04550.04200.09780.96671.4751
IR. 160.01220.22210.31300.11240.07510.29960.56610.6610
IR. 170.02770.05850.12020.17580.08700.17260.70981.5204
IR. 180.03490.06670.06110.15030.07280.05310.13800.3384
IR. 190.21210.09591.05070.18000.11490.99211.27941.1345
IR. 200.00440.01180.77930.48770.05420.68081.02941.0094
IR. 210.07740.17250.55600.22050.28270.55430.63951.3674
IR. 220.01210.93421.31070.34410.36031.31940.88481.1345
IR. 230.00370.21300.24520.08700.04800.27280.33090.3090
IR. 240.01620.19910.20920.11520.04620.20950.29040.3092
IR. 250.01360.03410.08910.08900.07420.04861.21921.2336
IR. 260.02010.06600.40770.31860.34130.44970.95180.9851
IR. 270.00730.39890.25210.10710.09780.33540.14010.3453
IR. 280.00860.12270.32410.15490.13750.31940.18430.8926
IR. 290.01310.03391.11190.38990.17620.92040.18880.7586
IR. 300.00871.11680.47390.18040.15120.51110.35421.0653
N. 10.01770.22450.31510.15770.15230.43330.16430.5183
N. 20.01170.14710.50650.04390.05430.40160.25640.7673
N. 30.00980.59400.47390.07450.08360.57260.08280.8730
N. 40.03590.21890.63990.32460.14250.38820.58341.8264
N. 50.02120.28180.49130.29280.16150.39160.16590.8487
RS. 10.02010.10510.06250.03640.04970.06900.03841.3818
RS. 20.03300.07820.09810.03180.04480.09570.09831.2781
RS. 30.00801.02871.70800.13650.06310.33650.23900.6780
RS. 40.04240.14030.38620.05660.08840.26160.45650.7772
RS. 50.04110.15051.31860.06620.11920.45680.26341.6900
Ave.0.02190.26650.44450.14880.10720.36200.41720.8575
Table 4. Comparison of iterations.
Table 4. Comparison of iterations.
GACCVSBGFRLSLBFILFECao’s ModelOurs
IR. 132018030301604020
IR. 238045040401005035
IR. 33602004035603535
IR. 48601003040462929
IR. 51802767351484933
IR. 6440744746825133
IR. 73807264731065945
IR. 876012055601055530
IR. 94001563135852926
IR. 10320180120305510550
IR. 112401204530604030
IR. 12700855754996031
IR. 1338037779603047546
IR. 1476016179501677951
IR. 1588024142651234430
IR. 163207627401952826
IR. 17280303235322922
IR. 185608840351325831
IR. 19740671714651812
IR. 202402066832386734
IR. 21500441818261715
IR. 22520701450551611
IR. 232008259452164330
IR. 247601803030803525
IR. 25400888665719386
IR. 263602363260213029
IR. 273602456035756035
IR. 283402607634837947
IR. 29240845833405650
IR. 304403003520803530
N. 128013773301273735
N. 2122016938801233233
N. 3400747770714540
N. 42409122202672210
N. 56606719291481717
RS. 1560753365474923
RS. 2260682362512321
RS. 3200302235882313
RS. 4480872427361812
RS. 53001191645571915
Average4281384642984431
Table 5. Comparison of running time (in seconds).
Table 5. Comparison of running time (in seconds).
GACCVSBGFRLSLBFILFECao’s ModelMSRMOurs
IR. 114.364849.398017.009213.756445.384516.41977.066110.7698
IR. 216.6241126.091523.267717.191328.365225.99036.704220.2464
IR. 315.594556.224223.865115.742018.231121.50478.052422.0967
IR. 428.063316.854520.310626.202713.492820.02037.375721.8969
IR. 511.14175.791149.773534.647445.823836.35027.125229.3925
IR. 627.221915.454236.401041.309427.0629144.80978.190729.4851
IR. 722.556715.267851.677457.465333.844652.27119.439039.0489
IR. 848.896935.3429144.518734.712033.441759.20405.923919.7322
IR. 912.434127.572420.358923.650325.097521.93266.790419.9093
IR. 1050.378962.399188.566647.449825.842387.02535.944366.6338
IR. 1117.462036.123632.634022.758422.211529.09429.187226.1399
IR. 1226.763115.277337.959934.860529.427341.18117.109523.7949
IR. 1349.543978.680858.008559.9297109.017758.350712.376653.3658
IR. 1485.109933.057767.869847.255860.914883.710611.725564.2202
IR. 1532.689941.077327.525842.423936.237230.45359.298623.7023
IR. 1612.392513.140717.977529.690656.714820.21248.682322.3312
IR. 178.41715.538620.420123.43449.528819.955326.714016.4456
IR. 1813.204414.714526.975221.695135.522136.36467.960320.7848
IR. 1913.602511.079612.69648.610817.950014.77057.33848.1545
IR. 207.127536.079243.618621.181910.590041.88495.661423.6258
IR. 219.96997.159712.860710.97777.262112.98747.691910.1184
IR. 2214.885411.508510.981830.212415.775813.18666.17688.5429
IR. 2315.159117.023143.590636.888671.179231.342311.620728.8746
IR. 2432.292750.882517.029214.304325.111221.25677.513815.4799
IR. 2519.322916.678456.373545.426122.476963.89599.883277.5913
IR. 269.286641.663620.017535.77796.139420.607510.253420.4785
IR. 2741.800078.145063.311324.106430.182866.83437.550836.1605
IR. 2832.414452.753694.080433.141831.773199.904312.581557.2206
IR. 2915.415318.539739.673333.017213.258239.881810.089141.6004
IR. 3019.168586.141323.013610.904024.322286.576410.212519.0091
N. 112.016224.199947.315022.551637.542330.56397.553827.1024
N. 252.593030.440526.306957.992239.527924.54098.285225.5377
N. 325.339014.979156.831251.261223.013833.99077.924451.3699
N. 46.474015.302316.148913.738472.260216.52466.26698.3545
N. 515.645710.927813.928917.850140.695913.83785.3266811.8827
RS. 117.222212.965522.388848.872514.064632.29075.572917.2348
RS. 26.108011.744416.490744.503214.925116.53884.330414.4676
RS. 36.21356.538015.246822.721924.880418.25014.936410.2431
RS. 410.332915.003916.798118.546510.250713.01255.33398.7737
RS. 56.088022.158613.153230.906016.213114.72766.073310.4180
Average22.033430.998036.174430.691730.638938.30647.846126.5559
Table 6. Statistic of F values (‘man’ and ‘plane’ image), and bold values indicate the top two results.
Table 6. Statistic of F values (‘man’ and ‘plane’ image), and bold values indicate the top two results.
GACCVSBGFRLSLBFILFECao’s ModelOurs
‘man5’Contour100.96780.96430.58300.69100.96720.9715
Contour20.05590.51800.96320.34350.27640.96430.9709
Contour30.01050.34730.96360.35900.25980.96410.9687
Contour40.01130.95310.96450.54060.69120.96470.9704
‘plane2’Contour10.06290.53940.89700.59720.60430.90510.9136
Contour20.19770.91940.89500.46970.07710.90400.9150
Contour300.50270.89410.41010.07650.90550.9145
Contour40.05060.90490.89480.59020.61360.90540.9143
Table 7. Statistic of BP values (‘man’ and ‘plane’ image), and bold values indicate the top two results.
Table 7. Statistic of BP values (‘man’ and ‘plane’ image), and bold values indicate the top two results.
GACCVSBGFRLSLBFILFECao’s ModelOurs
‘man5’Contour10.00390.46850.36160.08630.07120.44720.5272
Contour20.00930.03400.31200.09360.08290.32810.5227
Contour30.00660.01690.31220.09300.08450.32880.4809
Contour40.00950.31050.37150.08580.08500.44650.5221
‘plane2’Contour10.01280.08650.88640.18510.15890.90590.8704
Contour20.01300.97600.85450.23790.13370.88780.9905
Contour30.01310.06500.87320.23230.13380.88040.8864
Contour40.01320.90100.85120.18130.16250.85590.8688

Share and Cite

MDPI and ACS Style

Wan, M.; Gu, G.; Sun, J.; Qian, W.; Ren, K.; Chen, Q.; Maldague, X. A Level Set Method for Infrared Image Segmentation Using Global and Local Information. Remote Sens. 2018, 10, 1039. https://doi.org/10.3390/rs10071039

AMA Style

Wan M, Gu G, Sun J, Qian W, Ren K, Chen Q, Maldague X. A Level Set Method for Infrared Image Segmentation Using Global and Local Information. Remote Sensing. 2018; 10(7):1039. https://doi.org/10.3390/rs10071039

Chicago/Turabian Style

Wan, Minjie, Guohua Gu, Jianhong Sun, Weixian Qian, Kan Ren, Qian Chen, and Xavier Maldague. 2018. "A Level Set Method for Infrared Image Segmentation Using Global and Local Information" Remote Sensing 10, no. 7: 1039. https://doi.org/10.3390/rs10071039

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop