Next Article in Journal
New N-Methylimidazole-Functionalized Chitosan Derivatives: Hemocompatibility and Antibacterial Properties
Next Article in Special Issue
Binary Sand Cat Swarm Optimization Algorithm for Wrapper Feature Selection on Biological Data
Previous Article in Journal
Silica-Containing Biomimetic Composites Based on Sea Urchin Skeleton and Polycalcium Organyl Silsesquioxane
Previous Article in Special Issue
Multi-Agent Variational Approach for Robotics: A Bio-Inspired Perspective
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Criterion Sampling Matting Algorithm via Gaussian Process

1
Guizhou Key Laboratory of Pattern Recognition and Intelligent System, Guizhou Minzu University, Guiyang 550025, China
2
School of Computer Science, Zhongshan Institute, University of Electronic Science and Technology of China, Zhongshan 528400, China
3
Key Laboratory of Big Data and Intelligent Robot (SCUT), MOE of China, School of Software Engineering, South China University of Technology, Guangzhou 510006, China
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Biomimetics 2023, 8(3), 301; https://doi.org/10.3390/biomimetics8030301
Submission received: 1 June 2023 / Revised: 30 June 2023 / Accepted: 3 July 2023 / Published: 10 July 2023
(This article belongs to the Special Issue Nature-Inspired Computer Algorithms)

Abstract

:
Natural image matting is an essential technique for image processing that enables various applications, such as image synthesis, video editing, and target tracking. However, the existing image matting methods may fail to produce satisfactory results when computing resources are limited. Sampling-based methods can reduce the dimensionality of the decision space and, therefore, reduce computational resources by employing different sampling strategies. While these approaches reduce computational consumption, they may miss an optimal pixel pair when the number of available high-quality pixel pairs is limited. To address this shortcoming, we propose a novel multi-criterion sampling strategy that avoids missing high-quality pixel pairs by incorporating multi-range pixel pair sampling and a high-quality sample selection method. This strategy is employed to develop a multi-criterion matting algorithm via Gaussian process, which searches for the optimal pixel pair by using the Gaussian process fitting model instead of solving the original pixel pair objective function. The experimental results demonstrate that our proposed algorithm outperformed other methods, even with 1% computing resources, and achieved alpha matte results comparable to those yielded by the state-of-the-art optimization algorithms.

1. Introduction

Image matting is a crucial image processing technique with extensive applications in image synthesis [1,2], video editing [3,4], live broadcasting [5,6], and film special effects [7,8]. In image matting, alpha mattes can accurately extract foreground objects and merge them with new backgrounds to render new scenes [9,10]. The concept of the alpha matte was first proposed by Thomas in 1984 [11], whereby an alpha matte estimation model was constructed by introducing the alpha channel. Mathematically, the color value I p , including the RGB, of a pixel p can be linearly represented by the foreground color F p and the background color B p in the original image, as shown in Equation (1):
I p = α p F p + ( 1 α p ) B p  
where α p is the alpha matte of the foreground object at pixel p , and α p [ 0 , 1 ] . Specifically, α p takes 1 when p belongs to the foreground, it is assigned a value of 0 when p belongs to the background, and assigned a specific value in the range (0,1) when p is a semi-transparent pixel of which the color is a mixture of a foreground pixel color and a background pixel color. As both F p and B p are three-dimensional unknown vectors and α p is an unknown scalar, Equation (1) is an ill-defined problem. In order to accurately determine the value of α , Rhemann et al. [12] introduced a trimap, which divides the image into three non-overlapping regions, denoted as the known foreground, known background, and unknown regions, to impose additional constraints on the image matting problem, as shown in Figure 1 (where F is the foreground region with α = 1 , B is the background region with α = 0 , and U is the unknown region with α in the range (0, 1)).
Pixel pair optimization-based matting methods are a class of competitive image matting approaches that offer significant advantages in terms of parallelization [13]. These methods are particularly effective in processing an error-marked trimap [14] or a foreground that is spatially disconnected [15,16,17]. Essentially, the natural image matting problem is transformed into a pixel pair optimization problem, as shown in Equation (2).
min f ( x p ) s . t . x i = ( x i ( F ) , x i ( B ) ) T       p U ,   x i ( F ) F ,   x i ( B ) B
where f ( x p ) is the pixel pair evaluation function of unknown pixel p ; U , F , and B are the pixel sets of the unknown, known foreground, and known background regions, respectively; x p is the pixel pair decision vector of pixel p ; x i F and x i B represent pixels in the known foreground and background regions, respectively. Once the foreground and background colors are obtained by solving the pixel pair optimization problem, the alpha value of the pixel p can be estimated via the following expression:
α ^ p = ( I p B p ) ( F p B p ) F p B p 2  
where * 2 denotes the Euclidean norm of vector .
Pixel pair optimization-based matting methods can be further divided into sampling-based methods and evolutionary optimization-based matting methods. Sampling-based methods [18,19] to evaluate the alpha value of the unknown region by collecting the known foreground and background pixels as candidate samples, which narrows the search range through different sampling strategies. For example, Liang et al. [20] proposed a surrogate model based on natural image matting that effectively reduces computational resource consumption by building on top of random sampling. However, this approach also faces the issue of the potential loss of optimal pixel pairs due to the random sampling strategy. The main drawback of these sampling-based methods stems from the significant likelihood of high-quality pixel pair loss, resulting in unsatisfactory alpha mattes. To avoid this issue, He et al. [21] proposed a global sampling method that employs all pixel pairs as candidate samples to avoid the loss of high-quality pixel pairs. However, this approach also results in increased computational resource consumption. Then, the evolutionary optimization-based [22,23] methods were proposed, which can effectively mitigate the drawback of high-quality pixel pair loss in sampling-based methods and improve the quality of alpha mattes. For instance, Liang et al. [24] used the evolutionary algorithm instead of the pixel pair sampling process and proposed a multi-objective evolutionary algorithm based on multi-criteria decomposition. This algorithm utilizes all available computing resources by adjusting the number of iterations of the evolutionary algorithm and theoretically eliminates the risk of missing real samples. On the other hand, to address the problem of computing resource consumption, Liang et al. [25] developed a multi-scale evolutionary pixel pair optimization framework, which transforms the large-scale pixel-pair optimization problem into multiple sub-optimization problems of different scales by using an image pyramid. Although evolutionary optimization-based methods utilizing various evolutionary algorithms have improved the accuracy of the alpha matte, it may require thousands of iterations to find the optimal solution. Limited computing resources may restrict the applicability of this method and also compromise the quality of alpha mattes. In summary, neither the sampling-based method nor the existing evolutionary optimization-based method can provide satisfactory alpha mattes under limited computing resources. Regarding image matting tasks, there has been less discussion on matting under limited computing resources.
These limitations have motivated the present study, in which we designed a multi-criterion sampling strategy (MCSS) to ensure that high-quality pixel pairs are sampled. Furthermore, to reduce the consumption of computing resources, the multi-criterion matting algorithm via Gaussian process (GP-MCMatting) is proposed, which can provide a satisfactory alpha matte even when computing resources are limited. The contributions of the work presented in this paper are threefold:
  • By combining different features in a multi-criteria sampling strategy (MCSS), the problem of missing high-quality pixel pairs is alleviated;
  • This paper changes the traditional matting method, which only relies on one evaluation function, and combines multiple evaluation functions to comprehensively evaluate pixel pairs to select high-quality pixel pairs, avoiding the limitation of a single evaluation function;
  • In order to ensure that the matting problem can be solved even with limited computing resources, a new perspective was adopted. This paper proposes a new GP-MCMatting algorithm, in which we use the Gaussian process fitting model (GPFM) instead of the objective function to search for the optimal pixel pair. By using this algorithm, effective and accurate matching can be achieved with only 1% of computing resources.
Next, a description of the algorithm proposed in this article is presented. The article consists of three sections: Section 2 covers related work, Section 3 outlines the problem, and Section 4 introduces the GP-MCMatting algorithm, which includes the MCSS in Section 4.1 and the Gaussian process fitting model in Section 4.2.1. The related work in this field is presented first.

2. Related Work

Recently, the field of image matting has experienced significant advancements, as reflected by numerous publications that can be broadly categorized into those related to sampling-based methods and those related to evolutionary optimization-based methods. In the brief literature review provided below, we focus on the work most closely related to this article.
Sampling-based methods: Sampling-based methods take the sample pixel pairs of the known regions as the candidate sample and select the optimal foreground/background pixel pair through the evaluation function to solve the alpha matte. For example, to avoid the loss of real samples, Feng et al. [26] clustered foreground/background regions and selected representative pixels in each class as candidate samples. Inspired by image inpainting, Tang et al. [19] combined sampling with deep learning methods and used the image inpainting network to select foreground/background pixel pairs as candidate samples to evaluate the alpha mattes of unknown regions. As a part of their research, Huang et al. [27] designed a discrete multi-objective optimization algorithm based on pixel-level sampling. They, thus, effectively solved the problems of incomplete sampling space and optimal sample loss in the super-pixel sampling method and ensured the accuracy of the alpha matte. The strategy adopted by Cao et al. [28] was rooted in the patch-based image matting method, as these authors used a patch-based adaptive matting algorithm for high-resolution images and videos. As shown in their work, the algorithm extends the adaptive framework to video matting and reduces the consumption of computer resources. In short, in sampling-based methods, different sampling strategies [29,30,31] are utilized to select the sample pixel pairs, which reduces the scale of the decision space as well as the computational resource consumption. However, when there are few high-quality pixel pairs in the input image, the sampling-based methods may not identify the optimal pixel pair, resulting in an unsatisfactory alpha matte extraction accuracy.
Evolutionary optimization-based methods: Most evolutionary optimization-based methods [32] are based on the assumption that adjacent pixels have similar alpha values. For example, Liang et al. [33] modeled the image matting problem as a combinatorial optimization problem in which foreground and background pixel pairs are assumed to be known. These authors designed a heuristic optimization algorithm based on an adaptive convergence speed controller, which alleviated the problem of premature algorithm convergence when solving for the optimal pixel pair. In their work, Feng et al. [34] focused on addressing the large-scale problem of high-definition images and proposed a competitive swarm optimization algorithm based on group collaboration to realize the group collaborative solution of large-scale combinatorial optimization problems. These authors demonstrated that, compared to the sampling-based method, the evolutionary optimization-based method could effectively improve the alpha matte estimation accuracy. However, when this method is adopted, more than 5 × 10 3 pixel pairs need to be evaluated for each unknown pixel to ensure that high-quality pixel pairs are captured. Therefore, when the computing resources are limited, it is difficult to attain an adequate alpha value. In summary, when computing resources are limited, it is difficult to obtain a desired matting result with the existing evolutionary optimization-based and learning-based matting methods. Therefore, in this work, we propose a sampling-based multi-criterion sampling strategy (MCSS) to avoid the loss of high-quality pixel pairs. In addition, to achieve adequate matting with limited computing resources, we propose a novel algorithm—the multi-criterion matting algorithm via Gaussian process (GP-MCMatting)—incorporating the MCSS. The GP-MCMatting algorithm uses the Gaussian process fitting model (GPFM) instead of the original objective function to search for the optimal pixel pair, which effectively achieves matting under limited computing resources while ensuring the desired matting accuracy.

3. Problem Description

The authors of the existing sampling-based matting methods have utilized different strategies to select samples from the known regions (the foreground region F and the background region B in Figure 1b) to obtain candidate samples for pixels in the unknown region (the unknown region U in Figure 1b), thereby evaluating the alpha matte of the unknown region. One such sampling method is shown in Figure 2.
In the image above, I represents an RGB image, and F , B , U correspond to the known foreground, background, and unknown regions in the trimap respectively. S is the decision space composed of foreground and background samples, which is denoted by S F × S B . Finally, the evaluation function is used to determine the fitness value of each pixel pair, and the pixel pair with the best fitness is selected as the optimal pixel pair x o p t , based on the criteria shown in Equation (4):
min f ( x ) = ( x i F , x i B ) s . t .   x U , ( x i F , x i B ) S         S = S F × S B
where f ( ) is the evaluation function of the foreground/background pixel pair, x denotes the unknown pixel in the unknown region, ( x F , x B ) represents the decision variable corresponding to the pixels in the known foreground and background regions, and S F × S B is the Cartesian product of the foreground and background sample sets.
As the dimensions of the candidate sample obtained by sampling are smaller than those of the original decision space, sampling-based methods reduce the consumption of computing resources to a certain extent. However, when the number of high-quality pixel pairs in natural images is small (as shown in Figure 3), this strategy is prone to losing high-quality pixel pairs, resulting in an unsatisfactory matting quality.
As shown in Figure 3, depicting GT16 from the alpha matting dataset (with an 800 × 536 size), for the flag of the unknown region to be solved, there are fewer than 500 high-quality pixels confined to a smaller region. Thus, the probability of high-quality pixels being sampled is 0.0034. The sampling strategy proposed in this work that can be adopted to overcome this issue is described below.

4. Multi-Criterion Matting Algorithm via Gaussian Process

In this section, we introduce the GP-MCMatting algorithm, which mainly consists of two stages: (1) the multi-criteria sampling strategy (MCSS); (2) the Gaussian process fitting model (GPFM).

4.1. Multi-Criterion Sampling Strategy

This subsection details the multi-criterion sampling strategy (MCSS), which consists of (1) multi-range pixel pair sampling and (2) high-quality sample selection.

4.1.1. Multi-Range Pixel Pair Sampling

Sampling-based methods rely on local or global sampling strategies to sample pixel pairs, whereby the former approaches the risk of missing high-quality pixel pairs. In order to compensate for this shortcoming, when the latter strategy is adopted, all foreground and background pixel pairs are treated as the sample set to ensure that the optimal solution is found. While this avoids the loss of real samples, it also increases the consumption of computing resources. In this work, we mitigate the aforementioned shortcomings by assuming that there is an optimal pixel pair region in the input image, due to which high-quality pixel pairs can be identified by sampling this region. Guided by this assumption, we designed a multi-range pixel pair sampling method based on the color and spatial position of pixels. In multi-range pixel pair sampling, the color similarity score between the pixels can be reformulated as follows:
D c = I p I i U 2 ,   I p F B
where 2 denotes the Euclidean norm of vector , F B denotes the set of known regions, I p is the color value of the known pixel p , and I i U is the color value of the unknown pixel i .
In image matting, the spatial positions of pixels can reflect their structural features, which may help in the differentiation of the pixels. For this purpose, the similarity between the unknown region and the known boundary pixels, as well as the similarity between the unknown region and the long-range known regions (as shown in Figure 4), are calculated using the expressions shown below:
D s = I p , i n I i , i n U 2 ,   I p , i n F i n B i n  
D b = I p , i n b I i , i n U 2 ,   I p , i n b F i n B i n  
where D s denotes the spatial position distance between the unknown pixels and the long-range known pixels; D b represents the spatial distance between the pixels in the unknown region and the pixels in the known boundaries; I p , i n b represents the coordinate index values of the pixels in the known boundaries, I p , i n denotes the coordinate index value of pixels in the long-range known region; I i , i n U denotes the coordinate index value of the unknown pixel, and F i n B i n is the coordinate space of the known regions.
Let the color distance set between the unknown pixels and the known pixels be given by D c = { D 1 c , D 2 c , , D m c } , whereas the spatial distance between the unknown pixels and the long-range known region pixels is given by D s = { D 1 s , D 2 s , , D m s } , and the spatial distance between the unknown pixels and the pixels in the known boundaries is D b = { D 1 b , D 2 b , , D m b } . For a more intuitive view of the pixel distance scores, these three sets are rearranged in ascending order, i.e., D ( ) c = { D ( 1 ) c , D ( 2 ) c , , D ( m ) c } , D ( ) s = { D ( 1 ) s , D ( 2 ) s , , D ( m ) s } , and D ( ) b = { D ( 1 ) b , D ( 2 ) b , , D ( m ) b } . According to the sample sets of three different distances, the color is first used as the main collection feature, after which the pixels with similar colors and close distances in the set are considered as the first type of candidate samples. Then, the sampling process is described by the following expressions:
X c = { X c S k c ,   S k c < d max ( S c ) | S c |               ,   o t h e r ,   k = 1 , 2 , , | S c |  
S c = { S c { D ( j ) s } { D ( j ) b } ,   i f   D ( j ) c < ε                                         ,   o t h e r ,   j = 1 , 2 , , | D ( ) c |  
where X c is a sample set with similar colors and close distances, whereas S c is a set of pixels with a color similar to the unknown pixels, ε denotes the threshold of the color distance, and d * max ( S c ) | S c | represents the proportion of pixels with a small color distance and coordinate space distance in the S c set, where d denotes the number of samples.
Furthermore, according to the obtained multi-range sets, the coordinate space is the main acquisition feature, and the pixels in the same set characterized by a close spatial distance and similar color are considered as the second candidate sample type. This sampling method is described by the following equations:
X s = { X s S k s ,   S k s < d max ( S s ) | S s |              ,   o t h e r ,   k = 1 , 2 , , | S s |
S s = { S s { D ( j ) c } ,   i f   D ( j ) s < d                          ,   o t h e r ,   j = 1 , 2 , , | D ( ) c |
where X s is a sample set with a close spatial distance and similar colors, and S s represents a set of similar pixel distances. Thus, when sampling based on different primary features, the sample sets X c and X s are formed, allowing the final candidate sample X to be obtained as their union, i.e., X = X c × X s which is the Cartesian product of X c and X s .

4.1.2. High-Quality Sample Selection

As explained in the preceding section, multi-range pixel pair sampling yields the candidate sample set X = X c × X s . In cases with a large number of candidate samples that need to be considered for obtaining high-quality pixel pairs, the high-quality sample selection method proposed here can be adopted, which is combined with the multi-range pixel pair sampling method to form the MCSS. In the process of pixel pair evaluation, a single evaluation function is usually used to evaluate the optimal pixel pair. However, this may compromise the ability to determine the fitness value of a given pixel pair [34]. Therefore, we propose a novel approach that combines multiple evaluation functions to select high-quality pixel pairs as candidate samples. Specifically, we employ color difference evaluation f c ( x ) , fuzzy evaluation f f ( x ) , and fuzzy multi-criterion evaluation f m ( x ) as joint evaluation indices for sample assessment. Let f i denote the fitness value after averaging the three evaluation functions of pixel x . Accordingly, the fitness value corresponding to the elements in candidate sample X is F = { f 1 , f 2 , , f | X | } , where f i = 1 / 3 ( f c ( x ) + f f ( x ) + f m ( x ) ) . Reordering F in ascending order yields F ( ) = { f ( 1 ) , f ( 2 ) , , f ( | X | ) } . Then, according to the fitness value obtained by the joint evaluation function, the high-quality pixel pairs can be selected using the expression below:
Ω f = { Ω f f i ,   i f   min f i ( x ) + 1 > f i ( x ) ,   i = 1 , 2 , , | F | ,   f i F            ,   o t h e r Ω f Ω
where Ω f represents the set of fitness values of the high-quality pixel pairs; Ω is the pixel information corresponding to the fitness values of high-quality pixel pairs; and threshold min ( f i ( x ) ) + 1 is implemented to exclude excessively high values of the evaluation function. After obtaining high-quality pixel pairs according to Equations (12) and (13), the number of high-quality pixel pair samples is far smaller than the number of pixels in the decision space. As shown in Algorithm 1, the aforementioned approaches are combined to yield the MCSS.
Algorithm 1: Multi-criterion sampling strategy
Input: image and trimap.
1.//For information on image pixel.
2. F B = Known pixel information by trimap.
3. for  i = 1  to  | F B |  do
4.      //multi-range pixel-pairs sampling.
5.       { D c , D s , D b } = By Equations (5)–(7).
6.       { D ( ) c , D ( ) s , D ( ) b } = Sort the sets in ascending order.
7.      while  D i , ( ) c < ε  do
8.                 X c =  By Equations (8) and (9).
9.      end while
10.    while   D ( ) s < d   do
11.               X s = By Equations (10) and (11).
12.     end while
13.     X = X c × X s
14.end for
15.//High-quality sample selection.
16. F ( ) = { f ( 1 ) , f ( 2 ) , , f ( | X | ) } = By Equation (12).
17.if stop criterion is not met then
18.         Ω = Ω { x i }   / / x i   is   the   corresponding   sample   pixel   of   f i .
19. end if
Output: high-quality pixel pairs set   Ω .
Specifically, to ensure that high-quality pixel pairs are not missing, in Lines 1–2, the foreground and background pixel information is gathered according to the input image, while in Lines 3–14, the candidate sample set X is output after excessive range sampling, and in Lines 15–19, the final high-quality sample set Ω is obtained according to the high-quality selection method.

4.2. Multi-Criterion Matting Algorithm via Gaussian Process

In this subsection, we introduce the multi-criterion matting algorithm via Gaussian process (GP-MCMatting) proposed in this work, which enables image matting even when computing resources are limited while ensuring the required accuracy of the image matting.

4.2.1. Gaussian Process Fitting Model

The objective function of the natural image matting problem is typically complex due to the simultaneous consideration of the similarity between pixels in the foreground and background regions, as well as the similarity among pixels. Traditional objective functions usually require nonlinear optimization algorithms for solving, which demands intensive computational resources and time [25]. When computing resources are limited, it may be challenging to directly use traditional objective function-based methods. Accordingly, we assume that approximating the evaluation of the objective function by fitting the parameters of Gaussian processes (GPs) [20] can improve the accuracy of alpha matte. Therefore, the GPFM can be represented as follows:
min H ( x i , i n ) min f ( x ) s . t . min f ( x )       x = ( x i F , x i B ) ,   x i , i n = ( x i , i n F , x i , i n B )       x i F , x i , i n F F ;   x i B , x i , i n B B
where H ( x i , i n ) is the Gaussian process fitting model (GPFM). Because there are many redundant solution space regions in the coordinate space, in the proposed algorithm, the image index number space is used to construct the distribution of the objective function. Moreover, to avoid crossing the boundary when searching for the optimal solution, the GPFM is constructed in the index value space.
The GP is an infinite set of random variable distributions, where the joint distribution function of each finite subset is subject to Gaussian distribution [35,36]. Therefore, the GP is completely determined by the mean function and the covariance kernel function between any two random variables [37,38]. In this work, the optimization of the GPFM model is achieved by approximating the optimal solution of the pixel pair objective function, essentially by estimating the undetermined parameters in the GPFM and then using the evaluated parameters to obtain the optimal solution [20]. Let θ and θ ^ denote the parameter vector and the estimated parameter vector of the kernel function in the GPFM, respectively. Then, the likelihood function is used to estimate the kernel function parameters in the Gaussian process fitting model. Mathematically, the expression can be described as follows:
θ ^ = L ( θ | f ( x ) , x i , i n ) = arg min θ log P ( f ( x ) , x i , i n | θ )
Therefore, according to Equation (15), the GPFM of the pixel pair evaluation function can be obtained via the following expressions:
min H = H ( x i , i n | θ ^ )
Then, the index position of the optimal pixel pair can be evaluated by fitting the GPFM in Equation (16), allowing us to solve for the optimal pixel pair.

4.2.2. Multi-Criterion Matting Algorithm via Gaussian Process

Natural image matting is a large-scale problem, for which extensive computing resources are needed to solve the original objective function directly. Therefore, to achieve matting with limited computing resources, we propose a novel matting algorithm named the multi-criterion matting algorithm via Gaussian process (GP-MCMatting), which includes (1) the multi-criterion sampling strategy (MCSS) and (2) the Gaussian process fitting model (GPFM). The former is mainly used to select high-quality sample pixel pairs from a large number of decision variables. The latter mainly uses the GPFM to approximate the original objective function of natural images based on the MCSS, thereby reducing the consumption of computational resources. For constructing the GPFM basis, we define Θ as a penalty function of the GPFM:
Θ ( Ω , γ ( k ) ) = H ( Ω ) γ ( k ) log f ( Ω )
where γ ( k ) is the penalty factors that decrease sequentially and lim k + γ ( k ) = 0 , Ω is the set of high-quality pixel pair samples.
Once the GPFM is constructed, it is utilized to approximate the original objective function for a given pixel pair, allowing the optimal pixel pair to be obtained by solving the extreme points of the model via the high-quality pixel pair set Ω. In this work, the optimal solution of GPFM is obtained based on the ideas behind the interior-point algorithm (IPA) [39]. Within the solving process, the optimal solution x min of the GPFM is approximately equivalent to the optimal solution x b e s t of the objective function. In addition, we find that the optimal pixel pair is usually located in the X l o c of the known region closest to the unknown pixels. Therefore, selecting the closest pixel pair in the local foreground region and the background region as the initial IPA population often yields a viable solution. When this assumption does not hold, the IPA algorithm can also find the optimal foreground/background pixel pairs by solving the GPFM, because the algorithm is applied to the entire search space, thus theoretically avoiding the loss of the optimal pixel pair. The solving process of GP-MCMatting is shown in Algorithm 2.
Specifically, let ϵ represent the distance between the GPFM and the objective function that needs to be approximated, and τ is smaller than a specified small value. Θ x γ ( k ) is the partial derivative, where c is a randomly generated attenuation factor, and γ ( k ) is a penalty coefficient greater than 0. Ω is a set of high-quality pixel pairs obtained by Algorithm 1. In Algorithm 2, the required parameters are first initialized, and Lines 5–10 are used to construct GPFM. Lines 12–22 are the process of solving the optimal solution of the GPFM. The algorithm stops iterating when the optimal solution of the GPFM is approximately equivalent to the optimal solution of the objective function.
Algorithm 2: Multi-criteria matting algorithm via Gaussian process.
Input: image and trimap.
1. //Initialize parameters ϵ , τ .
2. x l o c = X l o c ( x ) .
3. γ ( k ) = a random number greater than 0.
4. c = a random number between [0, 1].
5.for  i = 1  to  | U |  do
6.      //Gaussian process fitting model construction.
7.       Ω = According to the Algorithm 1.
8.       ( f ( x ) , x i , i n ) = Ω .
9.       H = H ( L p ( θ | f ( x ) , x i , i n ) ) , x i , i n Ω .
10.    //Optimal pixel pair estimation.
11.     x min = min H ( x l o c ) .
12.    while  ϵ > τ  do
13.               x γ ( k ) = Θ x γ ( k ) = 0 .
14. x i n i t = x γ ( k ) .
15. ϵ = x γ ( k ) x i n i t .
16. γ ( k ) = γ ( k 1 ) c .
17.               if γ ( k ) 0 .  then
18.                       γ ( k ) =   a   random   number   greater   than   0 .
19. c =   a   random   number   between [0, 1].
20.end if
21.    end while
22.end for
23. x b e s t = x γ ( k )
Output :   x b e s t .

5. Experiments and Results

In this section, three sets of experiments are described, which were performed to verify the effectiveness of the high-quality pixel pair selection strategy and the effectiveness of the algorithm.

5.1. Experimental Setup

In the experiments, the alpha matting dataset presented by Rhemann et al. [9] (which contains 35 images, 27 of which are training images and 8 are testing images) was used as the benchmark dataset. All experiments were implemented in MATLAB, and we set up three groups of experiments, as outlined below.
  • This experiment was mainly performed to verify that the MCSS can effectively avoid the loss of high-quality pixel pairs, thus improving the matting performance.
  • In this experiment, a comparative analysis was conducted between GP-MCMatting and the state-of-the-art evolutionary optimization-based algorithms, such as the pyramid matting framework (PMF) [25], adaptive convergence speed controller based on particle swarm optimization (PSOACSC) [33], and the multi-objective evolutionary algorithm based on multi-criteria decomposition (MOEAMCD) [24], based on 1%, 2%, 5%, 10%, 20%, and 100% computing resources. The matting performance of the proposed algorithm under limited computing resources was verified.
  • This experiment was conducted to compare the GP-MCMatting performance with the aforementioned algorithms based on the availability of only 1% computing resources to verify its superiority.
In order to ensure the comparability of the experimental results based on different algorithms and to facilitate statistical analysis of their performance, the mean square error (MSE) and absolute value error (SAD) were adopted as the evaluation indices. In the sections that follow, the results and analyses of each group of experiments are presented.

5.2. Effectiveness of Multi-Criterion Sampling Strategy

In the MCSS, the spatial coordinates and color features of the pixels were sampled by solving Equations (8) and (9). Using Equation (9), pixels whose color distance was less than ε were considered candidate samples. In order to determine the optimal value of ε , we calculated the matting results under different distances. Figure 5 shows the sum of the MSE values of the 27 alpha matting training sets based on 1% computing resources and the different values of ε . The results show that when the color distance parameter ε was assigned the value of 10 (the red corresponds to the dashed line), the total MSE based on all 27 images was the smallest with a value of 0.789, i.e., the alpha matte accuracy was the highest. Therefore, the color distance parameter ε was set to 10 in the MCSS.
Additionally, to verify the effectiveness of the high-quality sample selection method in the MCSS, a comparative analysis of the matting quality obtained by selecting different evaluation functions was performed, and the findings are shown in Table 1. The comparisons were based on the sum of the MSE and SAD of the 27 images solved by different evaluation functions. By referring to Table 1, we can see that on the basis of the MSE and SAD evaluation metrics, the MCSS outperformed the method based on a single evaluation function. Therefore, the experimental results show that the combination of multiple evaluation functions to select high-quality pixel pairs was significantly better than a single evaluation function to select pixel pairs, which fully verifies the effectiveness of the MCSS.

5.3. Algorithm Evaluation and Comparison under the Conditions Characterized by Limited Computing Resources

In this subsection, GP-MCMatting is compared with the state-of-the-art evolutionary optimization-based algorithms to verify the matting performance of the proposed algorithm when computing resources are limited. As most algorithms require 5000 iterations per pixel [24,25,33], this was adopted as the benchmark, and 1%, 2%, 5%, 10%, 20%, and 100% availability of computing resources were considered in 5000 different scenarios to compare the matting performance of the studied algorithms. For the fairness of comparison, the MSE and SAD were used as the evaluation indicators to compare the alpha matte accuracy of the different algorithms. The results in Table 2 represent the sum of the MSE and SAD of different algorithms when applied to the training sets contained in the alpha matting dataset under different computing-resource conditions.
According to the results reported in Table 2, in terms of the SAD value, the matting performance of the proposed algorithm under limited and sufficient computing resources was superior to those associated with the PMF, MOEAMCD, and PSOACSC algorithms. According to the MSE evaluation index, the matting effect of the proposed GP-MCMatting algorithm was better than that of the algorithms used in the comparison only when 1%, 2%, and 5% of the computing resources were considered. Nonetheless, the GP-MCMatting algorithm was still superior to the PMF and PSOACSC in the scenarios based on 10%, 20%, and 100% computing resource availability. When both the SAD and MSE were considered, however, the proposed GP-MCMatting algorithm not only achieved competitive matting effects under limited computing resources, but also achieved good results even when computing resources were sufficient. Thus, the proposed algorithm based on the MCSS effectively realized the matting results irrespective of the computing resource availability.
In order to further analyze the effectiveness of GP-MCMatting under limited computing resources, the MSE results of different algorithms were compared based on their application on 27 images in the alpha matting dataset and the availability of 1% computing resources.

5.4. Comparison to the State-of-the-Art Methods

In order to demonstrate the effectiveness of the proposed GP-MCMatting algorithm and compare its performance to the state-of-the-art algorithms when computing resources are limited, the MSE values obtained by applying different matting algorithms to the 27 training sets under 1% computing resources are shown in Table 3.
According to the results reported in Table 3, GP-MCMatting outperformed the PMF, PAOACSC, and MOEAMCD algorithms in 24 of the 27 cases (with the exception of the GT24, and GT15 images). The reasons behind its suboptimal performance in these two cases are analyzed in detail in Section 5.5.
To further verify the matting effect of the proposed algorithm, the results yielded by the GP-MCMatting algorithm on the training set and test set are depicted in Figure 6. It is evident that GP-MCMatting effectively evaluated the object containing hair in the foreground target in the image (the visualization results in rows 1, 2, 3, 4, and 6). Conversely, according to the comprehensive visualization results and the MSE accuracy results, the matting effect of the compared algorithms under the limited computing resources was not satisfactory, and there were large unknown areas that could not be evaluated. Based on the analysis of the experimental results shown in Table 3 and the images included in Figure 6, it is evident that the algorithm proposed in this work avoided the loss of the optimal pixel pair by selecting high-quality pixel pairs, thus improving the extraction accuracy of the alpha matte. In addition, under limited computing resources, satisfactory matting was realized by approximately solving the GPFM, which reflects the superiority of the proposed GP-MCMatting.

5.5. Limitations of the GP-MCMatting Algorithm

The comparison and analysis of the experimental results in Section 5.4 fully reflect the superiority of the proposed algorithm under conditions characterized by limited computing resources. However, the results reported in Table 3 in Section 5.4 show that the extraction accuracy of the GP-MCMatting algorithm was low when applied to the GT15 and GT24 images, and there were failures in the acquisition of high-quality pixel pairs, as shown in Figure 7.
The results depicted in Figure 7 reveal that the color information of the foreground and background pixels in the annotated region was relatively similar, as depicted by the green box in the images. Although the pixels in this region met the color sampling requirements due to their minor color differences, the spatially close distribution of these pixels impeded the search for the optimal pixel pair during the evaluation of the kernel function parameters of the Gaussian process field model (GPFM) using pixel spatial distribution. Consequently, in complex images that exhibit similar foreground and background characteristics, GPFM approximations cannot replace the original evaluation function when solving for the optimal pixel pairs. This fundamental limitation constrains the practical implementation of the GP-MCMatting algorithm in such scenarios.

6. Conclusions

In this paper, we present a multi-criterion sampling matting algorithm via Gaussian process (GP-MCMatting) that effectively solves the matting problem with limited computing resources. To overcome the challenge of losing the optimal pixel pair, which occurs in sampling-based methods, an MCSS was designed based on multi-range pixel pair sampling and high-quality sample selection. Additionally, to address the challenge of matting under limited computing resources, the GP-MCMatting algorithm searches for the optimal pixel pair using an approximation of the GPFM instead of the original evaluation function. This technique avoids the protracted calculations required by the original evaluation function while still ensuring accurate matting. The experimental results demonstrate that the GP-MCMatting algorithm achieved similar results under sufficient computing resources and was superior when 1%, 2%, and 5% of computing resources were available.
However, when images with similar foreground and background color information were input, the GP-MCMatting algorithm failed to achieve satisfactory results. Therefore, to address the inherent limitations of the algorithm, future research will focus on exploring the color variation characteristics of adjacent pixels in images, taking into account the spatial continuity and correlation between pixels. This approach aims to solve the challenge of accurately extracting the alpha matte in images with extremely similar foreground and background color information, where the alpha matte is difficult to extract.

Author Contributions

Conceptualization, F.F. and M.T.; methodology, Y.Y. and H.G.; software, H.G. and Y.L.; validation, Y.X.; formal analysis, F.F.; investigation, M.T.; resources, H.H.; writing—original draft preparation, Y.Y.; writing—review and editing, L.W., H.H., Y.X., F.F., M.T. and Y.L.; supervision, H.H.; project administration, F.F., M.T., Y.L., Y.X. and H.H. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded by [the National Natural Science Foundation of China] grant number [62276103, 62002053, 61906069], [the Guizhou Provincial Science and Technology Projects] grant number [QKHJCZK2022YB195, QKHJCZK2023YB143, QKH-Platform Talent-ZCKJ [2021]007], [the Natural Science Foundation of Guangdong Province] grant number [2020A1515010696, 2022A1515011491, 2021A0101180005], [the Fundamental Research Funds for the Central Universities] grant number [2020ZYGXZR014], and The APC was funded by [the Youth Science and Technology Talents Cultivating Object of Guizhou Province] grant number [QJHKY2021104, QJHKY [2021]111], [the Science and Technology Support Program of Guizhou Province] grant number [QKHZC2021YB531], [the Natural Science Research Project of Department of Education of Guizhou Province] grant number [QJJ2023012, QJJ2023061, QJJ2023062], [the Zhongshan Science and Technology Research Project of Social welfare] grant number [2020B2017, 210714094038458] and [the high-level personnel Scientific Research Foundation Project of Zhongshan] grant number [2019A4018].

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The figures utilized to support the findings of this study are included in the article.

Acknowledgments

This work was supported by [the National Natural Science Foundation of China] grant number [62276103, 62002053, 61906069], [the Guizhou Provincial Science and Technology Projects] grant number [QKHJCZK2022YB195, QKHJCZK2023YB143, QKH-Platform Talent-ZCKJ [2021]007], [the Natural Science Foundation of Guangdong Province] grant number [2020A1515010696, 2022A1515011491, 2021A0101180005], [the Fundamental Research Funds for the Central Universities] grant number [2020ZYGXZR014], and The APC was funded by [the Youth Science and Technology Talents Cultivating Object of Guizhou Province] grant number [QJHKY2021104, QJHKY [2021]111], [the Science and Technology Support Program of Guizhou Province] grant number [QKHZC2021YB531], [the Natural Science Research Project of Department of Education of Guizhou Province] grant number [QJJ2023012, QJJ2023061, QJJ2023062], [the Zhongshan Science and Technology Research Project of Social welfare] grant number [2020B2017, 210714094038458] and [the high-level personnel Scientific Research Foundation Project of Zhongshan] grant number [2019A4018].

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Jiang, W.; Yu, D.; Xie, Z.; Li, Y.; Yuan, Z.; Lu, H. Trimap-guided feature mining and fusion network for natural image matting. Comput. Vis. Image Underst. 2023, 230, 1077–3142. [Google Scholar] [CrossRef]
  2. Yan, C.-G.; Meng, L.-X.; Li, L.; Zhang, J.-H.; Wang, Z.; Yin, J.; Zhang, J.-Z.; Sun, Y.-Q.; Zheng, B.-L. Age-Invariant Face Recognition by Multi-Feature Fusion and Decomposition with Self-Attention. ACM Trans. Multimed. Comput. Commun. Appl. 2021, 29, 1–18. [Google Scholar]
  3. Hu, L.-P.; Kong, Y.-T.; Li, J.-D.; Li, X.-Q. Effective Local-Global Transformer for Natural Image Matting. J. IEEE Trans. Circuits Syst. Video Technol. 2023, 99, 1. [Google Scholar] [CrossRef]
  4. Park, G.-T.; Son, S.-J.; Yoo, J.-Y.; Kim, S.-H.; Kwak, N. MatteFormer: Transformer-Based Image Matting via Prior-Tokens. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 11696–11706. [Google Scholar]
  5. Lin, S.-C.; Yang, L.; Saleemi, I.; Sengupta, S. Robust High-resolution Video Matting with Temporal Guidance. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–8 January 2022; pp. 3132–3141. [Google Scholar]
  6. Yan, C.-G.; Li, Z.-S.; Zhang, Y.-B.; Liu, Y.-T.; Ji, X.-Y.; Zhang, Y.-D. Depth Image Denoising Using Nuclear Norm and Learning Graph Model. ACM Trans. Multimed. Comput. Commun. Appl. 2020, 122, 1–17. [Google Scholar] [CrossRef]
  7. Sun, G.; Wang, Q.; Gu, C.; Tang, K.; Tai, Y.W. Deep Video Matting via Spatio-Temporal Alignment and Aggregation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition 2021, Nashville, TN, USA, 20–25 June 2021; pp. 6971–6980. [Google Scholar]
  8. Yanan, S.; Tang, C.-K.; Tai, Y.-W. Semantic Image Matting. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition 2021, Nashville, TN, USA, 20–25 June 2021; pp. 11120–11129. [Google Scholar]
  9. Huang, H.; Feng, F.-J.; Huang, S.; Chen, L.; Hao, Z.-F. Micro-scale Searching Algorithm for Coupling Matrix Optimization of Automated Microwave Filter Tuning. IEEE Trans. Cybern. 2023, 53, 2829–2840. [Google Scholar] [CrossRef] [PubMed]
  10. Li, J.; Yuan, G.; Fan, H. Robust trimap generation based on manifold ranking. Inf. Sci. 2020, 519, 200–214. [Google Scholar] [CrossRef]
  11. Thomas, P.; Tom, D. Compositing Digital Images. Proc. ACM Siggraph Comput. Graph. 1984, 18, 253–259. [Google Scholar]
  12. Rhemann, C.; Rother, C.; Wang, J.; Gelautz, M.; Kohli, P.; Rott, P. A Perceptually Motivated Online Benchmark for Image Matting. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 1826–1833. [Google Scholar]
  13. Ding, H.; Zhang, H.; Liu, C.; Jiang, X. Deep Interactive Image Matting with Feature Propagation. IEEE Trans. Image Process. 2022, 31, 2421–2432. [Google Scholar] [CrossRef]
  14. Liang, Y.; Huang, H.; Cai, Z.; Hao, Z.; Tan, K.C. Deep infrared pedestrian classification based on automatic image matting. Appl. Soft Comput. 2019, 77, 484–496. [Google Scholar] [CrossRef]
  15. Chen, X.; Zou, D.; Zhao, Q.; Tan, P. Manifold Preserving Edit Propagation. ACM Trans. Graph. 2012, 31, 1–7. [Google Scholar] [CrossRef]
  16. Aksoy, Y.; Aydin, T.-O.; Pollefeys, M. Designing Effective Inter-pixel Information Flow for Natural Image Matting. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition 2017, Honolulu, HI, USA, 21–26 July 2017; pp. 29–37. [Google Scholar]
  17. Li, X.; Li, J.; Lu, H. A Survey on Natural Image Matting with Closed-Form Solutions. IEEE Access 2019, 7, 136658–136675. [Google Scholar] [CrossRef]
  18. Cai, Z.-Q.; Liang, L.; Huang, H.; Hu, H.; Liang, Y.-H. Improving Sampling-based Image Matting with Cooperative Co-evolution Differential Evolution Algorithm. Appl. Soft Comput. 2017, 21, 4417–4430. [Google Scholar] [CrossRef]
  19. Tang, J.-W.; Aksoy, Y.; Oztireli, C.; Gross, M.; Aydin, T.-O. Learning-based Sampling for Natural Image Matting. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition 2019, Long Beach, CA, USA, 15–20 June 2019; pp. 3055–3063. [Google Scholar]
  20. Liang, Y.; Gou, H.; Feng, F.; Liu, G.; Huang, H. Natural image matting based on surrogate model. Appl. Soft Comput. 2023, 143, 110407. [Google Scholar] [CrossRef]
  21. He, K.-M.; Rhemann, C.; Rother, C.; Tang, X.-O.; Sun, J. A Global Sampling Method for Alpha Matting. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition 2011, Colorado Springs, CO, USA, 20–25 June 2011; pp. 2049–2056. [Google Scholar]
  22. Rother, C.; Kolmogorov, V.; Blake, A. GrabCut: Interactive Foreground Extraction using Iterated Graph Cuts. ACM Trans. Graph. 2004, 23, 309–314. [Google Scholar] [CrossRef]
  23. Levin, A.; Rav-Acha, A.; Lischinski, D. Spectral Matting. IEEE Trans. Pattern Anal. Mach. Intell. 2008, 30, 1699–1712. [Google Scholar] [CrossRef] [PubMed]
  24. Liang, Y.; Huang, H.; Cai, Z.; Hao, Z. Multiobjective Evolutionary Optimization Based on Fuzzy Multicriteria Evaluation and Decomposition for Image Matting. IEEE Trans. Fuzzy Syst. 2019, 27, 1100–1111. [Google Scholar] [CrossRef]
  25. Liang, Y.-H.; Feng, F.-J.; CAI, Z.-Q. Pyramid Matting: A Resource-Adaptive Multi-Scale Pixel-pair Optimization Framework for Image Matting. IEEE Access 2020, 8, 93487–93498. [Google Scholar]
  26. Feng, X.; Liang, X.; Zhang, Z. A Cluster Sampling Method for Image Matting via Sparse Coding. In Proceeding of the Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Springer International Publishing: New York, NY, USA, 2016; pp. 204–219. [Google Scholar] [CrossRef]
  27. Huang, H.; Liang, Y.-H.; Yang, X.; Hao, Z.-F. Pixel-level Discrete Multi-objective Sampling for Image Matting. IEEE Trans. Image Process. 2019, 28, 3739–3751. [Google Scholar] [CrossRef]
  28. Cao, G.; Li, J.; Chen, X.; He, Z. Patch-based self-adaptive matting for high-resolution image and video. Vis. Comput. 2019, 35, 133–147. [Google Scholar] [CrossRef]
  29. Shahrian, E.; Rajan, D.; Price, B.; Cohen, S. Improving Image Matting using Comprehensive Sampling Sets. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition 2013, Portland, OR, USA, 25–27 June 2013; pp. 636–643. [Google Scholar]
  30. Wang, J.; Cohen, M.-F. Optimized Color Sampling for Robust Matting. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition 2007, Honolulu, HI, USA, 21–26 July 2017; pp. 1–8. [Google Scholar]
  31. Eduardo, S.; Lopes, G.; Manuel, M.-O. Shared Sampling for Real-Time Alpha Matting. Comput. Graph. Forum 2010, 10, 575–584. [Google Scholar]
  32. Feng, F.-J.; Huang, H.; Liu, D.; Liang, Y.-H. Local Complexity Difference Matting based on Weight Map and Alpha Mattes. Multimed. Tools Appl. 2022, 81, 43357–43372. [Google Scholar] [CrossRef]
  33. Liang, Y.-H.; Huang, H.; Cai, Z.-Q. PSO-ACSC: A Large-scale Evolutionary Algorithm for Image Matting. Front. Comput. Sci. 2020, 14, 29–40. [Google Scholar] [CrossRef]
  34. Feng, F.-J.; Huang, H.; Wu, Q.-X.; Ling, X.; Liang, Y.-H.; Cai, Z.-Q. An Alpha Matting Algorithm based on Collaborative Swarm Optimization for High-resolution Images. J. Chin. Sci. Inf. Sci. 2020, 50, 424–437. [Google Scholar]
  35. Rasmussen, C.E. Evaluation of Gaussian Processes and Other Methods for Non-Linear Regression; University Toronto: Toronto, ON, Canada, 1999. [Google Scholar]
  36. Rasmussen, C.E.; Williams, K.-I. Gaussian Processes for Machine Learning (Adaptive Computation and Machine Learning); MIT Press: Cambridge, MA, USA, 2005. [Google Scholar]
  37. Khan, M.-E.; Immer, A.; Abedi, E.; Korzepa, M. Approximate Inference Turns Deep Networks into Gaussian processes. In Advances in Neural Information Processing Systems; Neural Information Processing Systems Foundation, Inc.: La Jolla, CA, USA, 2019; pp. 3088–3098. [Google Scholar]
  38. Zheng, Y.-J.; Yang, Y.-S.; Che, T.-T.; Hou, S.-J.; Huang, W.-H.; Gao, Y.; Tan, P. Image Matting with Deep Gaussian Process. IEEE Trans. Neural Netw. Learn. Syst. 2022, 7, 1–15. [Google Scholar] [CrossRef] [PubMed]
  39. Byrd, R.H.; Gilbert, J.C.; Nocedal, J. A trust region method based on interior point techniques for nonlinear programming. Math. Program. 2000, 89, 149–185. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Image and trimap, where (a) Input image; (b) Trimap; (c) Constraint regions by trimap.
Figure 1. Image and trimap, where (a) Input image; (b) Trimap; (c) Constraint regions by trimap.
Biomimetics 08 00301 g001
Figure 2. An example of a sampling-based matting method.
Figure 2. An example of a sampling-based matting method.
Biomimetics 08 00301 g002
Figure 3. Case with a small number of high-quality pixels in the original image.
Figure 3. Case with a small number of high-quality pixels in the original image.
Biomimetics 08 00301 g003
Figure 4. Spatial distance between pixels.
Figure 4. Spatial distance between pixels.
Biomimetics 08 00301 g004
Figure 5. Sensitivity analysis of the color distance parameter.
Figure 5. Sensitivity analysis of the color distance parameter.
Biomimetics 08 00301 g005
Figure 6. Visual comparison on the alpha matting dataset. (a) input image; (b) trimap; (c) groundtruth; (d) PMF; (e) MOEA-MCD; (f) PSOACSC; (g) GP-MCMatting.
Figure 6. Visual comparison on the alpha matting dataset. (a) input image; (b) trimap; (c) groundtruth; (d) PMF; (e) MOEA-MCD; (f) PSOACSC; (g) GP-MCMatting.
Biomimetics 08 00301 g006
Figure 7. The case of GP-MCMatting failure when applied to the GT15 and GT24 images.
Figure 7. The case of GP-MCMatting failure when applied to the GT15 and GT24 images.
Biomimetics 08 00301 g007
Table 1. Comparison of the matting results yielded by different evaluation functions.
Table 1. Comparison of the matting results yielded by different evaluation functions.
Color Difference Evaluation FunctionMulti-Objective Evaluation FunctionFuzzy Evaluation FunctionMulti-Criteria Sampling Strategy
MSE0.02990.02970.02930.0292
SAD7.26997.17317.16287.0999
Table 2. The SAD and MSE values for the GP-MCMatting and the state-of-the-art evolutionary optimization-based algorithms based on their application to 27 images at 1%, 2%, 10%, 20%, and 100% computer power and 5000 evaluations. Bold Arabic numerals indicate the index with the highest ranking.
Table 2. The SAD and MSE values for the GP-MCMatting and the state-of-the-art evolutionary optimization-based algorithms based on their application to 27 images at 1%, 2%, 10%, 20%, and 100% computer power and 5000 evaluations. Bold Arabic numerals indicate the index with the highest ranking.
Computing ResourcesSAD
1%2%5%10%20%100%
PSOACSC [33]604.425604.267604.418604.137603.950602.987
MOEAMCD [24]243.965243.346242.741242.179243.028242.741
PMF [25]349.160321.151294.067272.818253.481228.433
Ours191.697187.752202.500214.409216.114210.910
Computing resourcesMSE
1%2%5%10%20%100%
PSOACSC [33]5.061 5.063 5.063 5.062 5.056 5.039
MOEAMCD [24]1.127 1.121 1.113 1.1161.1181.113
PMF [25]2.231 1.910 1.689 1.430 1.257 1.028
Ours0.7890.8261.0291.151 1.163 1.171
Table 3. The MSE values of GP-MCMatting and the state-of-the-art evolutionary optimization-based methods based on their application on 27 images under 1% computing resources. Bold Arabic numerals indicate the index with the highest ranking.
Table 3. The MSE values of GP-MCMatting and the state-of-the-art evolutionary optimization-based methods based on their application on 27 images under 1% computing resources. Bold Arabic numerals indicate the index with the highest ranking.
AlgorithmsGT01GT02GT03GT04GT05GT06GT07GT08GT09
Ours3.33 × 10−36.40 × 10−39.68 × 10−31.23 × 10−21.49 × 10−21.49 × 10−27.89 × 10−33.93 × 10−21.29 × 10−2
PSOACSC [33]6.65 × 10−22.13 × 10−15.35 × 10−21.01 × 10−11.62 × 10−12.24 × 10−19.11 × 10−21.26 × 10−11.57 × 10−1
MOEAMCD [24]7.49 × 10−31.53 × 10−21.19 × 10−22.15 × 10−22.53 × 10−22.71 × 10−21.08 × 10−24.40 × 10−21.11 × 10−2
PMF [25]1.74 × 10−21.23 × 10−11.45 × 10−23.28 × 10−24.78 × 10−24.45 × 10−22.84 × 10−25.71 × 10−22.16 × 10−2
AlgorithmsGT10GT11GT12GT13GT14GT15GT16GT17GT18
Ours2.32 × 10−23.21 × 10−21.00 × 10−21.23 × 10−27.81 × 10−35.04 × 10−27.64 × 10−21.22 × 10−27.50 × 10−3
PSOACSC [33]2.22 × 10−13.02 × 10−14.61 × 10−22.08 × 10−11.55 × 10−11.79 × 10−13.62 × 10−11.41 × 10−12.04 × 10−1
MOEAMCD [24]2.77 × 10−23.93 × 10−21.47 × 10−22.29 × 10−22.61 × 10−24.02 × 10−21.87 × 10−11.59 × 10−21.66 × 10−2
PMF [25]6.36 × 10−27.65 × 10−21.84 × 10−26.66 × 10−25.37 × 10−29.36 × 10−23.43 × 10−13.01 × 10−26.15 × 10−2
AlgorithmsGT19GT20GT21GT22GT23GT24GT25GT26GT27
Ours8.82 × 10−37.56 × 10−31.57 × 10−27.15 × 10−37.35 × 10−39.84 × 10−21.51 × 10−15.95 × 10−28.06 × 10−2
PSOACSC [33]2.27 × 10−17.33 × 10−21.90 × 10−11.19 × 10−11.58 × 10−13.54 × 10−12.99 × 10−12.56 × 10−13.71 × 10−1
MOEAMCD [24]5.13 × 10−21.21 × 10−24.73 × 10−21.38 × 10−21.81 × 10−26.88 × 10−21.67 × 10−16.94 × 10−21.15 × 10−1
PMF [25]9.99 × 10−21.34 × 10−29.56 × 10−22.61 × 10−22.32 × 10−21.05 × 10−12.88 × 10−11.64 × 10−12.22 × 10−1
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yang, Y.; Gou, H.; Tan, M.; Feng, F.; Liang, Y.; Xiang, Y.; Wang, L.; Huang, H. Multi-Criterion Sampling Matting Algorithm via Gaussian Process. Biomimetics 2023, 8, 301. https://doi.org/10.3390/biomimetics8030301

AMA Style

Yang Y, Gou H, Tan M, Feng F, Liang Y, Xiang Y, Wang L, Huang H. Multi-Criterion Sampling Matting Algorithm via Gaussian Process. Biomimetics. 2023; 8(3):301. https://doi.org/10.3390/biomimetics8030301

Chicago/Turabian Style

Yang, Yuan, Hongshan Gou, Mian Tan, Fujian Feng, Yihui Liang, Yi Xiang, Lin Wang, and Han Huang. 2023. "Multi-Criterion Sampling Matting Algorithm via Gaussian Process" Biomimetics 8, no. 3: 301. https://doi.org/10.3390/biomimetics8030301

Article Metrics

Back to TopTop