Next Article in Journal
Non-Communication Decentralized Multi-Robot Collision Avoidance in Grid Map Workspace with Double Deep Q-Network
Previous Article in Journal
Pilot Study: Quantitative Photoacoustic Evaluation of Peripheral Vascular Dynamics Induced by Carfilzomib In Vivo
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Saliency Detection with Bilateral Absorbing Markov Chain Guided by Depth Information

1
Changchun Institute of Optics, Fine Mechanics and Physics, Chinese Academy of Sciences, Changchun 130033, China
2
School of Optoelectronics, University of Chinese Academy of Sciences, Beijing 100049, China
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(3), 838; https://doi.org/10.3390/s21030838
Submission received: 28 December 2020 / Revised: 19 January 2021 / Accepted: 22 January 2021 / Published: 27 January 2021
(This article belongs to the Section Sensing and Imaging)

Abstract

:
The effectiveness of depth information in saliency detection has been fully proved. However, it is still worth exploring how to utilize the depth information more efficiently. Erroneous depth information may cause detection failure, while non-salient objects may be closer to the camera which also leads to erroneously emphasis on non-salient regions. Moreover, most of the existing RGB-D saliency detection models have poor robustness when the salient object touches the image boundaries. To mitigate these problems, we propose a multi-stage saliency detection model with the bilateral absorbing Markov chain guided by depth information. The proposed model progressively extracts the saliency cues with three level (low-, mid-, and high-level) stages. First, we generate low-level saliency cues by explicitly combining color and depth information. Then, we design a bilateral absorbing Markov chain to calculate mid-level saliency maps. In mid-level, to suppress boundary touch problem, we present the background seed screening mechanism (BSSM) for improving the construction of the two-layer sparse graph and better selecting background-based absorbing nodes. Furthermore, the cross-modal multi-graph learning model (CMLM) is designed to fully explore the intrinsic complementary relationship between color and depth information. Finally, to obtain a more highlighted and homogeneous saliency map in high-level, we structure a depth-guided optimization module which combines cellular automata and suppression-enhancement function pair. This optimization module refines the saliency map in color space and depth space, respectively. Comprehensive experiments on three challenging benchmark datasets demonstrate the effectiveness of our proposed method both qualitatively and quantitatively.

1. Introduction

The salient object detection (SOD) is a fundamental task in computer vision, which attempts to imitate the human visual attention mechanism to locate and segment the interesting or attractive regions in a scene. It has been widely applied to a variety of vision tasks, such as image segmentation [1], resizing [2], enhancement [3], quality assessment [4], recognition [5], and matching [6]. In fact, the human visual system can not only intuitively capture the appearance of objects, but also perceive the depth information from the scene. Benefiting from the development of 3D sensing technology, the depth information can be captured more conveniently and accurately. Therefore, the RGB-D saliency detection using depth information is attracting more and more attention. Moreover, the effectiveness of depth information has been fully proved in other computer vision tasks, such as motion segmentation [7] and people re-identification [8].
Given a pair of RGB-D (RGB + depth) images, the task of the RGB-D saliency detection aims to predict a saliency map and extract the salient regions by exploring the complementary information between color image and depth data. Furthermore, existing RGB-D saliency detection models mainly use depth information in two ways. One is based on depth features [9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25], which focuses on taking depth information as an explicit supplementary feature of color features. In [12], Cheng et al. calculate the saliency map with additional depth information through color contrast, depth contrast, and spatial bias extended from 2D to 3D, which also proves that depth information is beneficial to visual saliency analysis in complex scenes. In order to fully explore the potential color and depth cues in the whole saliency processing process, Peng et al. [16] propose an evolution strategy to introduce depth information into super-pixel generation, initial saliency map generation, and saliency propagation. In [24], Fang et al. propose a united stereoscopic saliency model, which combines depth-guided background prior, boundary background, and compactness based on disparity to estimate the initial saliency map. The map is refined by using the spatial dissimilarity features under reduced dimensions and central preference. Zhu et al. [17,18] directly use the depth map to generate the depth feature saliency and merge it with the color features saliency, then optimize the saliency map by combining the center dark channel prior (CDCP) or background elimination model. In [21], Song et al. generate different saliency measures based on multi-level features at different scales and perform discriminative saliency fusion through a random forest regressor to obtain the final saliency result. Aiming at the problem that the robustness of the saliency detection algorithm is not satisfied in some complex situations containing multiple objects or complex background, Zhu et al. [20] propose a multilayer backpropagation algorithm based on depth mining, which extracts depth cues from four different saliency layers to improve performance.
The other is based on depth measurement [26,27,28,29,30,31,32,33,34,35,36], which aims to obtain implicit attributes such as shape and contour from the depth map by designing depth measurement algorithms. Ren et al. [27] propose the normalized depth prior and the global-context surface orientation prior. These prior can highlight near objects, weaken distant objects and reduce the saliency of severely inclined surfaces (such as the ground plane or ceilings). In [26], instead of using absolute depth, Ju et al. propose an anisotropic center-surround difference (ACSD) measure that considers the global depth structure to calculate and perceive the depth saliency map. Since the background usually contains the regions with a large change in depth compared to the neighborhood, this leads to a higher contrast in this region. In response to this problem, Feng et al. [28] design a local background enclosure (LBE) feature to capture the spread of angular directions, which quantifies the proportion of the object boundary that is in front of the background from the depth map. In [33], Wang et al. propose a multi-stage salient object detection framework based on minimum barrier distance transformation and multi-layer cellular automata (MCA). The framework integrates multiple visual features and priors including background prior, 3-D spatial prior and depth bias. In general, the depth-feature based method is an intuitive and simple to achieve the RGB-D saliency detection, which ignores the potential attributes in the depth map. By contrast, the depth-measurement based method aims to refine the saliency results by using implicit information.
However, limited by the technology of the depth sensor, not all depth information is accurate and practicable. In another word, when the depth maps are accurate, they can provide precise depth information to facilitate saliency detection, on the contrary, they may cause detection failure when the depth maps are poor. In order to handle this problem, Cong et al. [37] present a depth confidence measure to assess the reliability of the depth map and control the fusion ratio of depth features and color features in the saliency model. In addition, in [38], a novel saliency detection model is proposed that combines the implicit and explicit features of the depth map, its main idea is to transfer the existing RGB saliency detection model to RGB-D images with the help of depth constraint, so that it can inherit the saliency performance of RGB image. To a certain extent, the utilization efficiency of depth information is improved, but it also has a problem that the algorithm greatly relies on the performance of the RGB saliency detection algorithm. Therefore, how to effectively fuse depth information to enhance the detection of salient objects is still challenging. Moreover, the detection results of the above algorithms are mostly not ideal for scenes where the object touches the boundary.
To tackle these problems, we propose a saliency detection model with the bilateral absorbing Markov chain guided by depth information. The model includes three progressive processing stages. At the first stage, we explicitly combine depth features with color features to calculate the low-level saliency information based on background prior and contrast prior. In the second stage, we design a bilateral absorbing Markov chain model based on the background seed selection mechanism and cross-modal multi-graph learning model. In this stage, we can obtain mid-level foreground-based and background-based saliency maps by using low-level saliency cues of first stage. In the final stage, to further improve the performance of our algorithm, we propose a depth-guided optimization module to obtain a more homogeneous salient region.
The main contributions of our paper can be summarized as:
  • A multi-stage RGB-D saliency detection framework with the bilateral absorbing Markov chain model is proposed. The framework can make full use of the explicit and implicit information in the depth map and explore the complementary relationship between the modes.
  • The background seed screening mechanism is designed to solve the boundary touch problem. Moreover, the cross-modal multi-graph learning model is designed for implicitly fusing color and depth information by the learning.
  • To preferably highlight the salient regions, we design a depth-guided optimization module which combines cellular automata and suppression-enhancement function pair.

2. Methodology

This section describes the proposed method in detail, and the overall framework is shown in Figure 1. The algorithm mainly consists of four subsections: pre-processing, low-level saliency cues calculation, mid-level saliency maps generation and high-level saliency optimization.

2.1. Initial Two-Layer Sparse Graph Constrution

Given an RGB image and an aligned depth map, we first convert the RGB image to the CIELAB color space and segment it into N superpixels using mean shift [39] algorithm. The superpixel is a small region in the image composed of a series of adjacent pixels with similar features e.g., color, brightness, texture, etc. Then, we construct an initial two-layer sparse graph G = ( V , E ) such as [40], where V = { v i | 1 i N } denotes the nodes and E = { e i j | 1 i , j N } denotes the edges between nodes. The graph is generated by connecting each node to neighboring nodes and the most similar node sharing a common boundary with its neighboring nodes. It is worth to notice that the nodes on the four boundaries of the image are connected to each other to reduce the geodesic between the background nodes. As [40] proves, compared with the ordinary two-layer graph, the two-layer spares graph can effectively avoid the interference from surrounding redundant nodes.
In this work, we utilize the pre-trained FCN-32s network [41] to extract the color feature vector, the Euclidean distance c i j in RGB color space and depth difference d i j between superpixels i and j are defined as
c i j = x i x j
and
d i j = | d i d j |
where x i is the mean color feature vector of superpixel i , and d i denotes the mean depth value of superpixel i . The similarity a i j between superpixels i and j is defined as
a i j = a i j c ( a i j d ) ε
where the coefficient ε adjusts the weight of depth information and set as 0.5, a i j c and a i j d represent the color similarity and depth similarity respectively, and are defined as
a i j c = e c i j σ 2
and
a i j d = e d i j σ 2
where σ 2 is a parameter to control strength of the similarity which is set to 0.1. The affinity matrix W = [ w i j ] N × N of the graph is defined as the similarity between two superpixels,
w i j = { a i j , if j Ω i 0 , otherwise
where Ω i is the neighbors of superpixel i based on the initial two-layer sparse graph.

2.2. Low-Level Saliency Cues Calculation Using Color and Depth Cues

In this part, explicitly combining color and depth cues, we calculate low-level saliency information based on background prior and contrast prior. The saliency prior maps are shown in Figure 1.

2.2.1. Background Prior Calculation

We adopt boundary connectivity [42] to generate the background prior map, which is defined as
S b p ( i ) = 1 exp ( B n d C o n 2 ( i ) 2 σ b n d C o n 2 )
in which B n d C o n ( i ) refers to the value of boundary connectivity for superpixel i and σ b n d C o n is a weighting factor for boundary connectivity. Here empirically sets σ b n d C o n 2 = 1 . This background measure is robust to the normal cases and can effectively eliminate most background regions.

2.2.2. Region Contrast Prior Calculation

Human attention tends to focus on those image regions that contrast strongly with the surroundings. Therefore, we calculate a region contrast similar with [43], which integrates depth features and rich color features together. Then, compared to all other regions, we compute its saliency value by measuring its depth and color combined contrast,
S r c ( i ) = j = 1 , j i N a i j D o ( i , j ) A r e a ( j )
where D o ( i , j ) represents the Euclidean spatial distance between the superpixel i and j , A r e a ( v j ) is the area ratio of superpixel j compared with the whole image.

2.3. Mid-Level Saliency Maps Generation by Bilateral Absorbing Markov Chain

Inspired by [44], we design a bilateral absorbing Markov chain model, which combines multi-layer color features and depth features to obtain learned transition probability matrixes, and generate mid-level saliency maps. Most of the saliency models have poor detection results when the salient object is not in the center of the image, especially in the case of some salient regions touch the image boundary. To handle this situation in ours model, we propose a background seed screening mechanism (BSSM) to improve the graph model and better select background-based absorbing nodes. Moreover, we present a cross-modal multi-graph learning model (CMLM) to obtain the learned affinity and transition probability matrixes, which can make full use of the complementarity of color and depth information.

2.3.1. Absorbing Markov Chain for Saliency Detection

To facilitate the understanding, we give a brief introduction to the principle of absorbing Markov chain [45,46]. For a given set of states S = { s 1 , s 2 , , s k } , the probability of moving from state s i to the next state s j is expressed as the transition probability p i j , which does not depend on the chain before the current state. An absorbing Markov chain contains at least one absorbing state ( p i i = 1 ), and starts from every transient state, a certain absorbing state can be reached. For an absorbing chain with r absorbing states and t transient states, the canonical form of the transition matrix P is as follows,
P ( Q R 0 I )
where Q [ 0 , 1 ] t × t represents the transition probability of any pair of t transient states, while R [ 0 , 1 ] t × r represents the transition probability between any transient state and absorbing state. 0 is the r × t zero matrix and I is the r × r identity matrix. Furthermore, the fundamental matrix N is computed [45],
N = ( I Q ) 1 = I + Q + Q 2 +
where n i j of N can describe the expected number of times from transient state s i to transient state s j in the absorbing chain.
Then the absorption probability for each transient state to reach any absorbing state can be defined as [46],
B = N R
where b i j of B indicates the absorption probability from transient state s i to transient state s j .
Traditional saliency detection models based on absorbing Markov chain generally mirror image boundary superpixels as absorbing nodes (or states), and all others as transient nodes. Then, the transition matrix P is constructed according to the similarity (the transition probability) between nodes. The saliency value is measured by the absorption probability, the higher the absorption probability of the node, the more similar to the absorbing nodes.

2.3.2. Background Seed Screening Mechanism

Generally, traditional saliency detection models based on absorbing Markov chain [44,45,46] usually mirror image edge superpixels as absorbing nodes and simply connect all edge superpixels in pairs. However, as shown in Figure 2, when the salient object touches the image boundary, the mirroring will mistakenly regard the foreground nodes as background-based absorbing nodes, thus suppressing the saliency of the foreground regions or causing detection failure. Similarly, if the edge nodes contain foreground nodes, the full connections between them may be poorly robust. To overcome them, we propose a background seed screening mechanism (BSSM) for improving the two-layer sparse graph and selecting better background-based absorbing nodes. This mechanism removes the nodes that may belong to the foreground from the edge nodes. Furthermore, in order to increase the diversity of the background and restrain the background regions, a small number of random non-edge background nodes are selected to form a new edge node set and a background-based absorbing node set. Moreover, to obtain more homogeneous salient regions, we design the non-local connection similar to [47]. Next, we will introduce the construction process of the background seed screening mechanism and the non-local connections in detail.
To facilitate understanding, we provide a schematic diagram in Figure 3, which describes the main screening process of the background seeds. First, according to the attributes of saliency, position and depth, all nodes are classified as three categories. As shown in Figure 3a, based on the low-level background prior S b p , we divide all nodes into background seed set Ω B G , foreground seed set Ω F G and others.
Ω B G = { i | S b p ( i ) > 0.9 }
Ω F G = { i | S f p ( i ) > t h F G }
where S f p represents the foreground prior,
S f p ( i ) = 1 S b p ( i )
t h F G > 3 m e a n ( S f p ) + max ( S f p ) 4
According to the position attribute, the nodes can be classified as edge node set Ω e d g e = { i | i edge } and non-edge node set Ω n o n _ e d g e = { i | i edge } as shown in Figure 3b.
Considering that objects far away from the camera are likely to belong to the background, as shown in Figure 3c, we use the depth threshold to divide the nodes into depth-based background seed set Ω D e p and others.
Ω D e p = { i | d i > 1.2 t h D e p   and   i Ω F G }
t h Dep > 3 m e a n ( d i ) + max ( d i ) 4 , i Ω B G
To alleviate the boundary touch problem and select background seeds more accurately, we utilize k-means algorithm to filter out the foreground nodes in the background seed set Ω B G and edge node set Ω e d g e . More specifically, we cluster the sets of Ω B G , Ω e d g e and Ω F G to find the nodes that are similar with the foreground seeds Ω F G . Figure 3d is the filtered result: new edge node set Ω e d g e and background seed set Ω B G . In Figure 3e, we take depth information into consideration in the process of background seed screening. In Figure 3f, non-edge background seeds and depth-based background seeds are further divided into three sub-sets: Ω A , Ω B , and Ω C . It is obvious that the seeds in Ω A satisfy both background probability and depth with high values, while the seeds in Ω B and Ω C only satisfy the requirement of high background probability or high depth value, respectively.
Then, for guaranteeing the diversity of the background and suppressing the background more effectively, we combine a small number of non-edge nodes with Ω e d g e and further form the final edge nodes Ω f _ e d g e . These non-edge nodes are randomly composed of 50% Ω A , 10% Ω B , and 50% Ω C . In the initial two-layer sparse graph, to reduce the geodesic distances of nodes, all edge nodes are simply connected together. However, it may be poorly robust to the case when salient objects touch the image boundaries. Therefore, instead of the rough connections, we use the final edge nodes Ω f _ e d g e connected in pairs to obtain a new two-layer sparse graph G n e w . In addition, to obtain more consistent salient regions, we introduce the non-local connection into the graph. Specifically, it first sorts the foreground prior S f p and the region contrast prior S r c of all nodes, the top 50% of both are selected as foreground seeds, and the bottom 50% are selected as background seeds. For each superpixel, we connect it to two nodes that are randomly chosen from the two seed sets respectively. This connection mechanism is more conducive to highlight the foreground objects and suppress the background regions. The improved two-layer sparse graph with the non-local connection is visualized in Figure 4e.
Moreover, Figure 5 demonstrate the effects of the proposed background seed screening mechanism (BSSM) and non-local connection. In Figure 5e, it is clearly observed that the background is well suppressed by the improved two-layer sparse graph based on background seed screening mechanism (BSSM). Figure 5g illustrates that the non-local connection can achieve more complete and consistent salient regions.

2.3.3. Cross-Modal Multi-Graph Learning Model

The two-layer sparse graph constructs the connections among the local regions, which will restrict the range of random walk to the local regions. Therefore, the absorption time may be inaccurate, especially when the long-range smooth background distributes near the center of image. To overcome it, we have improved the graph model from the connection relationship in the above section. However, in the absorbing Markov chain model, another key influencing factor is the weight of the edges between nodes. Similar to Formula (3), most of the existing graph models directly weight depth and color cues to measure the similarity between nodes. However, the models do not consider the effect of color and depth information on saliency detection in different scenarios. For example, in some scenes, color is more reliable than depth, so a larger weight of color is needed. Conversely, if depth is more reliable, we need to strengthen the weight of depth. Therefore, we propose a cross-modal multi-graph learning model (CMLM), which fully explores the complementary relationship between color and depth in different scenarios. The learning model constructs a more accurate affinity matrix and captures the optimal fusion state of color and depth information.
Some algorithms [44,48] have constructed the affinity matrix by the learning. In [48], the learning model based on the single graph is proposed, which construct an approximate full affinity matrix by using the following equation,
min Y i , j = 1 N w i j y i y j 2 + μ i = 1 N y i i i 2
where Y = [ y i , y 2 , , y n ] N × N is an affinity matrix optimized by unsupervised learning based on the original sparse affinity matrix. y i = [ y i 1 , y i 2 , , y i N ] is a column vector indicating the degree of affinity between the node i and all other nodes, i i is the i - th column of an identity matrix Ι which indicates the similarity with itself. In Equation (18), the first item is a smoothing constraint item, which indicates the difference between y i and y j . The two nodes are more similar, the value of first item will be lower. The second item is a self-restraint item, which emphasizes that no matter how we update the value of y i of node i , it should not be too different from its initial value. μ is a parameter that balances the relationship between the two items, μ > 0 .
Formula (18) is the learning process under the single-layer graph. To make full use of the complementarity of color and depth information, we explore feature spaces of multiple modes and develop a cross-modal multi-graph model to learn an affinity matrix. We use β = [ β c , β d , ] to represent the set of multi-modal vectors, and its values indicate the importance of the corresponding affinity graph. In this work, we only adopt the modes of color and depth. β c = [ β c ( 1 ) , β c ( 2 ) , , β c ( m ) ] is a sub-vector of the color mode and m is the number of feature maps in color space. β d = [ β d ( 1 ) , β d ( 2 ) , , β d ( n ) ] is a sub-vector of the depth mode and n is the number of feature maps in depth space. W c ( ν ) = [ w i j c ( ν ) ] N × N is the graph affinity matrix computed by the ν - th color feature and W d ( τ ) = [ w i j d ( τ ) ] N × N is the graph affinity matrix computed by the τ - th depth feature. Then, the final learning affinity matrix optimization equation can be defined as
min β , Y ν = 1 m ( β c ( ν ) ) γ i , j = 1 N w i j c ( ν ) y i y j 2 + τ = 1 n ( β d ( τ ) ) γ i , j = 1 N w i j d ( τ ) y i y j 2 + μ i = 1 N y i i i 2 ,
s . t . ν = 1 m β c ( ν ) + τ = 1 n β d ( τ ) = 1 , 0 β c ( ν ) , β d ( τ ) 1
where the parameter γ controls the weight distribution of all affinity matrices, ensuring that different-mode features can be fully utilized. Without this parameter, in some cases, it is possible that only partial features participate in the learning of affinity matrix, which may utilize the complementarity between different features insufficiently. The parameter μ and γ are set to 0.001 and 4 respectively. To facilitate the derivation, we rewrite the above objective function (19) in the form of matrix,
J = ν = 1 m ( β c ( ν ) ) γ T r ( Y T L c ( ν ) Y ) + τ = 1 n ( β d ( τ ) ) γ T r ( Y T L d ( τ ) Y ) + μ Y I F 2
where L c ( ν ) = D c ( ν ) W c ( ν ) is the graph Laplacian matrix of the ν - th color feature, D c ( ν ) is the degree matrix and d i i c ( ν ) = j = 1 N w i j c ( ν ) . Similarly, L d ( τ ) = D d ( τ ) W d ( τ ) is the graph Laplacian matrix of the τ - th depth feature, D d ( τ ) is the degree matrix and d i i d ( ν ) = j = 1 N w i j d ( ν ) . T r ( ) and F compute the trace and the Frobenius norm of the matrix separately. We can see that there are two unknown items β and Y to be solved in Equation (20), so we decompose it into two sub-problems to solve this optimization problem by iteration.
Fix β , Update Y :
Y = μ ( ν = 1 m β c ( ν ) L c ( ν ) + τ = 1 n β d ( τ ) L d ( τ ) + μ I ) 1
Fix Y , Update β :
β c ( ν ) = ( T r ( Y T L c ( ν ) Y ) ) 1 1 γ ν = 1 m ( T r ( Y T L c ( ν ) Y ) ) 1 1 γ + τ = 1 n ( T r ( Y T L d ( τ ) Y ) ) 1 1 γ
β d ( τ ) = ( T r ( Y T L d ( τ ) Y ) ) 1 1 γ ν = 1 m ( T r ( Y T L c ( ν ) Y ) ) 1 1 γ + τ = 1 n ( T r ( Y T L d ( τ ) Y ) ) 1 1 γ
To get the optimal solution of sub-problems, we utilize partial derivative and Lagrange Multiplier Method. The specific derivation process can refer to [48]. With the learned affinity matrix Y , we can calculate the transition matrixes of absorbing Markov chain. The final learned affinity matrix W L = [ w i j L ] N × N is obtained by normalization,
W L = d i a g ( Y ) 1 × Y
Figure 6d shows the effects of the proposed cross-modal multi-graph learning model (CMLM). As it is illustrated, compared to single-mode multi-graph learning model (color mode), the proposed model is more precise to highlight the salient regions.

2.3.4. Background-Based Saliency Map via Absorbing Markov Chain

In this part, we select background-based absorbing nodes based on the above background seed screening mechanism. As is presented in Figure 7a, we mirror edge nodes Ω e d g e and some non-edge background nodes as virtual absorbing nodes, and all nodes in the image as transient nodes. The non-edge background nodes are randomly composed of 50% Ω A , 50% Ω B and 50% Ω C . The number of absorbing nodes is r. Then, the background-based affinity matrix W L B = [ w i j L ] N × r can be obtained with Formula (24). Furthermore, the learned transition matrix is defined as
P B = [ Q B N × N R B N × r 0 r × N I B r × r ]
where Q B N × N = D B 1 W L , R B N × r = D B 1 W L B , D B is the sum of the matrix D 1 and D 2 , D 1 = d i a g { d 1 W L , d 2 W L , , d N W L } is the degree matrix of W L , and d i W L = i = 1 N w i j L . D 2 = d i a g { d 1 W L B , d 2 W L B , , d N W L B } is the degree matrix of W L B , and d i W L B = i = 1 r w i j L . According to Formula (11), we can calculate the absorption probability matrix B B = N B R B , where N B = ( I Q B ) 1 . Based on the above work, the saliency of the node i is defined as
S b g ( i ) = 1 j = 1 r b i j
The background-based saliency map S b g is shown in Figure 1. Then, we mirror the nodes with the saliency value greater than the threshold t h as the foreground-based absorbing nodes, which is illustrated in Figure 7b. The number of absorbing nodes is k.
t h = ( m e a n ( S b g ) + max ( S b g ) ) / 2

2.3.5. Foreground-Based Saliency via Absorbing Markov Chain

Similarly, the foreground-based affinity matrix W L F = [ w i j L ] N × k can be obtained with Formula (24), and the learned transition matrix is as follows,
P F = [ Q F N × N R F N × k 0 k × N I F k × k ]
where Q F N × N = D F 1 W L , R F N × k = D F 1 W L F , D F is the sum of the matrix D 1 and D 2 , D 2 = d i a g { d 1 W L F , d 2 W L F , , d N W L F } is the degree matrix of W L F , d i W L F = i = 1 k w i j L . According to Formula (11), the absorption probability matrix B F = N F R F is obtained, where N F = ( I Q F ) 1 . In order to calculate the foreground-based saliency more accurately and eliminate the interference of weak correlated nodes, we sort each row of B F and select the top 60% of the nodes to calculate the final saliency value,
S f g ( i ) = i = 1 c b i j
where c = 0.6 k , and the foreground-based saliency map S f g is shown in Figure 1.

2.4. High-Level Saliency Map Optimization via Depth Guidance

In order to further highlight the salient regions and effectively explore the inner relationship between depth information and salient information, we design a depth-guided optimization module which combines cellular automata and suppression-enhancement function pair.

2.4.1. Optimization via Cellular Automata

We perform a primary fusion of the saliency maps produced by the bilateral absorbing Markov chain model,
S f b ( i ) = 0.4 S f g ( i ) + 0.6 S b g ( i )
Based on the improved two-layer sparse graph, we use the cellular automata [49] propagation mechanism to further optimize the fused saliency map. First, based on the learned affinity matrix W L and the color similarity matrix A c = [ a i j c ] N × N , we construct an impact factor matrix F = [ f i j ] N × N ,
F = A c W L
Furthermore, all superpixel nodes (cells) are updated simultaneously through the following iteration rules,
S h + 1 = C S h + ( I C ) F * S h
where I is the identity matrix. F = [ f i j ] N × N and C = d i a g { c 1 , c 2 , , c N } are normalized impact factor matrix and coherence matrix respectively,
F = D f 1 F
c i = a n o r m ( 1 / max ( f i j ) ) + b
where D f = d i a g { d f 1 , d f 1 , , d f N } is the degree of the matrix and d f i = j f i j . The constant coefficients a and b are set to 0.6 and 0.2, respectively, n o r m ( ) means normalization function. Each cell can automatically evolve into a more accurate and stable state, and under the influence of the neighborhood, the salient regions are easier to be detected. The initial S h when h = 0 is S f b in Equation (30), and the ultimate saliency map after h = 10 time steps is denoted as S C A , which is visualized in Figure 8g.

2.4.2. Refinement via Depth Information

Cellular automata mainly explores the neighborhood relationship between the nodes in the color feature space, but ignores the spatial position information in the scene. Therefore, we use depth cues to enhance and refine the salient regions and suppress the background regions. In this work, we design a depth selective refinement mechanism by a suppression–enhancement function pair: the suppression function is used to suppress the background, and then an enhancement function is used to emphasize the salient regions through high-confidence depth seeds.
Suppression function: The regions far away from the camera have a higher probability of being the background and need to be suppressed. Therefore, we defined the suppression function as follows,
S S F ( i ) = { S C A ( i ) , if   S C A ( i ) > 0.7 and   S d ( i ) > 0.5 and   S C A ( i ) > ( t h C A + 0.1 ) S C A ( i ) S d ( i ) , if   S C A ( i ) t h C A S C A ( i ) S d ( i ) , otherwise
where t h C A is the adaptive threshold of the saliency map S C A obtained by Otsu [50] algorithm, and S d ( i ) = n o r m ( d i ) is the depth prior. After filtering S S F through the Otsu algorithm, the suppressed saliency map S 1 is obtained.
Enhancement function: Although the suppression function inhibits background information to a certain extent, it may lose some saliency information. The enhancement function can play a complementary role. First of all, we need to determine which depth information is reliable and needs to be retained. Here we combine three saliency maps to filter out the potential depth seed set Ω D with high confidence. The depth seeds are all salient in saliency maps of S f g , S b g and S C A . The enhancement function is defined as follows,
S E F ( i ) = { S 1 ( i ) + 0.2 S d ( i ) , S d ( i ) 0.9 , i Ω D S 1 ( i ) + 0.05 S d ( i ) , S d ( i ) < 0.9 , i Ω D S 1 ( i ) , otherwise
After the suppression–enhancement function pair, we can get the final saliency map S E F , which is shown in Figure 8h.

3. Experiments and Discussion

3.1. Datasets

In this part, in order to effectively demonstrate our proposed algorithm, we evaluate the model in three most popular datasets, including NLPR [13], NJU2K [26], and STERE [9]. The NLPR dataset includes 1000 RGB-D images, where the depth maps are captured by Microsoft Kinect. The NJU2K dataset contains 1985 RGB-D images which are collected from the Internet, 3-D movies and photographs taken by stereo camera, and depth maps are estimated by the optical-flow method. The STERE dataset contains 1000 stereoscopic images with the corresponding pixel-level ground truth.

3.2. Evaluation Metrics

Following [51], we use the following five popular evaluation metrics to evaluate the performance of the saliency detection methods comprehensively.
MAE estimates a mean absolute error between a predicted saliency map S and ground-truth map GT, it is defined as
MAE = 1 W H x = 1 W y = 1 H | S ( x , y ) G T ( x , y ) |
where H and W are the height and the width of the saliency map.
PR curve is formed by a series of pairs of precision and recall scores calculated at fixed thresholds ranging from 0 to 255, which describes the model performance at different situations.
p r e c i s i o n = | S G T | S
r e c a l l = | S G T | G T
F-measure is a harmonic mean of average precision and recall, which is defined as,
F β = ( 1 + β 2 ) p r e c i s i o n r e c a l l β 2 p r e c i s o n + r e c a l l
We empirically set β 2 = 0.3 .
S-measure [52] is used to measure the spatial structure information, which is defined as,
S α = α S 0 + ( 1 α ) S r
where α is a balance parameter between the object-aware structural similarity S 0 and region-aware structural similarity S r , and it is set to 0.5.
E-measure [53] is to evaluate the foreground map (FM) and noise, which combines local pixel values with image-level mean values to jointly capture image-level statistics and local pixel matching information.
E m = 1 W H x = 1 W y = 1 H ϕ F M ( x , y )
where ϕ is an enhanced alignment matrix for the two properties of a binary map.

3.3. Ablation Study

Our algorithm combines background seed screening mechanism, non-local connection, cross-modal multi-graph learning model, and depth-guided optimization module. To further demonstrate the effectiveness of the components, a series of experiments are carried out. Figure 9 shows all the results of the above experiments intensively. In this part, we will combine the two-layer graph and the bilateral absorbing Markov chain based on single-modal multi-graph learning as the baseline model, which is the combination 1 in Figure 9. As is illustrated in Figure 9, the two-layer sparse graph and the background seed screening mechanism greatly improve the performance of our algorithm, which can be observed from the combinations 1, 2 and 3. Compared to the two-layer graph, the two-layer sparse graph suppresses most of the background better in Figure 5d. From Figure 5d, based on the background seed screening mechanism, background is further diluted, and the foreground is further strengthened. Compared with combination 3, the cross-modal multi-graph learning model has better improvement in precision-recall and S-measure, but the other evaluation parameters may be slightly lower. From comprehensive perspective, the cross-modal multi-image learning model and depth guided optimization module can achieve the best results which can refer to combinations 5 and 6. As Figure 6d shows, compared to the single-mode multi-graph learning model (color mode), the cross-mode multi-graph learning model can better pop foreground objects from various scenes. Figure 8h displays the effect of the depth-guided optimization module. Finally, from combinations 7 and 8, it obvious that the non-local connection can effectively improve the overall performance of the algorithm. The saliency maps with the non-local connection are more precise as shown in Figure 5g.

3.4. Comparisions with State-of-the-Art Methods

We compare our proposed algorithm with 10 state-of-the-art RGB-D saliency detection models, including ACSD [26], DESM [12], LHM [13], GP [27], DCMC [37], LBE [28], SE [16], CDCP [18], CDB [24], and DTM [38]. For fair comparison, we employ saliency maps provided by the [51]. Table 1 and Figure 10 show the quantitative results of different RGBD saliency detection models. We also report saliency maps with various scenes as shown in Figure 11.
We report PR curves of three datasets in Figure 10 and list S α , E m , F β and MAE in Table 1. As shown in Figure 10, our method achieves better PR curves on the three datasets, especially on NLPR and STERE datasets. This indicates that our method can obtain higher precision and recall compared with other methods. On the NJU2K dataset, although the end of our PR curve drops faster than some methods, we always maintain a robust curve on each dataset and keep a good balance between precision and recall overall.
As listed in Table 1, we can intuitively observe the superiority of our method among all the methods, which can be proved with the best results over all the three datasets. This demonstrates that our algorithm can generate more accurate salient regions and is more adaptable to various scenes than others.
In addition to the quantitative comparisons, to prove the effectiveness of our model visually, we also display some saliency maps in Figure 11. As we can see, the most saliency detection methods can effectively handle the cases with relatively simple backgrounds and homogenous objects. However, these methods fail to handle the complicated cases. In contrast, our method can deal with these intricate scenarios more effectively. To make it more convincing, we compare these methods in the following four aspects: (1) the effectiveness of dealing with boundary touch issues; (2) the effectiveness of the background suppression; (3) the effectiveness of solving similar appearances; and (4) the effectiveness of detection with a poor depth map.
Here combined with examples to vividly expand the above four aspects. First, as shown in the 7-th and 8-th rows of Figure 11a, the 3-th, 5-th, and 7-th rows of Figure 11b, and the 8-th row of Figure 11c, only the GP algorithm has certain resistance to boundary touch problem, but when the background is complex and the depth map is poor, as shown in the 3-th rows of Figure 11b, the detection will fail. In contrast, our algorithm achieves better results in various scenes when encountering this situation. Then, from the 3-th, 4-th, and 6-th rows of Figure 11a, we can find that most of the algorithms cannot effectively remove the background in front of the salient objects due to the interference from the depth near the camera. However, our method can availably eliminate them by using learning fusion. Moreover, as shown in the 7-th row of Figure 11b, the 8-th and 10-th rows of Figure 11c, our method works well when the color appearance of salient object is similar to the background. Finally, our model is still robust under the condition of poor depth map quality, which is demonstrated in the 3-th and 4-th rows of Figure 11b, the 1-th, 2-th, 3-th, and 6-th rows of Figure 11c.
In general, our algorithm has better robustness in the various complex scenarios. Especially, when the salient objects touch the image boundary or the depth map quality in the dataset is uneven, our method still has a good performance, which can obtain the uniform and highlighted salient objects.
Computational complexity. We utilize the computational complexity to prove the advantages of our proposed method compared to other methods (traditional-based and deep learning-based). In this paper, we adopt the floating point operations (FLOPs) to measure the computational complexity of the models. For fair comparisons, we obtain the deployment codes released by authors and use the same configuration as much as possible to estimate their computational complexity. As illustrated in Table 2, compared with the latest deep learning-based methods such as D3Net [51], BBS-Net [54], and UC-Net [55], our computational complexity is only one tenth or even one hundredth of theirs. Moreover, compared with the traditional-based methods such as DCMC [37], CDCP [18], and DTM [38], our model can achieve obvious higher performance in the relatively lower computational complexity combined with Table 1.

4. Conclusions and Future Work

In this paper, we propose a RGB-D saliency detection model with the bilateral absorbing Markov chain guided by depth information. Using the explicit combination of depth and color information, we first generate the low-level saliency cues based on the background prior and contrast prior. Then, to overcome the existing drawbacks in the absorbing Markov chain model, we propose a serial of methods: the background seed screening mechanism (BSSM) for boundary touch cases and the cross-modal multi-graph learning model for multi-modal fusion. Moreover, considering the limitation of local intrinsic correlation, a non-local intrinsic correlation is introduced to improved two-layer sparse graph. Based on the optimized bilateral absorbing Markov chain model, we obtain the mid-level saliency maps. Finally, we design a depth-guided optimization module to get more accurate high-level saliency map. The optimization module consists of two sub-modules: the cellular automata to optimize the integrated saliency map in the color space and the suppression-enhancement function pair to refine the saliency map in the depth space. Compared with most of the algorithms mentioned in this article, our method alleviates the boundary touch problem well and greatly suppresses the background. The comprehensive comparisons and ablation study on three RGB-D saliency detection datasets have demonstrated that the proposed method is effective and robust in various scenarios both qualitatively and quantitatively.
The literature [51] builds a new salient person (SIP) dataset with quite challenging which covers diverse real-world scenes from various viewpoints, poses, occlusion, illumination, and background. Moreover, deep learning-based RGB-D saliency detection methods [51,54,55] have developed vigorously and achieved the qualitative leap. Therefore, we look forward to extending our work to the deep learning in the future, exploring the complementarity of depth information and color information more fully, and dedicating ourselves to the studying of the saliency detection algorithm in real-world scenes.

Author Contributions

Conceptualization, J.W.; validation, G.H., P.L., and H.Y.; formal analysis, J.W.; investigation, J.W., G.H., P.L., and H.Y.; original draft preparation, J.W.; review and editing, J.W., G.H., P.L., H.Y., H.L., and Q.L.; funding acquisition, G.H., P.L., and H.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China, grant number 61602432 and 61401425.

Data Availability Statement

Data sharing not applicable.

Acknowledgments

The authors would like to thank the anonymous reviewers for their constructive comments.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Liu, Z.; Shi, R.; Shen, L.; Xue, Y.; Ngan, K.N.; Zhang, Z. Unsupervised Salient Object Segmentation Based on Kernel Density Estimation and Two-Phase Graph Cut. IEEE Trans. Multimed. 2012, 14, 1275–1289. [Google Scholar] [CrossRef]
  2. Achanta, R.; Susstrunk, S. Saliency Detection for Content-Aware Image Resizing. In Proceedings of the 2009 16th IEEE International Conference on Image Processing (ICIP), Cairo, Egypt, 7–10 November 2009; pp. 1005–1008. [Google Scholar]
  3. Li, C.; Guo, J.; Cong, R.; Pang, Y.; Wang, B. Underwater Image Enhancement by Dehazing With Minimum Information Loss and Histogram Distribution Prior. IEEE Trans. Image Process. 2016, 25, 5664–5677. [Google Scholar] [CrossRef] [PubMed]
  4. Jiang, Q.; Shao, F.; Gao, W.; Chen, Z.; Jiang, G.; Ho, Y.-S. Unified No-Reference Quality Assessment of Singly and Multiply Distorted Stereoscopic Images. IEEE Trans. Image Process. 2019, 28, 1866–1881. [Google Scholar] [CrossRef]
  5. Zhang, H.; Cao, X.; Wang, R. Audio Visual Attribute Discovery for Fine-Grained Object Recognition. In Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2017; pp. 7542–7549. [Google Scholar]
  6. Citak, E.; Bilgin, G. Visual Saliency Aided SAR and Optical Image Matching. In Proceedings of the 2019 Innovations in Intelligent Systems and Applications Conference (ASYU), Izmir, Turkey, 31 October–2 November 2019; pp. 1–5. [Google Scholar]
  7. Muthu, S.; Tennakoon, R.; Rathnayake, T.; Hoseinnezhad, R.; Suter, D.; Bab-Hadiashar, A. Motion Segmentation of RGB-D Sequences: Combining Semantic and Motion Information Using Statistical Inference. IEEE Trans. Image Process. 2020, 29, 5557–5570. [Google Scholar] [CrossRef]
  8. Patruno, C.; Marani, R.; Cicirelli, G.; Stella, E.; D’Orazio, T. People re-identification using skeleton standard posture and color descriptors from RGB-D data. Pattern Recognit. 2019, 89, 77–90. [Google Scholar] [CrossRef]
  9. Niu, Y.; Geng, Y.; Li, X.; Liu, F. Leveraging stereopsis for saliency analysis. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 454–461. [Google Scholar]
  10. Ciptadi, A.; Hermans, T.; Rehg, J. An In Depth View of Saliency. In Proceedings of the British Machine Vision Conference 2013, Bristol, UK, 9–13 September 2013; British Machine Vision Association and Society for Pattern Recognition: Durham, UK, 2013; pp. 112.1–112.11. [Google Scholar]
  11. Desingh, K.; Madhava, K.K.; Rajan, D.; Jawahar, C.V. Depth really Matters: Improving Visual Salient Region Detection with Depth. In Proceedings of the British Machine Vision Conference 2013, Bristol, UK, 9–13 September 2013; Machine Vision Association and Society for Pattern Recognition: Durham UK, 2013; pp. 91–98. [Google Scholar]
  12. Cheng, Y.; Fu, H.; Wei, X.; Xiao, J.; Cao, X. Depth Enhanced Saliency Detection Method. In Proceedings of the 2014 International Conference on Internet Multimedia Computing and Service, Xiamen, China, 10–12 July 2014; pp. 23–27. [Google Scholar]
  13. Peng, H.; Li, B.; Xiong, W.; Hu, W.; Ji, R. Rgbd salient object detection: A benchmark and algorithms. In Computer Vision—ECCV, Proceedings of the 2004 European Conference on Computer Vision, Prague, Czech Republic, 11–14 May 2004; Springer: Cham, Switzerland, 2014; pp. 92–109. [Google Scholar]
  14. Fan, X.; Liu, Z.; Sun, G. Salient region detection for stereoscopic images. In Proceedings of the 2014 19th International Conference on Digital Signal Processing, Hong Kong, China, 20–23 August 2014; pp. 454–458. [Google Scholar]
  15. Sheng, H.; Liu, X.; Zhang, S. Saliency analysis based on depth contrast increased. In Proceedings of the 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Shanghai, China, 20–25 March 2016; pp. 1347–1351. [Google Scholar]
  16. Quo, J.; Ren, T.; Bei, J. Salient object detection for RGB-D image via saliency evolution. In Proceedings of the 2016 IEEE International Conference on Multimedia and Expo (ICME), Seattle, WA, USA, 11–15 July 2016; pp. 1–6. [Google Scholar]
  17. Zhu, C.; Li, G. A Three-Pathway Psychobiological Framework of Salient Object Detection Using Stereoscopic Technology. In Proceedings of the 2017 IEEE International Conference on Computer Vision Workshop (ICCVW), Venice, Italy, 22–29 October 2017; pp. 3008–3014. [Google Scholar]
  18. Zhu, C.; Li, G.; Wang, W.; Wang, R. An Innovative Salient Object Detection Using Center-Dark Channel Prior. In Proceedings of the 2017 IEEE International Conference on Computer Vision Workshops (ICCVW), Venice, Italy, 22–29 October 2017; pp. 1509–1515. [Google Scholar]
  19. Qu, L.; He, S.; Zhang, J.; Tian, J.; Yang, Q. RGBD Salient Object Detection via Deep Fusion. IEEE Trans. Image Process. 2016, 26, 2274–2285. [Google Scholar] [CrossRef]
  20. Zhu, C.; Li, G.; Guo, X.; Wang, W.; Wang, R. A Multilayer Backpropagation Saliency Detection Algorithm Based on Depth Mining. In Computer Analysis of Images and Patterns; Felsberg, M., Heyden, A., Krüger, N., Eds.; Springer International Publishing: Cham, Switzerland, 2017; pp. 14–23. [Google Scholar]
  21. Hangke, S.; Liu, Z.; Du, H.; Sun, G.; Le Meur, O.; Ren, T. Depth-Aware Salient Object Detection and Segmentation via Multiscale Discriminative Saliency Fusion and Bootstrap Learning. IEEE Trans. Image Process. 2017, 26, 4204–4216. [Google Scholar]
  22. Tang, C.; Hou, C. RGBD salient object detection by structured low-rank matrix recovery and Laplacian constraint. Trans. Tianjin Univ. 2017, 23, 176–183. [Google Scholar] [CrossRef]
  23. Zhu, C.; Cai, X.; Huang, K.; Li, T.H.; Li, G. PDNet: Prior-Model Guided Depth-Enhanced Network for Salient Object Detection. In Proceedings of the 2019 IEEE International Conference on Multimedia and Expo (ICME), Shanghai, China, 8–12 July 2019; pp. 199–204. [Google Scholar]
  24. Liang, F.; Duan, L.; Ma, W.; Qiao, Y.; Cai, Z.; Qing, L. Stereoscopic Saliency Model using Contrast and Depth-Guided-Background Prior. Neurocomputing 2018, 275, 2227–2238. [Google Scholar] [CrossRef]
  25. Huang, P.; Shen, C.H.; Hsiao, H.F. RGBD Salient Object Detection using Spatially Coherent Deep Learning Framework. In Proceedings of the 2018 IEEE 23rd International Conference on Digital Signal Processing (DSP), Shanghai, China, 19–21 November 2018; pp. 1–5. [Google Scholar]
  26. Ju, R.; Ge, L.; Geng, W.; Ren, T.; Wu, G. Depth saliency based on anisotropic center-surround difference. In Proceedings of the 2014 IEEE international conference on image processing (ICIP), Paris, France, 27–30 October 2014; pp. 1115–1119. [Google Scholar]
  27. Ren, J.; Gong, X.; Yu, L.; Zhou, W.; Yang, M.Y. Exploiting global priors for RGB-D saliency detection. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Boston, MA, USA, 7–12 June 2015; pp. 25–32. [Google Scholar]
  28. Feng, D.; Barnes, N.; You, S.; Mccarthy, C. Local Background Enclosure for RGB-D Salient Object Detection. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 2343–2350. [Google Scholar]
  29. Du, H.; Liu, Z.; Song, H.; Mei, L.; Xu, Z. Improving RGBD Saliency Detection Using Progressive Region Classification and Saliency Fusion. IEEE Access 2016, 4, 8987–8994. [Google Scholar] [CrossRef]
  30. Shigematsu, R.; Feng, D.; You, S.; Barnes, N. Learning RGB-D Salient Object Detection using background enclosure, depth contrast, and top-down features. In Proceedings of the 2017 IEEE International Conference on Computer Vision Workshops (ICCVW), Venice, Italy, 22–29 October 2017; pp. 2749–2757. [Google Scholar]
  31. Han, J.; Chen, H.; Liu, N.; Yan, C.; Li, X. CNNs-Based RGB-D Saliency Detection via Cross-View Transfer and Multiview Fusion. IEEE Trans. Cybern. 2018, 48, 3171–3183. [Google Scholar] [CrossRef] [PubMed]
  32. Cong, R.; Lei, J.; Fu, H.; Lin, W.; Huang, Q.; Cao, X.; Hou, C. An Iterative Co-Saliency Framework for RGBD Images. IEEE Trans. Cybern. 2019, 49, 233–246. [Google Scholar] [CrossRef] [Green Version]
  33. Wang, A.; Wang, M. RGB-D Salient Object Detection via Minimum Barrier Distance Transform and Saliency Fusion. IEEE Signal Process. Lett. 2017, 24, 663–667. [Google Scholar] [CrossRef]
  34. Chen, H.; Li, Y.F.; Su, D. Attention-Aware Cross-Modal Cross-Level Fusion Network for RGB-D Salient Object Detection. In Proceedings of the 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 1–5 October 2018; pp. 6821–6826. [Google Scholar]
  35. Hao, C.; Youfu, L.; Dan, S. Multi-modal Fusion Network with Multi-scale Multi-path and Cross-modal Interactions for RGB-D Salient Object Detection. Pattern Recognit. 2018, 86, 376–385. [Google Scholar]
  36. Chen, H.; Li, Y. Progressively Complementarity-Aware Fusion Network for RGB-D Salient Object Detection. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 3051–3060. [Google Scholar]
  37. Cong, R.; Lei, J.; Zhang, C.; Huang, Q.; Cao, X.; Hou, C. Saliency Detection for Stereoscopic Images Based on Depth Confidence Analysis and Multiple Cues Fusion. IEEE Signal Process. Lett. 2016, 23, 819–823. [Google Scholar] [CrossRef] [Green Version]
  38. Cong, R.; Lei, J.; Fu, H.; Hou, J.; Huang, Q.; Kwong, S. Going From RGB to RGBD Saliency: A Depth-Guided Transformation Model. IEEE Trans. Cybern. 2020, 50, 3627–3639. [Google Scholar] [CrossRef] [PubMed]
  39. Comaniciu, D.; Meer, P. Mean shift: A robust approach toward feature space analysis. IEEE Trans. Pattern Anal. Mach. Intell. 2002, 24, 603–619. [Google Scholar] [CrossRef] [Green Version]
  40. Zhou, L.; Yang, Z.; Zhou, Z.; Hu, D. Salient Region Detection Using Diffusion Process on a Two-Layer Sparse Graph. IEEE Trans. Image Process. 2017, 26, 5882–5894. [Google Scholar] [CrossRef]
  41. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  42. Zhu, W.; Liang, S.; Wei, Y.; Sun, J. Saliency optimization from robust background detection. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 2814–2821. [Google Scholar]
  43. Cheng, M.-M.; Mitra, N.J.; Huang, X.; Torr, P.H.; Hu, S.-M. Global contrast based salient region detection. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 569–582. [Google Scholar] [CrossRef] [Green Version]
  44. Zhang, L.; Ai, J.; Jiang, B.; Lu, H.; Li, X. Saliency Detection via Absorbing Markov Chain With Learnt Transition Probability. IEEE Trans. Image Process. 2018, 27, 987–998. [Google Scholar] [CrossRef]
  45. Jiang, B.; Zhang, L.; Lu, H.; Yang, C.; Yang, M.H. Saliency Detection via Absorbing Markov Chain. In Proceedings of the 2013 IEEE International Conference on Computer Vision, Sydney, NSW, Australia, 1–8 December 2013; pp. 1665–1672. [Google Scholar]
  46. Sun, J.; Lu, H.; Liu, X. Saliency Region Detection Based on Markov Absorption Probabilities. IEEE Trans. Image Process. 2015, 24, 1639–1649. [Google Scholar] [CrossRef]
  47. Luo, H.; Han, G.; Liu, P.; Wu, Y. Salient Region Detection Using Diffusion Process with Nonlocal Connections. Appl. Sci. 2018, 8, 2526. [Google Scholar] [CrossRef] [Green Version]
  48. Bai, S.; Sun, S.; Bai, X.; Zhang, Z.; Tian, Q. Smooth Neighborhood Structure Mining on Multiple Affinity Graphs with Applications to Context-Sensitive Similarity. In Computer Vision–ECCV 2016, Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 8–16 October 2016; Springer: Cham, Switzerland, 2016; pp. 592–608. [Google Scholar]
  49. Qin, Y.; Lu, H.; Xu, Y.; Wang, H. Saliency detection via Cellular Automata. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 110–119. [Google Scholar]
  50. Nobuyuki, O. A Threshold Selection Method from Gray-Level Histograms. IEEE Trans. Syst. Man Cybern. 1979, 9, 62–66. [Google Scholar]
  51. Fan, D.P.; Lin, Z.; Zhao, J.X.; Liu, Y.; Zhang, Z.; Hou, Q.; Zhu, M.; Cheng, M.M.J.A. Rethinking RGB-D Salient Object Detection: Models, Data Sets, and Large-Scale Benchmarks. IEEE Trans. Neural Netw. Learn. Syst. 2020. [Google Scholar] [CrossRef] [PubMed]
  52. Fan, D.; Cheng, M.; Liu, Y.; Li, T.; Borji, A. Structure-Measure: A New Way to Evaluate Foreground Maps. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 4558–4567. [Google Scholar]
  53. Fan, D.; Gong, C.; Cao, Y.; Ren, B.; Cheng, M.; Borji, A. Enhanced-alignment Measure for Binary Foreground Map Evaluation. arXiv 2018, arXiv:1805.10421. [Google Scholar]
  54. Fan, D.-P.; Zhai, Y.; Borji, A.; Yang, J.; Shao, L. BBS-Net: RGB-D Salient Object Detection with a Bifurcated Backbone Strategy Network. In Computer Vision–ECCV 2020, Proceedings of the European Conference on Computer Vision 2020, Glasgow, UK, 23–28 August 2020; Springer: Cham, Switzerland, 2020; pp. 275–292. [Google Scholar]
  55. Zhang, J.; Fan, D.-P.; Dai, Y.; Anwar, S.; Saleh, F.; Aliakbarian, S.; Barnes, N.J.A.P.A. UC-Net: Uncertainty Inspired RGB-D Saliency Detection via Conditional Variational Autoencoders. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 8579–8588. [Google Scholar]
Figure 1. Flowchart of the proposed method. BSSM: background seed screening mechanism; CMLM: cross-modal multi-graph learning model; bgAMC and fgAMC denote background-based and foreground-based saliency maps based on absorbing Markov chain respectively; SE function pair represents suppression-enhancement function pair.
Figure 1. Flowchart of the proposed method. BSSM: background seed screening mechanism; CMLM: cross-modal multi-graph learning model; bgAMC and fgAMC denote background-based and foreground-based saliency maps based on absorbing Markov chain respectively; SE function pair represents suppression-enhancement function pair.
Sensors 21 00838 g001
Figure 2. Traditional saliency detection models based on absorbing Markov chain. (a) RGB images. (b) Ground truth. (c) Saliency maps generated by [45]. (d) Saliency maps generated by [46]. (e) Saliency maps generated by [44].
Figure 2. Traditional saliency detection models based on absorbing Markov chain. (a) RGB images. (b) Ground truth. (c) Saliency maps generated by [45]. (d) Saliency maps generated by [46]. (e) Saliency maps generated by [44].
Sensors 21 00838 g002
Figure 3. The process diagram of the background seed screening. (a) The node sets based on saliency attribute; (b) The node sets based on location attribute; (c) The node sets based on depth attribute; (d) The node sets after removing the foreground nodes; (e) The diagram of the comprehensive relationship between the three attributes; (f) The final node sets after classifying.
Figure 3. The process diagram of the background seed screening. (a) The node sets based on saliency attribute; (b) The node sets based on location attribute; (c) The node sets based on depth attribute; (d) The node sets after removing the foreground nodes; (e) The diagram of the comprehensive relationship between the three attributes; (f) The final node sets after classifying.
Sensors 21 00838 g003
Figure 4. The construction and comparison of the proposed graph model. (a) Input RGB image. (b) Input depth image. (c) Ground truth. (d) A diagram of the connections of one of the nodes based on initial two-layer sparse graph. A node (illustrated by a pink dot) connects to its adjacent nodes (blue dots and connections) and the most similar node (dark green dots and connections) sharing common boundaries with its adjacent nodes. All edge nodes are connected to pairs (yellow dots and local connections). (e) A diagram of the connections of one of the nodes based on improved two-layer sparse graph. Different from the initial graph, the new edge nodes first remove some foreground nodes which are in the image boundary (the nodes at the bottom edge of image), and further join a small number of non-edge background nodes (black nodes). Each pair of the new edge nodes connects to each other (yellow and black dots and connections). Additionally, each node connects to the background seeds (light green dots and connections) and the foreground seeds (purple dots and connections).
Figure 4. The construction and comparison of the proposed graph model. (a) Input RGB image. (b) Input depth image. (c) Ground truth. (d) A diagram of the connections of one of the nodes based on initial two-layer sparse graph. A node (illustrated by a pink dot) connects to its adjacent nodes (blue dots and connections) and the most similar node (dark green dots and connections) sharing common boundaries with its adjacent nodes. All edge nodes are connected to pairs (yellow dots and local connections). (e) A diagram of the connections of one of the nodes based on improved two-layer sparse graph. Different from the initial graph, the new edge nodes first remove some foreground nodes which are in the image boundary (the nodes at the bottom edge of image), and further join a small number of non-edge background nodes (black nodes). Each pair of the new edge nodes connects to each other (yellow and black dots and connections). Additionally, each node connects to the background seeds (light green dots and connections) and the foreground seeds (purple dots and connections).
Sensors 21 00838 g004
Figure 5. Visual comparisons of different graph-based results. (a) Original RGB images. (b) Original depth images. (c) Saliency maps produced by two-graph neighborhood graph. (d) Saliency maps produced by two-layer sparse graph. (e) Saliency maps produced by improved two-layer sparse graph based on background seed screening mechanism (BSSM). (f) Saliency maps produced by improved two-layer sparse graph without the non-local connections. (g) Saliency maps produced by improved two-layer sparse graph with the non-local connections. (h) Ground truth.
Figure 5. Visual comparisons of different graph-based results. (a) Original RGB images. (b) Original depth images. (c) Saliency maps produced by two-graph neighborhood graph. (d) Saliency maps produced by two-layer sparse graph. (e) Saliency maps produced by improved two-layer sparse graph based on background seed screening mechanism (BSSM). (f) Saliency maps produced by improved two-layer sparse graph without the non-local connections. (g) Saliency maps produced by improved two-layer sparse graph with the non-local connections. (h) Ground truth.
Sensors 21 00838 g005
Figure 6. Visual comparisons of our proposed cross-mode multi-graph learning model (CMLM). (a) Original RGB images. (b) Original depth images. (c) Saliency maps based on single-mode multi-graph learning model (SMLM). (d) Saliency maps based on cross-modal multi-graph learning model (CMLM). (e) Ground truth.
Figure 6. Visual comparisons of our proposed cross-mode multi-graph learning model (CMLM). (a) Original RGB images. (b) Original depth images. (c) Saliency maps based on single-mode multi-graph learning model (SMLM). (d) Saliency maps based on cross-modal multi-graph learning model (CMLM). (e) Ground truth.
Sensors 21 00838 g006
Figure 7. Relevant schematic diagrams of the bilateral absorbing Markov chain model. (a) The construction of background-based absorbing Markov chain model: mirror edge superpixels (yellow dots) and a few non-edge background superpixels (black dots) as virtual absorbing nodes; (b) The construction of foreground-based absorbing Markov chain model: mirror foreground superpixels (blue dots) as virtual absorbing nodes. Both set all superpixels on the entire image as transient nodes.
Figure 7. Relevant schematic diagrams of the bilateral absorbing Markov chain model. (a) The construction of background-based absorbing Markov chain model: mirror edge superpixels (yellow dots) and a few non-edge background superpixels (black dots) as virtual absorbing nodes; (b) The construction of foreground-based absorbing Markov chain model: mirror foreground superpixels (blue dots) as virtual absorbing nodes. Both set all superpixels on the entire image as transient nodes.
Sensors 21 00838 g007
Figure 8. Visualization of the main saliency refinement process. (a) Original RGB images. (b) Original depth images. (c) Background prior probability maps combining color and depth cues. (d) Contrast prior probability maps combining color and depth cues. (e) Background-based saliency maps by absorbing Markov chain model. (f) Foreground-based saliency maps by absorbing Markov chain model. (g) Saliency maps optimized by cellular automata. (h) Saliency maps generate by the depth selective refinement mechanism based on suppression-enhancement function pair. (i) Ground truth.
Figure 8. Visualization of the main saliency refinement process. (a) Original RGB images. (b) Original depth images. (c) Background prior probability maps combining color and depth cues. (d) Contrast prior probability maps combining color and depth cues. (e) Background-based saliency maps by absorbing Markov chain model. (f) Foreground-based saliency maps by absorbing Markov chain model. (g) Saliency maps optimized by cellular automata. (h) Saliency maps generate by the depth selective refinement mechanism based on suppression-enhancement function pair. (i) Ground truth.
Sensors 21 00838 g008
Figure 9. Valuation of different components. (a) PR curves. (b) S-measure, E-measure, and F-measure at adaptive threshold, mean absolute error (MAE). G1: two-layer graph; G2: two-layer spare graph; BAMC (SL): bilateral absorbing Markov chain based on single-modal multi-graph learning (color mode), CL: cross-modal multi-graph learning model (color and depth modes); BSSM: background seed screening mechanism; DOM: depth-guided optimization model; NLC: non-local connection.
Figure 9. Valuation of different components. (a) PR curves. (b) S-measure, E-measure, and F-measure at adaptive threshold, mean absolute error (MAE). G1: two-layer graph; G2: two-layer spare graph; BAMC (SL): bilateral absorbing Markov chain based on single-modal multi-graph learning (color mode), CL: cross-modal multi-graph learning model (color and depth modes); BSSM: background seed screening mechanism; DOM: depth-guided optimization model; NLC: non-local connection.
Sensors 21 00838 g009
Figure 10. PR curves of the proposed and 10 state-of-the-art methods on 3 datasets. (a) NLPR dataset; (b) NJU2K dataset; (c) STERE dataset.
Figure 10. PR curves of the proposed and 10 state-of-the-art methods on 3 datasets. (a) NLPR dataset; (b) NJU2K dataset; (c) STERE dataset.
Sensors 21 00838 g010
Figure 11. Visual examples of different methods on different datasets. (a) NLPR dataset; (b) NJU2K dataset; (c) STERE dataset.
Figure 11. Visual examples of different methods on different datasets. (a) NLPR dataset; (b) NJU2K dataset; (c) STERE dataset.
Sensors 21 00838 g011aSensors 21 00838 g011b
Table 1. Quantitative comparisons of different RGB-D saliency detection methods on three popular datasets. Red, green and blue indicate the best, second and third performances. ↑ denotes larger is better, and ↓ denotes smaller is better.
Table 1. Quantitative comparisons of different RGB-D saliency detection methods on three popular datasets. Red, green and blue indicate the best, second and third performances. ↑ denotes larger is better, and ↓ denotes smaller is better.
MethodsYearNLPRNJU2KSTERE
S α E m F β MAE S α E m F β MAE S α E m F β MAE
ACSD20140.67280.74180.53450.17870.69920.78630.69640.20210.69190.79320.66070.2000
DESM20140.57220.69780.56330.31240.66480.68240.63210.28350.64250.67510.59420.2951
LHM20140.62980.81310.66360.10770.51360.70820.63830.20480.56170.77000.70290.1719
GP20150.65450.80450.65930.14610.52650.71610.65540.21060.58760.78420.71060.1822
DCMC20160.72440.78560.61410.11670.68610.79050.71730.17160.73060.83140.74250.1476
LBE20160.76190.85500.73550.08130.69520.79130.74000.15280.66010.74850.59510.2498
SE20160.75610.83880.69150.09130.66420.77220.73350.16870.70820.82500.74760.1427
CDCP20170.72700.80010.60760.11210.66850.74720.62380.18030.71340.79640.66550.1489
CDB20180.62860.80940.61320.11420.62390.74480.64840.20280.61510.80790.71270.1655
DTM20200.67870.76560.52710.16110.64900.74540.60820.22170.70490.79780.65850.1910
OURS20200.81310.87510.78450.07120.73610.79250.74940.13590.77740.83470.77240.1110
Table 2. Computational complexity comparison with traditional-based and deep learning-based RGB-D saliency detection methods. Red, green and blue indicate the best, second and third performances.
Table 2. Computational complexity comparison with traditional-based and deep learning-based RGB-D saliency detection methods. Red, green and blue indicate the best, second and third performances.
MethodsDCMCCDCPDTMD3NetBBS-NetUC-NetOURS
Year2016201720202020202020202020
PlatformMatlabMatlabMatlabPyTorchPyTorchPyTorchMatlab
Image size640 × 480640 × 480640 × 480224 × 224352 × 352352 × 352640 × 480
FLOPs(G)3.08911.25650.410455.072231.139616.15020.2002
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Wu, J.; Han, G.; Liu, P.; Yang, H.; Luo, H.; Li, Q. Saliency Detection with Bilateral Absorbing Markov Chain Guided by Depth Information. Sensors 2021, 21, 838. https://doi.org/10.3390/s21030838

AMA Style

Wu J, Han G, Liu P, Yang H, Luo H, Li Q. Saliency Detection with Bilateral Absorbing Markov Chain Guided by Depth Information. Sensors. 2021; 21(3):838. https://doi.org/10.3390/s21030838

Chicago/Turabian Style

Wu, Jiajia, Guangliang Han, Peixun Liu, Hang Yang, Huiyuan Luo, and Qingqing Li. 2021. "Saliency Detection with Bilateral Absorbing Markov Chain Guided by Depth Information" Sensors 21, no. 3: 838. https://doi.org/10.3390/s21030838

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop