Next Article in Journal
Solid Concentration Estimation by Kalman Filter
Previous Article in Journal
Virtual Network Embedding for Multi-Domain Heterogeneous Converged Optical Networks: Issues and Challenges
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Scale Global Contrast CNN for Salient Object Detection

1
College of Computer and Information Engineering, Tianjin Normal University, Tianjin 300387, China
2
Postdoctoral Innovation Practice Base, Huafa Industrial Share Co., Ltd., Zhuhai 519000, China
3
Key Wireless Laboratory of Jiangsu Province, School of Telecommunication and Information Engineering, Nanjing University of Posts and Telecommunications, Nanjing 210003, China
4
School of Computer Science and Engineering, Beihang University, Beijing 100191, China
5
Hangzhou Innovation Institute, Beihang University, Hangzhou 310051, China
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(9), 2656; https://doi.org/10.3390/s20092656
Submission received: 1 March 2020 / Revised: 24 April 2020 / Accepted: 29 April 2020 / Published: 6 May 2020
(This article belongs to the Section Intelligent Sensors)

Abstract

:
Salient object detection (SOD) is a fundamental task in computer vision, which attempts to mimic human visual systems that rapidly respond to visual stimuli and locate visually salient objects in various scenes. Perceptual studies have revealed that visual contrast is the most important factor in bottom-up visual attention process. Many of the proposed models predict saliency maps based on the computation of visual contrast between salient regions and backgrounds. In this paper, we design an end-to-end multi-scale global contrast convolutional neural network (CNN) that explicitly learns hierarchical contrast information among global and local features of an image to infer its salient object regions. In contrast to many previous CNN based saliency methods that apply super-pixel segmentation to obtain homogeneous regions and then extract their CNN features before producing saliency maps region-wise, our network is pre-processing free without any additional stages, yet it predicts accurate pixel-wise saliency maps. Extensive experiments demonstrate that the proposed network generates high quality saliency maps that are comparable or even superior to those of state-of-the-art salient object detection architectures.

1. Introduction

Salient object detection (SOD) is a fundamental task in computer vision that attempts to mimic human visual systems that rapidly respond to visual stimuli and locate visually salient objects in a scene. Estimating salient regions from an image could facilitate a lot of vision tasks, ranging from low-level ones such as segmentation [1] and image resizing [2] to high-level ones such as image captioning [3]; thus, it has been receiving increasing interest in the computer vision community and has been extended to other relevant topics, such as video SOD [4,5] and RGB-D SOD [6,7]. Numerous methods have been developed in the past decades. Most of them focus on two topics; the first one works on predicting eye fixations, and the other one aims at detecting salient object/regions from an image. In this work, we mainly focus on the latter one, i.e., detecting salient objects from clutter scenes.
Since the pioneer work of Itti’s computational saliency model [8], extensive efforts have been devoted to develop saliency methods identifying objects or locating regions that attract the attention of a human observer at the first sight of an image. Most of these methods draw inspiration from bottom-up human visual attention mechanisms, e.g., Feature Integration Theory (FIT) [9], and dedicate to measure uniqueness, distinctness and rarity of scenes to infer saliency maps, where the basic assumption is that objects might be salient if they differ from their surroundings. The central idea of these models is capturing contrast information of visual features either in a local manner [10] or in a global manner [11,12].
Since saliency computation is simulation of low-level stimuli-driven attention, early approaches mainly rely on low level features such as color, intensity, texture, etc. However, methods employing only low level features are very hard to produce satisfactory results. They suffer from the limitation that the detected regions may contain only parts of the salient objects, and the background is easily mixed up with the salient regions. One solution to overcome these limitations is extracting features from small regions (such as super-pixels) instead of pixels to incorporate context information [12]; another one is combining mid- or high-level cues [13] to introduce semantic knowledge of the salient objects. These methods could obtain promising results; however, they may fail to detect salient objects embedded in clutter backgrounds or sharing similar appearance with some distractors. Salient object detection remains a challenging problem especially under such complex scenes.
Recently, deep convolutional neural networks (CNNs) have achieved astonishing performance on a variety of computer vision tasks, such as object detection [14] and image segmentation [15]. The great success of CNNs has been motivating researchers to employ CNNs as feature extractors to represent salient objects [16,17]. In these methods, saliency can be estimated in pixel level or in region level. Inspired by FCN [15], some works model saliency computation as a dense labeling problem, they directly learn mapping functions from input images to their saliency maps [18]. These methods, which impel CNNs to learn internal features that capture saliency information of images in an implicit way, are pure computational methods, and they do not explain the psychophysical evidence. Given the fact that visual contrast is the most important factor in visual attention, capturing contrast information is the key idea in designing biologically-inspired saliency models. Therefore, many CNN based methods employ contrast strategy to build network architectures, in which a super-pixel segmentation step is usually added to obtain homogeneous regions, and contrast measurement between them is computed by using CNN features [16,17].
Recall that global (contrast) information is critical when detecting salient objects [12], and neurons in high level layers of a CNN have large receptive fields that could cover the entire image, thus capturing its global information, while neurons in low level layers only possess small receptive fields capturing local information. Given local and global features of an image, we can design a CNN architecture that learns their contrast information, thus estimating a saliency map of the image. Based on this idea, in this paper, we develop an end-to-end CNN architecture to automatically capture global contrast information between each local feature and the global image feature. Furthermore, we implement it in a multi-scale manner to leverage different level features; see Figure 1 for an example. We test the proposed method on four widely used saliency datasets and compare it with 10 state-of-the-art models. Experimental results demonstrate that the proposed method could effectively detect salient objects and achieve comparable or even superior performance to that of the state-of-the-art ones.
The rest of this paper is organized as follows. In Section 2, we give a brief review on contrast and CNN based saliency methods. Section 3 elaborates the detail architecture of the proposed multi-scale global contrast CNN. Section 4 gives experiments and comparisons with the state-of-the-art methods. Finally, this paper is concluded in Section 5.

2. Related Work

A great number of salient object detection methods have been proposed in the past decades; a comprehensive survey can be found from [19]. In this section we give a brief review of saliency computation models closely related to our method.

2.1. Contrast Based Models

Recent studies [20] have suggested that visual contrast is at the central of saliency attention. Most existing visual saliency computation models are designed based on either local or global contrast cues.
Local contrast based methods investigate the rarity of image regions with respect to their local neighborhoods [12]. The pioneer work on these models is Itti’s model [8], in which saliency maps are generated by measuring center-surround difference of color, orientation, and intensity features. Later, Harel et al. [21] estimates center-surround saliency maps in a graph computation manner and achieves superior performance to that of Itti’s model. Similarly, Klein et al. [22] encodes local center-surround divergence in multi-feature channels and computes them in an efficient scale-space to deal with scale variations. Liu et al. [23] incorporates multi-scale contrast features with center-surround histogram and color spatial distribution by Markov random fields to detect salient objects. Without knowing the size of the salient object, contrast is usually computed at multiple scales. Jiang et al. [24] integrates regional contrast, regional property and regional background descriptors to form saliency maps. One major drawback of local contrast based methods is that they tend to highlight strong edges of salient objects thus producing salient regions with holes.
Global contrast based methods compute saliency of a small region by measuring its contrast with respect to all other parts of the image. Achanta et al. [11] proposes a simple frequency-tuned salient region detection method, in which saliency value of a pixel is defined as difference between its color and mean color of the image. Cheng et al. [12] introduces a global contrast based salient object detection algorithm, in which saliency of a region is assigned by the histogram difference between the target region and all other regions. Later, they propose a soft image abstraction method to capture large scale perceptually homogeneous elements, which enables more effective estimation of global saliency cues [25]. Differently, in [10], contrast and saliency estimation is formulated in a unified way using high-dimensional Gaussian filter.

2.2. Cnn Based Models

Representing pixels or regions efficiently and compactly is critical for saliency models. The aforementioned methods only employ low-level features such as color and texture. Recently, inspired by the great success of CNNs in many computer vision tasks, researchers in the community are encouraged to leverage power of CNNs to capture high level information from the images. Vig et al. [26] is probably the first attempt at modeling saliency computation using deep neural networks. This work focuses on predicting eye fixation by assembling different layers using a linear SVM. Zhao et al. [16] and Li et al. [27] extract a global feature of an image and a local feature of a small region in it using different CNNs, and then, the saliency of this region is formulated as a classification problem. Wang et al. [28] proposes a saliency detection model composed of two CNNs; one learns features capturing local cues such as local contrast, textures and shape information, and the other one learns the complex dependencies among global cues. He et al. [29] learns the hierarchical contrast features using multi-stream CNNs. To obtain accurate salient boundaries, images are first segmented into super-pixels in multi-scales. Two sequences, color uniqueness and color distribution, are extracted from each super-pixel and fed into CNNs to obtain features. Saliency maps are generated by fusing saliency results inferred from each scale. Li et al. [17] adopts a two-stream deep contrast CNN architecture. One stream accepts original images as input, infers semantic properties of salient objects and captures visual contrast among multi-scale feature maps to output their coarse saliency maps. The other stream extracts segment wise features and models visual contrast between regions and saliency discontinuities along region boundaries. Ren et al. [30] puts forward a multi-scale encoder-decoder network (MSED) by fusing multi-scale features from the image-level. Li et al. [31] presents a multi-scale cascade network (MSC-Net) for saliency detection in a coarse-to-fine manner, which encodes abundant contextual information whilst progressively incorporating the saliency prior knowledge to improve the detection accuracy. Li et al. [32] discloses the importance of inference module in the saliency detection and presents a deep yet lightweight architecture which extracts multi-scale features by leveraging a multi-dilated depth-wise convolution operation. Different from them, in this paper, we design an end-to-end multi-scale global contrast network that explicitly learns hierarchical contrast information among global and local features of an image to infer its salient object regions. Compared with the aforementioned multi-scale CNN-based models, our proposed model is lightweight and without any pre-processing operations.

3. Multi-Scale Global Contrast CNN

In this section, we will give details of our multi-scale global contrast CNN (denoted as MGCC) architecture.

3.1. Formulation

Salient object detection can be considered as a binary classification problem. Given an image I, the saliency value of a pixel i (i also could be a super-pixel) in it can be represented as follows,
S i I = Pr ( y i = 1 | f i , f I ; W )
where S I is the saliency map of the image I (for notational simplicity, we will drop the superscript I in the remainder of this paper), S i is the saliency value of pixel i, f i and f I are features of the pixel i and image I, respectively. y i = 1 indicates the probability of the pixel i being salient, while y i = 0 indicates background, W is collection of parameters.
In global contrast based methods, S i can be estimated through measuring the distance of the two features,
S i = C d ( f i , f I ) ; W
where C [ · ] is a function estimating saliency maps from d ( · ) , and d ( · ) is a metric function measuring the distance of f i and f I , which could be a simple Euclidean distance or other pre-defined distance metrics. For example, in [12], features are represented using color histograms, the saliency of a super-pixel is defined as its color contrast to all other regions in the image, which can be inferred from the weighted summation of color distances between the current region and all other ones.
Since S i is a probability value ranging from 0 to 1, C [ · ] often adopts the following form,
S i = C d ( f i , f I ) ; W σ ( d ( f i , f I ) ; W )
where σ ( · ) is a nonlinear function, e.g., sigmoid function, mapping d ( · ) to [ 0 , 1 ] . If we represent f i and f I using deep features and define d ( · ) as a metric learned from training data, then Equation (3) can be solved using a convolutional neural network. In the following section, we will give details of the proposed network architecture to achieve this.

3.2. Global Contrast Learning

The essence of obtaining contrast information between two regions is quantifying a “distance" between features of them, thus inducing a measure of similarity. As discussed above, the function d ( · ) can be viewed as a metric function that captures distance between f i and f I , in which larger distance indicates higher contrast, thus higher probability being salient. There are multiple ways to calculate d ( · ) . For instance, it can be formulated as pre-defined metrics, such as L 1 or L 2 norms. However, this requires the two features to have the same dimension, which is hard to achieve in CNNs. Suppose f i l is a feature of pixel i extracted from the l-th convolutional layer of a CNN (e.g., VGG-16 [33]). Although we can apply global pooling on this layer to obtain f I , thus making these two features have the same dimension, i.e., the channels of feature maps in this layer, lots of information will be lost during pooling process, especially when l is in low layers. Furthermore, low level features lack of semantic information, which is very important in detecting salient objects [34]. An alternative solution is adding an additional layer to project both of them into an embedding space, making them to have equal size and then calculating a distance matrix. However, it is hard to achieve satisfactory results by inferring salient objects directly from distance matrices; this is mainly because important semantic information about the salient objects is missing when computing distances.
In addition to the pre-defined metrics, another solution is defining metric with the knowledge of the data, that is, learning the metric functions from the training samples. As a powerful tool as it is, CNNs have been proved to be very effective in approximating very complex functions and in learning visual similarities. To achieve this end, we attempt to design a CNN architecture that learns the distance function between f i and f I . One important thing that should be noted is that the semantic information of the object should be preserved because we intend to recover accurate object boundaries. To achieve this, we design a very simple architecture that could capture global contrast of f i and f I . Firstly, VGG-16 [33] is employed to extract features from input images. VGG-16 consists of 13 convolutional layers, 5 pooling layers and 3 fully connected layers. We modify it by removing the last 3 fully connected layers and using 256 × 256 input instead of original 224 × 224 . The last pool layer of the modified VGG-16 (with size of 8 × 8 × 512 ) is used to represent the global feature. To emphasize contrast information and reduce distractions from semantic information, we apply an additional 1 × 1 convolutional layer to obtain compact 8 × 8 × 256 representations of global features. Then, we concatenate it with previous layers in a recurrent manner, and introduce more convolutional layers to learn visual contrast information, as shown in Figure 2. At the end of the network, output is up-sampled to meet the size of the ground truth maps. Although it is simple, this repeating concatenation strategy can successfully characterize contrast information of the image while preserving semantic information of salient objects.

3.3. Multi-Scale Global Contrast Network

Layers in a CNN from low to high levels capture different levels of abstraction. Neurons in early layers have small receptive fields that only respond to local regions of an input, thus producing low level features representing texture, edges, etc., while neurons in late layers have large receptive fields that may cover most of or even the entire image, thus capturing semantic information of the image or objects in it.
It is very important to employ low level features when generating output with accurate and clear boundaries [15]. Inspired by HED [35], we design multi-scale outputs to capture features in different layers and integrate them together to produce finer results. Specifically, we propose a Multi-scale Global Contrast CNN, abbreviated as MGCC, which adopts truncated VGG16 as the backbone, there are five convolutional segments, each of which contain two or three convolution layers, followed by one pooling layer to down-sample the size of the feature maps. Our proposed model takes the final output feature map, i.e., the fifth convolution segment, as the global feature. Then, we concatenate it with previous layers in a recurrent channel-concatenation manner by first resizing the global feature map to the same size with the corresponding feature maps at previous layers (the global contrast module, which corresponds to the left-part in Figure 2). This process is somewhat similar to feature pyramid network (FPN) [36] but different from it in that we respectively take the outputs of the previous four layers to concatenate with the fifth convolution layer, i.e., global features. For example, the output feature map of the fourth segment has the size of 16 × 16 × 256 ; thus, we resize the global feature whose size is 8 × 8 × 512 by upsampling two times, to the size of 16 × 16 × 512 . Then, we concatenate them in a channel-wise manner. To learn more visual contrast information, we introduce several more convolutional layers (referred to the right-part of Figure 2). Consequently, the proposed MGCC generates four scale outputs, each of which could produce accurate saliency maps. We resize all the saliency maps of the four scale outputs to the same size of the original image and then fuse them in an element-wise summation to obtain the final finer saliency map. Figure 1 shows several examples. The architecture of the proposed MGCC is shown in Figure 3. The detail parameters are given in Table 1.
As discussed above, the salient object detection task can be formulated as a binary prediction problem; thus we use binary cross entropy as loss function to train our network. Given a set of training samples ( X n , Y n ) n = 1 N , where N is the number of samples, X n is an image, and Y n is the corresponding ground truth, the loss function L m for the m-th scale output is defined as
L m ( W m , W vgg ) = j = 1 | Y | Y j log Pr ( Y ^ j m = 1 | X ; W m , W vgg ) + ( 1 Y j ) log Pr ( Y ^ j m = 0 | X ; W m , W vgg )
where Y ^ j m is the predicted saliency value for pixel j. The fused loss L fused takes a similar form to Equation (4), and the fusion weights w are also learned form training samples. Finally, the loss function for training is given by
L ( W ) = α L fused ( W ) + m = 1 4 β m L m ( W m , W vgg )
where W = { W vgg , W 1 , , W 4 , w } is the collection of the parameters in the proposed network. w denotes the trainable parameters in the additional convolution layer for scale-4, which has been described in Table 1. α and β s are weights balancing different loss functions and all set to 1 in our experiments.

4. Experiments

4.1. Datasets

All experiments are conducted on four widely used datasets, including ECSSD [37], HKU-IS [27], PASCAL-S [38] and DUT-OMRON [39].
  • ECSSD [37] is a challenge dataset which contains 1000 images with semantically meaningful but structurally complex natural contents.
  • HKU-IS [27] is composed by 4447 complex images, each of which contains many disconnected objects with diverse spatial distribution. Furthermore, it is very challenging for the similar foreground/background appearance.
  • PASCAL-S [38] contains a total of 850 images, with eye-fixation records, roughly pixel-wise and non-binary salient object annotations included.
  • DUT-OMRON [39] consists of 5168 images with diverse variations and complex background, each of which has pix-level ground truth annotations.

4.2. Evaluation Metrics

Three metrics, including precision-recall (P-R) curves, F-measure and Mean Absolute Error (MAE) are used to evaluate the performance of the proposed and other methods. For an estimated saliency map with values ranging from 0 to 1, its precision and recall can be obtained by comparing the thresholded binary mask with the ground truth. Making these comparisons at each threshold and averaging them on all images will generate P-R curves of this dataset.
The F-measure is a harmonic mean of average precision and recall, which is defined as,
F β = ( 1 + β 2 ) · precision · recall β 2 · prcision + recall
As suggested by many existing works [16,40], β 2 is set as 0.3. MAE reflects absolute difference of the estimated S and the ground truth saliency maps G.
MAE = 1 W × H x = 1 W y = 1 H | S ( x , y ) G ( x , y ) |
where W and H are width and height of the maps.
Both metrics of MAE and F-measure are based on pixel-wise errors and often ignore the structural similarities, as demonstrated in [41,42]. In many applications, it is desired that the results of the salient object detection model retain the structure of objects. Therefore, three more metrics, i.e., weighted F-measure F β w [41], S-measure ( S α ) [42] and E-measure ( E m ) [43] are also introduced to further evaluate our proposed method.
Specifically, F β w [41] is computed as follows:
F β w = ( 1 + β 2 ) · precisio n w · recal l w β 2 · prcisio n w + recal l w
where precisio n w and recal l w are the weighted precision and recall. Note that the difference between F β w and F β is that it can compare a non-binary map against ground-truth with thresholding operation, to avoid the interpolation flaw. As suggested in [41,44,45,46], we empirically set β 2 = 0 . 3 .
S α [42] is proposed to measure the spatial structure similarities between saliency maps.
S α = α × S o + ( 1 α ) × S r
where α is a balance parameter between object-aware structural similarity S o and region-aware structural similarity S r , as suggested in [42,47,48].
E-measure ( E m ) [43,44,49,50] is to evaluate the foreground map (FM) and noise, which can correctly rank the maps consistent with the application rank.
E m = 1 w × h x = 1 w y = 1 h ϕ F M ( x , y )
where ϕ denotes the enhanced alignment matrix, which is to capture pixel-level matching and image-level statistics of a binary map.

4.3. Implementation Details

We implement the proposed network in PyTorch [51]. As mentioned above, we utilize VGG-16 [33] pre-trained on ImageNet [52] as backbone to extract features. The MSRA10K dataset [12] is employed to train the network. Before feeding into the network, all images are resized to 256 × 256 . During training, parameters are optimized using Adam optimizer. The learning rates for VGG-16 and other newly added layers are initially set as 10 4 and 10 3 and decreased by a factor of 0.1 in every 30 epochs. In addition, we set momentum as 0.9. The training was conducted on a single NVIDIA Titan X GPU with a batch size of 8. It will converge in 80 epochs. It should be noted that no data augmentation was used during training.

4.4. Comparison with the Sate-of-the-Art

We compare the proposed MGCC with 10 state-of-the-art saliency models, including 5 CNN based methods: LEGS [28], MDF [27], MCDL [16], ELD [53], DCL [17] and 5 classical models: SMD [40], DRFI [24], RBD [54], MST [55] and MB+ [56]. These methods are chosen because the first 5 are also CNN and contrast based methods, and the last five traditional methods are either reported as benchmarking methods in [19] or developed recently. For fair comparison, we employ either implementation or saliency maps provided by the authors.
We report P-R curves in Figure 4 and list Max F-measure (Max F β ), MAE, F β w , S α and E m in Table 2. From Figure 4 we can see that our method achieves better P-R curves on the four datasets; especially on ECSSD and HKU-IS datasets, it obtains the best results, showing that MGCC can achieve the highest precision and the highest recall comparing with other methods. On PASCAL-S and DUT-OMRON datasets, although MGCC drops faster than DCL [17] and ELD [53] on the right side of the P-R curves, we can observe that the MGCC obtains better or at least comparable break-even points (i.e., the points on the curves where precision equals recall), which indicates that our method can keep a good balance between precision and recall.
From Table 2, we can see that deep learning based approaches significantly outperform traditional saliency models, which clearly demonstrate the superiority of deep learning techniques. Among all the methods, the proposed MGCC achieves almost the best results over all the four datasets, except for the HKU-IS dataset, on which, DCL, a leading contrast based saliency model, performs slightly better than ours in terms of Max F β and F β w ; however, it underperforms ours in terms of MAE, S α , and E m . The proposed MGCC and DCL [17] obtain identical Max F β on the PASCAL-S dataset, yet lower MAE is achieved by our MGCC. It can be seen that MGCC improves MAE with a considerable margin on all four datasets. This demonstrates that our method can produce more accurate salient regions than other methods.
Some example results of our and other methods are shown in Figure 5 for visual comparison, from which we can see our method performs well even under complex scenes. It is worth mentioning that to achieve better performance and obtain accurate salient regions, many CNN based models adopt two- or multi-stream architectures to incorporate both pixel-level and segment-level saliency information [16,17,27,28,53]. For instance, DCL consists of two complementary components; one stream generates low resolution pixel-level saliency maps, and one stream generates full resolution segment-level saliency maps. They combine the two saliency maps to obtain better results. While our network only has one stream and predicts saliency maps in pixel wise, with simpler architecture and without additional processing (e.g., super-pixel segmentation or CRF), our method achieves comparable or even better results than other deep saliency models. Another thing that should be mentioned is that, with simple architecture and completely end-to-end feed-forward inference, our network produces saliency maps at a near real time speed of 19 fps on a Titan X GPU.

4.5. Ablation Study

To further demonstrate the effectiveness of the multi-scale fusion strategy, we compare our proposed model with the results output from scale-1, scale-2, scale-3, and scale-4, as illustrated in Table 2.
From Table 2, we can observe that when merging global feature with the features of previous layers, the performance gradually increases from scale-1 to scale-4, which verifies that merging higher-level semantic features can further boost the performance. Additionally, from the metrics, we can see that fusing multi-scale information (i.e., our proposed MGCC model), the performance has significantly improved, which indeed demonstrates the effectiveness and superiority of our proposed multi-scale fusion strategy.

5. Conclusions and Future Work

In this paper, we have proposed an end-to-end multi-scale global contrast CNN for salient object detection. In contrast to previous CNN based methods, designing complex two- or multi-stream architectures to capture visual contrast information or directly mapping images to their saliency maps and learning internal contrast information in an implicit way, our network is simple yet good at capturing global visual contrast, thus achieving superior performance both at detecting salient regions and processing speed.
As demonstrated in existing literature [57], the SOC dataset [58] is the most challenging dataset. Some attempts have been made on this dataset in Deepside [44] and SCRNet [59]. We look forward to conducting some experiments on this dataset in our future work to further demonstrate the effectiveness and superiority of our proposed approach.

Author Contributions

Conceptualization, W.F., G.G., X.L. and Q.L.; methodology, W.F., X.C. and Q.L; software, W.F., G.G. and X.C.; validation, W.F., G.G., X.C.; formal analysis, W.F., X.L. and Q.L.; investigation, W.F., G.G., X.C.; resources, W.F., G.G., X.C. and Q.L.; data curation, W.F., X.L. and Q.L.; writing—original draft preparation, W.F., X.L. and X.C.; writing—review and editing, W.F., G.G, X.C., X.L. and Q.L.; visualization, W.F. and X.L.; supervision, W.F. and Q.L; project administration, W.F.; funding acquisition, W.F. All authors have read and agreed to the published version of the manuscript.

Funding

This study is funded by NSFC (Natural Science Foundation of China): 61602345; National Key Research and Development Plan: 2019YFB2101900; Application Foundation and Advanced Technology Research Project of Tianjin (15JCQNJC01400).

Acknowledgments

We are grateful to the reviewers for their valuable comments and suggestions that help us to improve this work. We also thank authors who kindly provide their codes for comparison.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Liu, Z.; Shi, R.; Shen, L.; Xue, Y.; Ngan, K.; Zhang, Z. Unsupervised salient object segmentation based on kernel density estimation and two-phase graph cut. In Proceedings of the TMM 2012, Liberec, Czech Republic, 4–6 September 2012; Volume 14, pp. 1275–1289. [Google Scholar]
  2. Achanta, R.; Süsstrunk, S. Saliency detection for content-aware image resizing. In Proceedings of the 2009 IEEE International Conference on Image Processing, Cairo, Egypt, 7–10 November 2009; pp. 1005–1008. [Google Scholar]
  3. Andrej, K.; Li, F. Deep visual-semantic alignments for generating image descriptions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2015, Boston, MA, USA, 7–12 June 2015; pp. 3128–3137. [Google Scholar]
  4. Fan, D.; Wang, W.; Cheng, M.; Shen, J. Shifting More Attention to Video Salient Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2019, Long Beach, CA, USA, 16–20 June 2019; pp. 8554–8564. [Google Scholar]
  5. Wang, W.; Shen, J.; Shao, L. Video Salient Object Detection via Fully Convolutional Networks. IEEE Trans. Image Process. 2018, 27, 38–49. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Zhao, J.; Cao, Y.; Fan, D.; Cheng, M.; Li, X.; Zhang, L. Contrast Prior and Fluid Pyramid Integration for RGBD Salient Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2019, Long Beach, CA, USA, 16–21 June 2019; pp. 3927–3936. [Google Scholar]
  7. Liu, Z.; Zhang, W.; Zhao, P. A cross-modal adaptive gated fusion generative adversarial network for RGB-D salient object detection. Neurocomputing 2020, 387, 210–220. [Google Scholar] [CrossRef]
  8. Itti, L.; Koch, C.; Niebur, E. A model of saliency-based visual attention for rapid scene analysis. IEEE Trans. Pattern Anal. Mach. Intell. 1998, 11, 1254–1259. [Google Scholar] [CrossRef] [Green Version]
  9. Treisman, A.; Gelade, G. A feature-integration theory of attention. Cogn. Psychol. 1980, 12, 97–136. [Google Scholar] [CrossRef]
  10. Perazzi, F.; Krähenbühl, P.; Pritch, Y.; Hornung, A. Saliency filters: Contrast based filtering for salient region detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2012, Providence RI, USA, 16–21 June 2012; pp. 733–740. [Google Scholar]
  11. Achanta, R.; Hemami, S.; Estrada, F.; Süsstrunk, S. Frequency-tuned salient region detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2009, Miami, FL, USA, 20–25 June 2009; pp. 1597–1604. [Google Scholar]
  12. Cheng, M.; Mitra, N.; Huang, X.; Torr, P.; Hu, S. Global contrast based salient region detection. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 569–582. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  13. Sun, X.; Huang, Z.; Yin, H.; Shen, H. An integrated model for effective saliency prediction. In Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence (AAAI 2017), San Francisco, CA, USA, 4–9 February 2017; pp. 274–281. [Google Scholar]
  14. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2014, Columbus, OH, USA, 24–27 June 2014; pp. 580–587. [Google Scholar]
  15. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2015, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  16. Zhao, R.; Ouyang, W.; Li, H.; Wang, X. Saliency detection by multi-context deep learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2015, Boston, MA, USA, 7–12 June 2015; pp. 1265–1274. [Google Scholar]
  17. Li, G.; Yu, Y. Deep contrast learning for salient object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2016, Las Vegas, NV, USA, 27–30 June 2016; pp. 478–487. [Google Scholar]
  18. Zhang, P.; Wang, D.; Lu, H.; Wang, H.; Yin, B. Learning uncertain convolutional features for accurate saliency detection. In Proceedings of the 2017 IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 212–221. [Google Scholar]
  19. Borji, A.; Cheng, M.; Jiang, H.; Li, J. Salient object detection: A benchmark. IEEE Trans. Image Process. 2015, 24, 5706–5722. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  20. Einhäuser, W.; König, P. Does luminance-contrast contribute to a saliency map for overt visual attention? Eur. J. Neurosci. 2003, 17, 1089–1097. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  21. Harel, J.; Koch, C.; Perona, P. Graph-based visual saliency. In Proceedings of the Advances in Neural Information Processing Systems 20 (NIPS 2007), Vancouver, BC, Canada, 3–6 December 2007; pp. 545–552. [Google Scholar]
  22. Klein, D.; Frintrop, S. Center-surround divergence of feature statistics for salient object detection. In Proceedings of the 2011 IEEE International Conference on Computer Vision, Barcelona, Spain, 6–13 November 2011; pp. 2214–2219. [Google Scholar]
  23. Liu, T.; Yuan, Z.; Sun, J.; Wang, J.; Zheng, N.; Tang, X.; Shum, H. Learning to detect a salient object. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 33, 353–367. [Google Scholar]
  24. Jiang, H.; Wang, J.; Yuan, Z.; Wu, Y.; Zheng, N.; Li, S. Salient object detection: A discriminative regional feature integration approach. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2013, Portland, OR, USA, 23–28 June 2013; pp. 2083–2090. [Google Scholar]
  25. Cheng, M.; Warrell, J.; Lin, W.; Zheng, S.; Vineet, V.; Crook, N. Efficient salient region detection with soft image abstractionn. In Proceedings of the 2013 IEEE International Conference on Computer Vision, Sydney, Australia, 1–8 December 2013; pp. 1529–1536. [Google Scholar]
  26. Vig, E.; Dorr, M.; Cox, D. Large-scale optimization of hierarchical features for saliency prediction in natural images. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2014, Columbus, OH, USA, 24–27 June 2014; pp. 2798–2805. [Google Scholar]
  27. Li, G.; Yu, Y. Visual saliency detection based on multiscale deep CNN features. IEEE Trans. Image Process. 2016, 25, 5012–5024. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  28. Wang, L.; Lu, H.; Ruan, X.; Yang, M. Deep networks for saliency detection via local estimation and global search. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2015, Boston, MA, USA, 7–12 June 2015; pp. 3183–3192. [Google Scholar]
  29. He, S.; Lau, R.W.; Liu, W.; Huang, Z.; Yang, Q. SuperCNN: A Superpixelwise Convolutional Neural Network for Salient Object Detection. Int. J. Comput. Vis. 2015, 115, 330–344. [Google Scholar] [CrossRef]
  30. Ren, Q.; Hu, R. Multi-scale deep encoder-decoder network for salient object detection. Neurocomputing 2018, 316, 95–104. [Google Scholar] [CrossRef]
  31. Li, X.; Yang, F.; Cheng, H.; Chen, J.; Guo, Y.; Chen, L. Multi-Scale Cascade Network for Salient Object Detection. In Proceedings of the 2017 ACM International Conference on Multimedia, Silicon Valley, CA, USA, 23–27 October 2017; pp. 439–447. [Google Scholar]
  32. Li, Z.; Lang, C.; Chen, Y.; Liew, J.; Feng, J. Deep Reasoning with Multi-scale Context for Salient Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2019, Long Beach, CA, USA, 16–20 June 2019. [Google Scholar]
  33. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  34. Nguyen, T.; Liu, L. Salient object detection with semantic priors. arXiv 2017, arXiv:1705.08207. [Google Scholar]
  35. Xie, S.; Tu, Z. Holistically-nested edge detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2015, Boston, MA, USA, 7–12 June 2015; pp. 1395–1403. [Google Scholar]
  36. Lin, T.; Dollar, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature Pyramid Networks for Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2017, Honolulu, HI, USA, 21–26 July 2017; pp. 936–944. [Google Scholar]
  37. Yan, Q.; Xu, L.; Shi, J.; Jia, J. Hierarchical saliency detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2013, Portland, OR, USA, 23–28 June 2013; pp. 1155–1162. [Google Scholar]
  38. Li, Y.; Hou, X.; Koch, C.; Rehg, J.M.; Yuille, A.L. The secrets of salient object segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2014, Columbus, OH, USA, 24–27 June 2014; pp. 280–287. [Google Scholar]
  39. Yang, C.; Zhang, L.; Lu, H.; Ruan, X.; Yang, M. Saliency detection via graph-based manifold ranking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2013, Portland, OR, USA, 23–28 June 2013; pp. 3166–3173. [Google Scholar]
  40. Peng, H.; Li, B.; Ling, H.; Hu, W.; Xiong, W.; Maybank, S.J. Salient object detection via structured matrix decomposition. TPAMI 2017, 39, 818–832. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  41. Margolin, R.; Zelnikmanor, L.; Tal, A. How to Evaluate Foreground Maps. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2014, Columbus, OH, USA, 24–27 June 2014; pp. 248–255. [Google Scholar]
  42. Fan, D.; Cheng, M.; Liu, Y.; Li, T.; Borji, A. Structure-Measure: A New Way to Evaluate Foreground Maps. In Proceedings of the 2017 IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 4558–4567. [Google Scholar]
  43. Fan, D.; Gong, C.; Cao, Y.; Ren, B.; Cheng, M.; Borji, A. Enhanced-alignment Measure for Binary Foreground Map Evaluation. In Proceedings of the IJCAI 2018, Stockholm, Sweden, 13–19 July 2018; pp. 698–704. [Google Scholar]
  44. Fu, K.; Zhao, Q.; Gu, I.Y.; Yang, J. Deepside: A general deep framework for salient object detection. Neurocomputing 2019, 356, 69–82. [Google Scholar] [CrossRef]
  45. Su, J.; Li, J.; Zhang, Y.; Xia, C.; Tian, Y. Selectivity or Invariance: Boundary-aware Salient Object Detection. In Proceedings of the 2019 IEEE International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019. [Google Scholar]
  46. Zhao, T.; Wu, X. Pyramid Feature Attention Network for Saliency Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2019, Long Beach, CA, USA, 16–20 June 2019; pp. 3085–3094. [Google Scholar]
  47. Zhang, P.; Lu, H.; Shen, C. Troy: Give Attention to Saliency and for Saliency. arXiv 2018, arXiv:1808.02373. [Google Scholar]
  48. Zhang, P.; Lu, H.; Shen, C. HyperFusion-Net: Densely Reflective Fusion for Salient Object Detection. arXiv 2018, arXiv:1804.05142. [Google Scholar]
  49. Piao, Y.; Ji, W.; Li, J.; Zhang, M.; Lu, H. Depth-Induced Multi-Scale Recurrent Attention Network for Saliency Detection. In Proceedings of the 2019 IEEE International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 7254–7263. [Google Scholar]
  50. Wei, J.; Wang, S.; Huang, Q. F3Net: Fusion, Feedback and Focus for Salient Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2019, Long Beach, CA, USA, 16–20 June 2019. [Google Scholar]
  51. Paszke, A.; Gross, S.; Chintala, S.; Chanan, G. PyTorch: Tensors and Dynamic Neural Networks in Python with Strong GPU Acceleration. 2017. Available online: https://github.com/pytorch/pytorch (accessed on 6 May 2020).
  52. Deng, J.; Dong, W.; Socher, R.; Li, L.J.; Li, K.; Li, F. Imagenet: A large-scale hierarchical image database. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2019, Long Beach, CA, USA, 16–20 June 2019; pp. 248–255. [Google Scholar]
  53. Lee, G.; Tai, Y.; Kim, J. Deep saliency with encoded low level distance map and high level features. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2016, Las Vegas, NV, USA, 27–30 June 2016; pp. 660–668. [Google Scholar]
  54. Zhu, W.; Liang, S.; Wei, Y.; Sun, J. Saliency optimization from robust background detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2014, Columbus, OH, USA, 24–27 June 2014; pp. 2814–2821. [Google Scholar]
  55. Tu, W.; He, S.; Yang, Q.; Chien, S. Real-time salient object detection with a minimum spanning tree. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2016, Las Vegas, NV, USA, 27–30 June 2016; pp. 2334–2342. [Google Scholar]
  56. Zhang, J.; Sclaroff, S.; Lin, Z.; Shen, X.; Price, B.; Mech, R. Minimum barrier salient object detection at 80 fps. In Proceedings of the 2015 IEEE International Conference on Computer Vision, Santiago, Chile, 13–16 December 2015; pp. 1404–1412. [Google Scholar]
  57. Wang, W.; Lai, Q.; Fu, H.; Shen, J.; Ling, H. Salient Object Detection in the Deep Learning Era: An In-Depth Survey. arXiv 2019, arXiv:1904.09146. [Google Scholar]
  58. Fan, D.; Cheng, M.; Liu, J.; Gao, S.; Hou, Q.; Borji, A. Salient Objects in Clutter: Bringing Salient Object Detection to the Foreground. In Proceedings of the 15th European Conference on Computer Vision (ECCV 2018), Munich, Germany, 8–14 September 2018; pp. 196–212. [Google Scholar]
  59. Wu, Z.; Su, L.; Huang, Q. Stacked Cross Refinement Network for Edge-Aware Salient Object Detection. In Proceedings of the 2019 IEEE International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 7264–7273. [Google Scholar]
Figure 1. Saliency maps generated by the proposed multi-scale global contrast convolutional neural network (CNN). Our method predicts saliency maps in multiple scales and fuses them to obtain the final results.
Figure 1. Saliency maps generated by the proposed multi-scale global contrast convolutional neural network (CNN). Our method predicts saliency maps in multiple scales and fuses them to obtain the final results.
Sensors 20 02656 g001
Figure 2. A sketch of learning global contrast module used in this paper. W i ( l ) , H i ( l ) and C i ( l ) respectively represent the width, height and channel of the feature maps at the previous layers, and C ( g ) means the channel of the global feature map, resizing the global feature map to the same size as the feature maps at previous layers and concatenating them in a channel-wise manner.
Figure 2. A sketch of learning global contrast module used in this paper. W i ( l ) , H i ( l ) and C i ( l ) respectively represent the width, height and channel of the feature maps at the previous layers, and C ( g ) means the channel of the global feature map, resizing the global feature map to the same size as the feature maps at previous layers and concatenating them in a channel-wise manner.
Sensors 20 02656 g002
Figure 3. The architecture of the proposed multi-scale global contrast CNN (MGCC).
Figure 3. The architecture of the proposed multi-scale global contrast CNN (MGCC).
Sensors 20 02656 g003
Figure 4. P-R curves of the proposed and 10 state-of-the-art methods on ECSSD, HKU-IS, PASCAL-S and DUT-OMRON datasets.
Figure 4. P-R curves of the proposed and 10 state-of-the-art methods on ECSSD, HKU-IS, PASCAL-S and DUT-OMRON datasets.
Sensors 20 02656 g004
Figure 5. Visual comparison with the state-of-the-art methods.
Figure 5. Visual comparison with the state-of-the-art methods.
Sensors 20 02656 g005
Table 1. Detail architectures of the proposed network. ( m , n ) / ( k × k ) means that there are m channels in previous layer and n channels in current layer; the filters connecting them have size k × k . Scale-4 architecture is slightly different to the other three ones in that it has one additional convolutional layer.
Table 1. Detail architectures of the proposed network. ( m , n ) / ( k × k ) means that there are m channels in previous layer and n channels in current layer; the filters connecting them have size k × k . Scale-4 architecture is slightly different to the other three ones in that it has one additional convolutional layer.
Layer 1Layer 2Layer 3Layer 4Layer 5
Scale-1(320, 320)/(3×3)(320, 256)/(1×1)(256, 256)/(3×3)(256, 1)/(1×1)-
Scale-2(384, 384)/(3×3)(384, 256)/(1×1)(256, 256)/(3×3)(256, 1)/(1×1)-
Scale-3(512, 512)/(3×3)(512, 256)/(1×1)(256, 256)/(3×3)(256, 1)/(1×1)-
Scale-4(768, 768)/(3×3)(768, 512)/(3×3)(512, 256)/(1×1)(256, 256)/(3×3)(256, 1)/(1×1)
Table 2. Performance of the proposed MGCC and other 10 state-of-the-art methods on 4 popular datasets. Red, green and blue indicate the best, the second best and the third best performances. “–” represents no reported.
Table 2. Performance of the proposed MGCC and other 10 state-of-the-art methods on 4 popular datasets. Red, green and blue indicate the best, the second best and the third best performances. “–” represents no reported.
DatasetsECSSD [37]HKU-IS [27]PASCAL-S [38]DUT-OMRON [39]
MethodsMax F β MAE↓ F β w S α E m Max F β MAE↓ F β w S α E m Max F β MAE↓ F β w S α E m Max F β MAE↓ F β w S α E m
DRFI [24]0.7820.1700.4620.7200.7630.7770.1450.5280.7270.8320.6940.2010.4690.6480.7450.6640.1500.3260.6970.793
RBD [54]0.7160.1710.4300.6950.6970.7230.1420.4880.6830.7360.6590.1970.4290.6170.6700.6300.1440.3970.6680.721
MB+ [56]0.7390.1710.4280.6070.6910.7280.1500.4910.5340.6430.6800.1930.4530.7140.8140.6240.1680.3860.5790.693
MST [55]0.7310.1490.4450.6010.6860.7220.1680.4850.6930.7530.6700.1870.4580.6360.7150.6000.1490.3130.6530.688
SMD [40]0.7600.1730.4530.7160.7450.7430.1560.5020.7020.7960.6900.2010.4630.6450.7370.6240.1660.3850.6860.716
LEGS [28]0.8270.1180.8050.7860.8720.7670.1920.7360.7430.9310.7590.1550.7280.6700.2040.6310.713
MCDL [16]0.8370.1100.8160.8030.8890.8080.0910.7680.7860.9270.7430.1460.7870.7210.7060.7020.0880.6700.7520.670
DCL [17]0.8870.0720.8380.8680.9160.8800.0580.8410.8770.9310.8080.1100.7330.7850.8490.7170.0940.6390.7710.826
MDF [27]0.8340.1050.8100.7760.8860.8140.1120.7540.8100.8720.7680.1500.7040.6960.7940.6940.0920.6430.7210.820
ELD [53]0.8660.0790.7860.8380.9100.8390.0730.7800.8210.9100.7710.1260.6690.7610.8180.7000.0920.5960.7510.797
Scale-10.7830.1650.7440.7320.7650.7820.1430.6730.7290.8340.6930.1840.6820.7070.7520.6720.1570.4940.7010.762
Scale-20.8070.1430.7660.7630.8030.8070.1240.7150.7520.8580.7340.1530.7040.7250.7620.6880.1350.5580.7130.794
Scale-30.8230.1280.7830.7910.8240.8250.1060.7530.7900.8920.7570.1380.7180.7430.7780.6930.1170.5830.7210.807
Scale-40.8350.1040.8110.8030.8650.8340.0910.7760.8190.9130.7650.1270.7300.7540.7940.6990.1050.6230.7390.815
MGCC(Proposed)0.8910.0660.8470.8870.9310.8780.0570.8380.8860.9550.8080.1000.7960.7930.8570.7260.0740.6980.7930.838

Share and Cite

MDPI and ACS Style

Feng, W.; Li, X.; Gao, G.; Chen, X.; Liu, Q. Multi-Scale Global Contrast CNN for Salient Object Detection. Sensors 2020, 20, 2656. https://doi.org/10.3390/s20092656

AMA Style

Feng W, Li X, Gao G, Chen X, Liu Q. Multi-Scale Global Contrast CNN for Salient Object Detection. Sensors. 2020; 20(9):2656. https://doi.org/10.3390/s20092656

Chicago/Turabian Style

Feng, Weijia, Xiaohui Li, Guangshuai Gao, Xingyue Chen, and Qingjie Liu. 2020. "Multi-Scale Global Contrast CNN for Salient Object Detection" Sensors 20, no. 9: 2656. https://doi.org/10.3390/s20092656

APA Style

Feng, W., Li, X., Gao, G., Chen, X., & Liu, Q. (2020). Multi-Scale Global Contrast CNN for Salient Object Detection. Sensors, 20(9), 2656. https://doi.org/10.3390/s20092656

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop