Next Article in Journal
Optimization of PTFE Coating on PDMS Surfaces for Inhibition of Hydrophobic Molecule Absorption for Increased Optical Detection Sensitivity
Next Article in Special Issue
DNet: Dynamic Neighborhood Feature Learning in Point Cloud
Previous Article in Journal
Dynamic Set Planning for Coordinated Multi-Point in B4G/5G Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

NRA-Net—Neg-Region Attention Network for Salient Object Detection with Gaze Tracking

1
Department of Plasma Bio Display, Kwangwoon University, 20 Kwangwoon-ro, Nowon-gu, Seoul 01897, Korea
2
Department of Smart Convergence, Kwangwoon University, 20 Kwangwoon-ro, Nowon-gu, Seoul 01897, Korea
3
Ingenium College of Liberal Arts, Kwangwoon University, 20 Kwangwoon-ro, Nowon-gu, Seoul 01897, Korea
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(5), 1753; https://doi.org/10.3390/s21051753
Submission received: 25 January 2021 / Revised: 23 February 2021 / Accepted: 25 February 2021 / Published: 4 March 2021

Abstract

:
In this paper, we propose a detection method for salient objects whose eyes are focused on gaze tracking; this method does not require a device in a single image. A network was constructed using Neg-Region Attention (NRA), which predicts objects with a concentrated line of sight using deep learning techniques. The existing deep learning-based method has an autoencoder structure, which causes feature loss during the encoding process of compressing and extracting features from the image and the decoding process of expanding and restoring. As a result, a feature loss occurs in the area of the object from the detection results, or another area is detected as an object. The proposed method, that is, NRA, can be used for reducing feature loss and emphasizing object areas with encoders. After separating positive and negative regions using the exponential linear unit activation function, converted attention was performed for each region. The attention method provided without using the backbone network emphasized the object area and suppressed the background area. In the experimental results, the proposed method showed higher detection results than the conventional methods.

1. Introduction

Video processing technology using deep learning has been studied in various fields, such as monochrome image colorization, object detection and recognition, super-resolution technology, character detection and object recognition. Particularly, in the field of object detection and recognition, the underlying method is salient object detection (SOD). The purpose of this method is to detect objects that are of interest to a person, detect objects that cause the line of sight when a person first sees a video or a single image, and track the line of sight based on the results. SOD predicts where the gaze is focused when gaze tracking without equipment. Dramatic scene changes when predicting gaze can cause tracking to fail. SOD can supplement gaze tracking information by learning the area where the gaze is concentrated and predicting the input image. SOD’s training dataset generates an area where the eyes of 20 to 30 people are focused through gaze tracking when displaying images on the screen instantly. This method recognizes objects and is used in various areas, such as scene classification, tracking and detection. Other typical object detection methods include the contour-based, division-based, and deep learning-based methods.
SOD is the creation of a saliency map by detecting objects that are of interest to people or objects that are considered to be the most important in a video or image. The correct saliency map of the SOD input image is called the ground truth. Typical split-based SOD methods include superpixel, contour-based, and deep learning-based methods.
The deep learning-based detection method does not require complicated pretreatments and posttreatment processes and shows a high detection rate. Most of the existing SOD methods have autoencoder structures, and the use of some methods results in a significant deformation of the model structure and loss function. The model structure can be improved to reduce losses during the feature extraction process due to the shallow structure and to maximize the error of the loss function during the learning process. Research is underway to improve the detection rate of salient objects based on deep learning-based methods. However, the detection rate drops due to the high degree of similarity between the object and the background or the existence of several objects.
The existing deep learning-based method typically includes a fully convolutional network (FCN) [1], which uses skip connection to minimize losses during the feature extraction process. However, various values are extracted at the feature extraction stage, and feature values can be expressed as negative in this process. Traditional methods focus on positive region values that do not utilize negative region values, which causes feature loss. Also, since deep learning research is being conducted based on the backbone network, a backbone network is always required. A backbone network is a feature extractor that has learned a lot of data in advance. Although it has the advantage of being able to select various features, the type of backbone is limited and shows poor performance when extracting features that have not been learned.
In this study, the negative region that was not used in the existing method is utilized. A new attention module is created by using the spatial attention technique in the negative region. This module minimizes loss of functionality during feature extraction. We also propose a deep learning model called the Neg-Region Attention (NRA), which aims to minimize the feature loss of salient objects due to complex environmental problems.The proposed method did not use a backbone network to extract desired features and does not require additional pre-trained weights. It aims to construct a relatively light model without a backbone network. In addition, it aims to improve the performance of feature extraction by providing a new module using the negative region.

2. Related Works

Deep learning is a machine learning algorithm that summarizes the core contents and features of complex data, with nonlinear transformation methods composed of multiple layers. In the existing machine learning algorithm, a person directly analyzes and evaluates by extracting the kind of features present in the data to be learned. However, in deep learning, necessary features are extracted and learned from the data that the machine automatically learns. In these studies, deep learning was developed as a convolutional neural network-based method with high data recognition and detection performance. In addition, deep learning technology has been established, and excellent performance methods have been developed in the field of SOD, where FCN is a typical example.

2.1. Hand Crafted-Based Detection Method

Superpixel-based split methods [2,3] split the salient object and background with the internal information of the image, such as brightness, color, contrast and texture. Because salient objects have movements in the video, the method considers the position of the object according to the time using a superpixel partition. Contour-based detection methods [4] include detecting salient objects using a fast Fourier transform and a Gaussian filter. Such a method maintains the contour of the object and shows a high detection rate, but it requires a large amount of calculation due to pre-processing and post-processing process requirements. In addition, the detection rate of salient objects decreases due to problems such as complicated backgrounds, high similarity between backgrounds and objects, or the existence of several objects.

2.2. Deep Learning-Based Detection Method

The deep learning-based SOD method shows high accuracy in trained images without requiring complicated pre-processing and post-processing processes. In the present research, this method is performed with an autoencoder structure, and it can be further classified to methods transforming the network structure and transforming the loss function. When the network structure is deformed, the loss and shallow structure in the process of extracting the features are improved, which consequently improves the results. When the loss function is transformed, the loss function is improved, which consequently minimizes the error in the learning process. Both methods show improved performance, and even when transforming the loss function, it is necessary to improve the network structure in the process of minimizing the error [5,6,7,8,9].
Recently, many deep learning-based detection methods have been studied to improve the performance of SOD. However, the detection rate drops due to problems, such as high similarity between the background and object and complicated background. In addition, feature loss occurs during feature extraction through several convolution layers.

2.3. Autoencoder

An autoencoder [10] is a type of artificial neural network used to compress and restore image data. It is a learning model with a structure similar to that of feed-forward neural networks (FNNs) [11]. Different from FNNs, the sizes of the input and output layers of an autoencoder are always the same.
An autoencoder is largely composed of an encoder and a decoder. The encoder is likely a network that extracts features from the input data or compresses it into an internal representation. The decoder is a generation network that converts extracted features and compressed internal representations into the output. An autoencoder is a deep learning network structure that is often used in the field of SOD and partitioning. The autoencoder copies the input to the output only on the same side as the input layer with the same number of nodes in the hidden layer. Therefore, the number of nodes in the hidden layer is smaller than that in the input layer and the data are compressed, as shown in Figure 1. In this method, control is used to represent data efficiently. The upsampling of the decoding process causes feature loss as it is simply used in the process of increasing the size of the feature map.

3. Proposed Method

The proposed method, that is, NRA, has an autoencoder structure and research has been conducted to reduce losses that occur in the process of compressing and decompressing features and losses that occur in the process of expanding extracted features. Existing methods use the rectified linear unit (ReLU) [12] as an activation function in the encoder process of extracting features. ReLU treats negative regions as 0, so feature loss occurs, but the learning speed is fast. Instead of using ReLU, the conventional method uses a model in the backbone network where a large dataset is trained to prevent feature loss. The backbone network requires pre-trained weights, and there is a limit to model transformation. Most of the input images are also fixed and the model becomes heavier.
The proposed method aims to improve the heaviness of the model, change the size of the input image according to the user’s computer performance, and minimize the feature loss at the encoder stage. ELU is used as an activation function to avoid the feature loss problem that occurs in ReLU. The exponential linear unit (ELU) activation function is used in the encoding process to compress and extract features and utilize them in the negative region. The NRA provided for the extracted features can be used to suppress non-object areas in the negative areas and emphasize the contour and texture information of the objects in the positive areas. The decoding process, which enlarges the extracted features to the size of the input image, utilizes the features in the encoding stage through concatenation. Through this process, an improved saliency map is generated. The detection flowchart using NRA proposed in this paper is shown in Figure 2.

3.1. Feature Extraction Using the Proposed Attention

In a deep learning-based method, the convolution operation results involve negative and positive regions. The conventional method uses ReLU activation function to determine which node to pass to the next layer. The ReLU activation function treats the negative region as 0 and causes feature loss in the feature extraction process. Accordingly, the proposed method utilizes the negative region and uses the Exponential Linear Unit (ELU) [13] activation function to prevent feature loss. The positive region of the ELU activation function is processed similarly as the ReLU activation function, and the negative region has a convergence form of (1) and (2). Equations (1) and (2) are the equations of the ReLU and ELU activation functions, respectively, where the e x p ( ) function was used in the negative region in ELU. The graphs of the activation functions are presented in Figure 3.
R e L U ( x ) = m a x ( 0 , x )
E L U ( x ) = x , if   x 0 a ( e x 1 ) , if   x < 0 .
As shown in Figure 3, the graph of the ELU activation function directly outputs the input in the positive region, but in the negative region, it is normalized so that it is not outputted immediately and converges to a . By setting the values of a, the influence of the negative region can be limited. The proposed method uses all the characteristics of the positive and negative regions by setting a to 1, but the negative region has little effect.
When the value of a is set large, the effect of the negative region becomes large in the complete data representation, and the texture information around the boundary information in which the amount of data change is large is expressed in various ways. The proposed method uses integrated texture information rather than various texture information in the negative area and sets the value of a to 1 to suppress non-object areas.
We propose NRA to suppress the non-object area of the positive region and emphasize the object area using the negative region of the ELU. The proposed method was separately emphasized in the negative and positive regions after applying the ELU activation function. Figure 4 shows the proposed NRA structure.
The proposed NRA method is configured as follows after applying ELU, the process of separating the negative and positive areas results in an ELU feature map and an element-wise product spatial attention course using a 1 × 1 convolution in the separated area and an element of the emphasized feature map, which consists of a joint process using an element-wise sum. In Figure 4, neg_ELU represents the negative region of the ELU activation function and pos_ELU represents the positive region.
Equation (2) is the process of separating into a negative region (neg_ELU) and a positive region (pos_ELU), which are respectively shown in (3) and (4), respectively. Figure 5 is a graph showing the separation in the ELU.
E L U n ( x ) = 0 , if   x 0 a ( e x 1 ) , if   x < 0 , a = 1
E L U p ( x ) = x , if   x 0 0 , if   x < 0 .
Using the ELU activation function, the feature map was separated into a negative area and a positive area. In Equation (3), E L U n ( x ) is the negative region of the ELU when the input is x, and the positive region is treated as 0. Because a is set to 1 and converges to 1 , it can be confirmed that the texture information with a small amount of data change is unified and displayed. In Equation (4), E L U p ( x ) is a positive region of the ELU when the input is x and is processed in the same way as the ReLU function. The proposed method separates the positive and negative regions to take advantage of the properties of the ELU.
Figure 6 shows the module of Figure 4, which has a feature map separated into positive and negative regions, as shown in Figure 5. Figure 6c is a feature map showing the separation in the ELU. This map is different from that when the ReLU activation function is applied in the separation of the ELU feature map (b) into a positive region, so the normalization range is different and ReLU shows different results. In the positive region, various textures and boundary information are extracted according to the amount of data change in the image. These features are affected by color and brightness. The texture information of the salient object of lighting is extracted from the features of the non-object area, and the shadow features of the non-object area are extracted in the same way as the salient object. Figure 6d shows the results of dividing the saliency map into negative and positive regions, where a converges to 1 .

3.1.1. Extraction of Attention Region from Each Region

Because the SOD needs to detect the area of the salient object, not only the boundary information but also the texture information of the object to be detected is important. Various texture information can be obtained by outputting features through an ELU function with a wide range of feature expressions. However, environmental conditions, such as light reflection, affect color and brightness, and texture information is extracted based on such conditions, resulting in a loss area. Although the loss can be minimized by unifying the various texture information through an emphasis technique, a region with texture information similar to that of a salient object is emphasized, resulting in false detection. Therefore, the proposed method does not use the ELU function as it is and proceeds with the enhancement technique by separating the saliency map into a positive region representing various texture information and a negative region containing mainly unified texture information and boundary information with a large amount of change.
Areas of objects and backgrounds in the image are separated based on the boundary information. Spatial attention can be used for positive areas that have various textures, and boundary information can be used to unify the information of various textures and emphasize the salient object area. The feature separated into the negative region is different from that in the positive region, such that the feature converges to 1 and outputs unified texture information and boundary information with a large amount of data change. If these features are utilized without emphasis, then the area of the salient object can be suppressed as a non-object area. Spatial attention can be used to emphasize only the salient object area based on the boundary information to suppress the non-object area emphasized in the positive area.
The proposed method performs spatial attention in the positive and negative regions using Equations (5) and (6), respectively, and is a transformation of spatial attention. In the case of using the average and maximum pooling, a representative value is outputted in the separate positive region, resulting in a loss of various texture information. Because the sigmoid is normalized to a value between 0 and 1, the amount of data change is altered, so the boundary information is lost. Therefore, in the proposed method, spatial attention through a convolution and element-wise product is used without pooling and sigmoid.
A p ( x ) = f c o n v 1 × 1 ( E L U p ( f c o n v 3 × 3 ( x ) ) ) E L U ( f c o n v 3 × 3 ( x ) )
A n ( x ) = f c o n v 1 × 1 ( E L U n ( f c o n v 3 × 3 ( x ) ) ) E L U ( f c o n v 3 × 3 ( x ) ) .
In Equations (5) and (6), A p ( x ) and A n ( x ) represent the positive and negative regions of the input x; f c o n v 1 × 1 and f c o n v 3 × 3 are the 1 × 1 and 3 × 3 convolutions, respectively; the ⨂ is an element-wise product; E L U p and E L U n are the positive and negative regions separated from the ELU in Equations (3) and (4), respectively; and a was set to 1 in the ELU. The results of such a modified spatial attention equation are shown in Figure 7.
Figure 7 shows a feature map of each region separated from the ELU and a feature map emphasizing the positive and negative regions. Spatial attention was performed on the salient object to emphasize the detailed information from Figure 7a. Negative areas also emphasized the contour and texture information via spatial attention. Unlike the result of applying the ELU activation function, the area of the salient object was emphasized. The emphasis of the positive area emphasizes the texture information of the entire image, and the negative area suppresses the non-object area and emphasizes the contour and texture information.

3.1.2. Combination of Attention Positive and Negative Region

The texture information of the salient object was emphasized based on the boundary information between the salient object and the background, and the positive area where the shadow area was emphasized and the negative area where the non-object area was suppressed were combined with the element-wise sum. As a result of emphasizing the negative region through the combination, the shadow region, which is a non-object region, is suppressed, and the feature that the region of the salient object is emphasized through the emphasis on the positive and negative regions is obtained. Equation (7) shows the combination of the emphasized feature maps.
A e l e m e n t ( x ) = A p ( x ) + A n ( x ) .
The combination of the feature maps emphasized for input x is represented by A e l e m e n t ( x ) . A p represents a feature map with emphasized positive areas, and A n represents a feature map with emphasized negative areas.
Figure 8 shows the result of a combination of feature maps and shows an emphasis on each area. Bases on the spatial attention results of the positive region, which contains various detailed information, we have element-wise-summed the positive and negative feature maps emphasized to suppress the shadow features that are non-object regions. By combining the results of emphasizing the unified texture information in (c) and the result of emphasizing various texture information in (b), the non-object area is suppressed, as shown in Figure 8d. The boundary and texture information of the salient objects is also emphasized.
In the deep learning process, when the distance between the input and output increases as shown above, the slope value is saturated with a large or small value in the backpropagation process in which the weight is transmitted between layers when learning the network, resulting in an ineffective learning, a slope that slows learning, and loss problem. To prevent these problems, a structure that learns the difference between the input value and output value was constructed by applying skip connection after combining the highlighted feature maps.
Figure 9 compares the results of the highlighted feature map combination with the results of applying skip connection to the combined feature map. The problems of weights being propagated directly from the output to the input and the slope disappearing in deep structures when learning the network through a structure that applies skip connection are avoided. The final NRA result is shown in Figure 9c, in which the non-object area is suppressed and the detailed information of the object is emphasized. Equation (8) is the formula for NRA.
A N A ( x ) = A e l e m e n t ( x ) + x .
The result of NRA is represented by A N A ( x ) on the input x, and A e l e m e n t ( x ) and the input are element-wise-summed to utilize skip connection. A e l e m e n t ( x ) is a combination of the positive and negative feature maps emphasized in Equation (7).
Figure 10 is a comparison of the feature map to which the ELU activation function is applied when performing the convolution operation on the input image and the result of applying the proposed NRA to the result emphasized by these feature maps. When spatial attention is applied to the feature map extracted using the ELU activation function, the texture information of the salient object is extracted in the non-object area by lighting, as shown in Figure 10c. The proposed NRA method is separated into positive and negative regions by the ELU activation function, and then spatial attention is applied to each region to suppress non-object regions and emphasize the detailed information of salient objects.

3.2. Decoder for Extending Extracted Features

Because the decoding process is a stage where feature extraction and compressed features are expanded to the size of the input image and restored at the encoder stage, feature loss occurs during the expansion process. In this process, information on the correlation of surrounding pixels is lost. The proposed method uses concatenation to utilize the features of each stage extracted from the encoder at each stage of the decoder to prevent feature loss.
Unlike skip connection, which adds a feature map, concatenation simply follows. The number of feature maps is increasing, following the feature maps of the same size. Equation (9) is the formula for concatenation:
C o n c a t ( w , h ) = [ A N A ( x ) w × h ; D ( y ) w × h ] .
A N A ( x ) is the result of performing NRA on the input x, and w × h indicates the size of the resulting feature map. D ( y ) is the result of the inverse convolution of the input y with the decoder, and w × h indicates the size of the result feature map. C o n c a t ( w , h ) is the result of the concatenation of a feature map of the same size as w × h .

3.3. Residual Block in the Process of Concatenation Operation

When the NRA result of the encoder step is directly concatenated to the decoder, only the salient region is expressed because the highlighted feature is not refined. A residual block was used to generate a saliency map close to the ground truth through the feature refinement process of this highlighted region. Figure 11 shows the results of the concatenation of the NRA without feature refinement.
The results show only the emphasized regions where the features have not been purified. The features in the residual block were reconstructed to improve the quality of the saliency maps and generate them closer to the ground truth.
The existing residual block consists of two ( 3 × 3 ) convolutions and two ReLU activation functions. Such a structure does not take advantage of the negative region features using the ReLU activation function. Because the proposed method also utilizes features in the negative region, we used the ELU activation function to prevent the loss of features highlighted by the residual block. The information transmitted by skip connection in the proposed method emphasizes the salient area. When the ELU activation function was applied after receiving the emphasized feature information, the features in the emphasized negative region were normalized and feature loss occurred. Therefore, unlike the conventional method, the result was outputted without using the activation function after skip connection. Equation (10) is an equation of the residual block by the proposed method.
f r e s ( x ) = f c o n v 3 × 3 ( E L U ( f c o n v 3 × 3 ( x ) ) ) + x .
f r e s ( x ) is the result of the residual block in the proposed way of the input x, and f c o n v 3 × 3 is the 3 × 3 convolution. After extracting the features in the 3 × 3 convolution as the input, the features were enabled as an ELU function and the features were extracted again via a 3 × 3 convolution. Then, the input was added to the element-wise sum, and the skip-connection structure was used. Figure 12 shows the structure of the residual block.
The features of the structure of these proposed residual blocks were purified. This method also avoids the problem of slope disappearance in the skip connection, reduces loss, and generates a saliency map close to the ground truth. Figure 13 shows the final SOD results.

4. Experimental Environment

4.1. Environment and Dataset

In this paper, MSRA10K (10,000) Salient Object Database was used as the training dataset, and ECSSD (1000) was used as the validation dataset. The MSRA10K dataset was trained with a total of 80,000 images using rotation (0 , 90 , 180 , 270 ) and flipping. During the learning process, the verification dataset was used to compare the degree of convergence between the other datasets and to confirm the overfitting phenomenon in which the accuracy of only the training dataset increases. Experimental datasets were compared and analyzed using ECSSD, HKU-IS (4447), and DUT-OMRON (5182). Adaptive Moment Estimation (Adam) optimizer [14] was used as the optimization function. The initial learning rate was set to 0.0001, the batch size was set to 48, and the epoch was set to 80. GPU was trained and experimented using NVIDIA GeForce RTX 3090 24 GB. The learning rate was set through a number of experiments, and if the initial learning rate exceeds 0.0001, the learning convergence speed is fast and learning is not performed. It was adjusted through the learning rate scheduler according to the learning convergence speed. The size of the input image was set to 224 × 224, which is the most used for comparison with existing methods.

4.2. Loss Function

The proposed method uses the L2 loss function. It is used when there is only one type of object to be detected, such as SOD, or when only correct and incorrect answers are identified. When there are various types of objects to be detected, such as object recognition, loss is calculated for each type of object using cross-entropy. The L2 loss function calculates the error by comparing the saliency map predicted with the mean squared error (MSE) and the squared error of the ground truth. When calculating the error, there are outliers in which the value rapidly changes. MSE is greatly affected by these outliers, and the weights are adjusted accordingly. Equation (11) is the equation of the L2 loss function.
f L o s s ( x ) = i = 1 n ( y i y i ^ ) 2 .
y represents the ground truth, and y ^ represents the saliency map predicted by the proposed method. The result of summing the difference between the ground truth and the predicted saliency map is f L o s s , and reducing this value entails the weight adjustment of the learning process.

4.3. Evaluation Index

To compare and analyze the experimental results, the mean absolute error (MAE) [15], precision, recall, and F-measure [16] were used as evaluation indicators. Equation (12) is an expression of the evaluation index MAE.
M A E = 1 W × H x = 1 W y = 1 H S ( x , y ) G ( x , y ) .
S ( x , y ) represents the predicted saliency map, and G ( x , y ) represents the ground truth. W × H represents the size of the image. MAE is an error rate that represents the absolute error value between the ground truth and the predicted result, so the lower the value, the better the performance. Equation (13) is the expression of the precision and recall, and Equation (14) is the F-measure.
p r e c i s i o n = T P T P + F P , r e c a l l = T P T P + F N
F β = ( 1 + β 2 ) p r e c i s i o n × r e c a l l ( β 2 × p r e c i s i o n ) + r e c a l l , β 2 = 0.3 .
The precision, recall and F-measure are values that indicate accuracy, and a value higher than 0 indicates better performance. Precision and recall are calculated based on whether the ground truth and the pixel value at the same location in the saliency map are the same.
S-measure (Structure-measure) [17] simultaneously evaluates object-aware structural similarity and region-aware between a predicted saliency map and a ground truth. S-measure is Equation (15).
S = α × S o + ( 1 α ) × S r ,
where S o is the object-aware structural similarity measure and S r is the region-aware structural similarity measure. S-measure is a combination of two evaluations, and α = 0.5 was used.
E-measure(Enhanced-alignment measure) [18] combines the image-level mean value and local pixel values into one. Jointly capture local pixel matching information and image level statistics. E-measure is defined by Equation (16).
Q F M = 1 W × H x = 1 W y = 1 H ϕ F M ( x , y ) ,
where h and w are height and width of the map.

5. Experimental Results

5.1. Learning Convergence Experiment

Figure 14 is a comparative analysis graph of the loss convergence and precision convergence in the learning process of the conventional FCN method using the ReLU activation function and the FCN method using the ELU activation function. The loss converged faster in the learning process of the FCN (ELU) method using the ELU activation function than the conventional FCN (ReLU). The findings confirmed that using ELU instead of ReLU resulted in a faster convergence.

5.2. Training MSRA10K Dataset

Figure 15 is a comparison image of the proposed method and other existing methods and experimental results. The experimental results were compared with ECSSD (4 images), HKU-IS (4 images), and DUT-OMRON (4 images) as examples, and the comparison methods were ELD (Encoded Low level Distance map) [19], DS (Deep Saliency) [20], DCL (Deep Contrast Learning) [21], Amulet [22], DGRL (Detect Globally Refine Locally) [23] and AFNet [24], which are all deep learning-based methods. All of these methods use a backbone network, whereas the proposed method was trained without a backbone network. Other detectors use the backbone network for the encoding process, which is the feature extraction step, so various features can be easily extracted. However, the proposed method improves the performance of the feature extraction step by applying the NRA without a backbone to the encoding step and minimizes the loss of texture and contour information. ELD and DS were greatly affected by color and brightness, and detected the surroundings of the target object. DCL mainly detected a single object and detects other objects together. It was vulnerable to multi-object detection and showed a result that is sensitive to contour information. The Amulet detects the area of the target object, but if the input image is complex, the surroundings are also detected. In some of the result images, a background area other than the surrounding area was also detected. DGRL showed a clear detection results, but loss occurred in the detailed part and the surrounding area was detected together. AFNet showed the best performance compared to the previous methods and the area and contour of the object were preserved. If there were multiple small objects, some detection fails, and if the background was complex, the surroundings were detected together. The proposed method showed excellent detection performance for small objects and detected large and multiple objects well. As with the existing methods, when the background is complex, the surroundings were detected together, but false detection was reduced. When compared to existing methods by learning without a backbone network, it showed excellent performance, and the performance of the attention module using ELU was also proven.
Table 1 and Table 2 are comparison tables for the evaluation of the proposed method and other deep learning-based algorithms. The number of parameters of the proposed method is less than the average and shows excellent performance without using the backbone network. The MAE, mean F-measure ( β 2 = 0.3 ), S-measure and E-measure were measured for the datasets ECSSD, HKU-IS, and DUT-OMRON. The best performance numbers are expressed in red, the second is blue, and the third is green. The proposed method showed superior performance in the MAE, mean F-measure, S-measure and E-measure compared to the conventional methods using the backbone network.
Figure 16 and Figure 17 are the comparison diagrams of the precision and recall curves and F-measure curves of the proposed method and other deep learning algorithms. The proposed method shows excellent performance in both indicators. In this curve, the minimum recall value can be used as an indicator of robustness, where the higher the precision value of the minimum recall value, the more accurate the salient object prediction, which means that the background and foreground are well separated.
Figure 18 shows the result of detection of protruding objects for the motocross-jump video dataset. When performing gaze tracking, gaze detection may fail if a dramatic scene change occurs as in the video above. When a dramatic scene change occurs, there are factors such as the position of the object or the rotation of the camera. The detection of salient objects compensates for this problem and predicts objects in which human gaze is concentrated even with scene changes. This prediction result can supplement information on which object is mainly focused on gaze detection.

6. Conclusions

In this paper, we propose a deep learning-based method to detect salient objects in images in various environments. Existing deep learning-based methods proceed with an autoencoder structure, and feature loss occurs in the encoding process for extracting and compressing features and the decoding process for expanding and restoring the extracted features. Due to this feature loss, a background other than an object is detected, or an object with complex internal information fails to be detected. Most of the existing methods require a backbone network, and improve the network based on the backbone. However, feature extraction is limited, and it is difficult to extract specialized features for any object. The efficiency of the proposed method to reduce the feature loss in the autoencoder structure was studied. After separating the positive and negative regions through the NRA proposed in the encoding process of the autoencoder structure, the enhancement technique was performed. Positive numbers represent various textures and boundary information, and negative numbers mainly represent boundary information with a large amount of change in data. To utilize this characteristic information, spatial attention technique was performed in each domain. The proposed method prevents feature loss and creates a final saliency map by reconstructing features with a modified residual block. Existing deep learning methods extract features using a backbone network, but the proposed method achieves an excellent performance by extracting features using the attention technique without a backbone network.

Author Contributions

Conceptualization, H.K.; Methodology, H.K.; Software, H.K.; Investigation, H.K., and S.K.; Writing—Original Draft Preparation, H.K.; Writing—Review & Editing, S.L., and S.K.; Supervision, S.L.; Project Administration, S.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Code and dataset will be made available on request to the first author’s email with appropriate justification. The public site for each dataset is as follows. MSRA10K: https://mmcheng.net/msra10k/; ECSSD: http://www.cse.cuhk.edu.hk/leojia/projects/hsaliency/dataset.html; DUT-OMRON: http://saliencydetection.net/dut-omron/; HKU-IS: https://sites.google.com/site/ligb86/hkuis.

Acknowledgments

This research was supported by Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Science and ICT (2020R1F1A1069079).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  2. Liu, Z.; Zhang, X.; Luo, S.; Le Meur, O. Superpixel-based spatiotemporal saliency detection. IEEE Trans. Circuits Syst. Video Technol. 2014, 24, 1522–1540. [Google Scholar] [CrossRef]
  3. Li, J.; Liu, Z.; Zhang, X.; Le Meur, O.; Shen, L. Spatiotemporal saliency detection based on superpixel-level trajectory. Signal Process. Image Commun. 2015, 38, 100–114. [Google Scholar] [CrossRef]
  4. Achanta, R.; Hemami, S.; Estrada, F.; Susstrunk, S. Frequency-tuned salient region detection. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 1597–1604. [Google Scholar]
  5. Badrinarayanan, V.; Kendall, A.; SegNet, R.C. A deep convolutional encoder-decoder architecture for image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar] [CrossRef] [PubMed]
  6. Jiang, X.; Gao, Y.; Fang, Z.; Wang, P.; Huang, B. An end-to-end human segmentation by region proposed fully convolutional network. IEEE Access 2019, 7, 16395–16405. [Google Scholar] [CrossRef]
  7. Meng, F.; Guo, L.; Wu, Q.; Li, H. A new deep segmentation quality assessment network for refining bounding box based segmentation. IEEE Access 2019, 7, 59514–59523. [Google Scholar] [CrossRef]
  8. Wang, L.; Wang, L.; Lu, H.; Zhang, P.; Ruan, X. Saliency detection with recurrent fully convolutional networks. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 8–16 October 2016; pp. 825–841. [Google Scholar]
  9. Han, L.; Li, X.; Dong, Y. Convolutional edge constraint-based U-net for salient object detection. IEEE Access 2019, 7, 48890–48900. [Google Scholar] [CrossRef]
  10. Baldi, P. Autoencoders, unsupervised learning, and deep architectures. In Proceedings of the ICML Workshop on Unsupervised and Transfer Learning, Bellevue, WA, USA, 2 July 2012; pp. 37–49. [Google Scholar]
  11. Huang, G.B.; Zhu, Q.Y.; Siew, C.K. Extreme learning machine: A new learning scheme of feedforward neural networks. In Proceedings of the 2004 IEEE International Joint Conference on Neural Networks, Budapest, Hungary, 25–29 July 2004; pp. 985–990. [Google Scholar]
  12. Agarap, A.F. Deep learning using rectified linear units (relu). arXiv 2018, arXiv:1803.08375. [Google Scholar]
  13. Clevert, D.A.; Unterthiner, T.; Hochreiter, S. Fast and accurate deep network learning by exponential linear units (elus). arXiv 2015, arXiv:1511.07289. [Google Scholar]
  14. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  15. Chai, T.; Draxler, R.R. Root mean square error (RMSE) or mean absolute error (MAE)?—Arguments against avoiding RMSE in the literature. Geosci. Model Dev. 2014, 7, 1247–1250. [Google Scholar] [CrossRef] [Green Version]
  16. Powers, D.M. Evaluation: From precision, recall and F-measure to ROC, informedness, markedness and correlation. arXiv 2011, arXiv:2010.16061. [Google Scholar]
  17. Fan, D.P.; Gong, C.; Cao, Y.; Ren, B.; Cheng, M.M.; Borji, A. Enhanced-alignment measure for binary foreground map evaluation. arXiv 2018, arXiv:1805.10421. [Google Scholar]
  18. Fan, D.P.; Cheng, M.M.; Liu, Y.; Li, T.; Borji, A. Structure-measure: A new way to evaluate foreground maps. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 4548–4557. [Google Scholar]
  19. Lee, G.; Tai, Y.W.; Kim, J. Deep saliency with encoded low level distance map and high level features. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 660–668. [Google Scholar]
  20. Li, X.; Zhao, L.; Wei, L.; Yang, M.H.; Wu, F.; Zhuang, Y.; Wang, J. Deepsaliency: Multi-task deep neural network model for salient object detection. IEEE Trans. Image Process. 2016, 25, 3919–3930. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  21. Li, G.; Yu, Y. Deep contrast learning for salient object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 478–487. [Google Scholar]
  22. Zhang, P.; Wang, D.; Lu, H.; Wang, H.; Ruan, X. Amulet: Aggregating multi-level convolutional features for salient object detection. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 202–211. [Google Scholar]
  23. Wang, T.; Zhang, L.; Wang, S.; Lu, H.; Yang, G.; Ruan, X.; Borji, A. Detect globally, refine locally: A novel approach to saliency detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3127–3135. [Google Scholar]
  24. Feng, M.; Lu, H.; Ding, E. Attentive feedback network for boundary-aware salient object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 1623–1632. [Google Scholar]
Figure 1. Autoencoder architecture example.
Figure 1. Autoencoder architecture example.
Sensors 21 01753 g001
Figure 2. The overall flowchart of our Salient Object detection method. (Neg-Region Attention Network).
Figure 2. The overall flowchart of our Salient Object detection method. (Neg-Region Attention Network).
Sensors 21 01753 g002
Figure 3. The rectified linear unit (ReLU) activation function graph before (a) and after (b) is an Exponential Linear Unit (ELU) activation function graph.
Figure 3. The rectified linear unit (ReLU) activation function graph before (a) and after (b) is an Exponential Linear Unit (ELU) activation function graph.
Sensors 21 01753 g003
Figure 4. Neg-Region Attention module architecture.
Figure 4. Neg-Region Attention module architecture.
Sensors 21 01753 g004
Figure 5. The separated ELU activation function graph. (a) is the negative region and (b) is the positive region.
Figure 5. The separated ELU activation function graph. (a) is the negative region and (b) is the positive region.
Sensors 21 01753 g005
Figure 6. The ELU activation function and separated results. (a) is the input image and (b) is the ELU activation function results, (c) is the positive region, (d) is the negative region.
Figure 6. The ELU activation function and separated results. (a) is the input image and (b) is the ELU activation function results, (c) is the positive region, (d) is the negative region.
Sensors 21 01753 g006
Figure 7. Results of applying attention in each area. (a) is the positive region and (b) is the positive attention results, (c) is the negative region, (d) is the negative attention results.
Figure 7. Results of applying attention in each area. (a) is the positive region and (b) is the positive attention results, (c) is the negative region, (d) is the negative attention results.
Sensors 21 01753 g007
Figure 8. Sum of the elements of the attention result for each region. (a) is hte input image and (b) is the positive attention result, (c) is the negative attention result, (d) is the element-wise sum result.
Figure 8. Sum of the elements of the attention result for each region. (a) is hte input image and (b) is the positive attention result, (c) is the negative attention result, (d) is the element-wise sum result.
Sensors 21 01753 g008
Figure 9. The skip-connection after element-wise sum. (a) is the input image and (b) is the element-wise sum result, (c) is the skip-connection result.
Figure 9. The skip-connection after element-wise sum. (a) is the input image and (b) is the element-wise sum result, (c) is the skip-connection result.
Sensors 21 01753 g009
Figure 10. The Neg-Region Attention module results. (a) is the input image and (b) is the ELU activation function result, (c) is the attention result using the ELU activation function, (d) is the Neg-Region Attention module result.
Figure 10. The Neg-Region Attention module results. (a) is the input image and (b) is the ELU activation function result, (c) is the attention result using the ELU activation function, (d) is the Neg-Region Attention module result.
Sensors 21 01753 g010
Figure 11. Prediction result of the model using only the attention module. (a) is the input image and (b) is using only the attention module, (c) is groundtruth.
Figure 11. Prediction result of the model using only the attention module. (a) is the input image and (b) is using only the attention module, (c) is groundtruth.
Sensors 21 01753 g011
Figure 12. Composition of residual blocks. (a) is existing residual block and (b) is residual block.
Figure 12. Composition of residual blocks. (a) is existing residual block and (b) is residual block.
Sensors 21 01753 g012
Figure 13. Results of using the proposed residual block. (a) is the input image and (b) is using only the attention module, (c) is using the proposed residual block. (d) is ground truth.
Figure 13. Results of using the proposed residual block. (a) is the input image and (b) is using only the attention module, (c) is using the proposed residual block. (d) is ground truth.
Sensors 21 01753 g013
Figure 14. Loss convergence graph for each activation function in the FCN model. Red used the ReLU function, and blue used the ELU function.
Figure 14. Loss convergence graph for each activation function in the FCN model. Red used the ReLU function, and blue used the ELU function.
Sensors 21 01753 g014
Figure 15. Comparison of experimental results of the proposed method and other deep learning methods. (a) Input image, (b) ELD, (c) DS, (d) DCL, (e) Amulet, (f) DGRL, (g) AFNet, (h) NRA-Net (proposed method), (i) Groundtruth.
Figure 15. Comparison of experimental results of the proposed method and other deep learning methods. (a) Input image, (b) ELD, (c) DS, (d) DCL, (e) Amulet, (f) DGRL, (g) AFNet, (h) NRA-Net (proposed method), (i) Groundtruth.
Sensors 21 01753 g015
Figure 16. Comparison with other methods of precision and recall curve in each dataset. (a) PR curve of ECSSD dataset and (b) HKU-IS dataset, (c) DUT-OMRON dataset.
Figure 16. Comparison with other methods of precision and recall curve in each dataset. (a) PR curve of ECSSD dataset and (b) HKU-IS dataset, (c) DUT-OMRON dataset.
Sensors 21 01753 g016
Figure 17. Comparison with other methods of F-measure curve in each dataset. (a) F-measure curve of ECSSD dataset and (b) HKU-IS dataset, (c) DUT-OMRON dataset.
Figure 17. Comparison with other methods of F-measure curve in each dataset. (a) F-measure curve of ECSSD dataset and (b) HKU-IS dataset, (c) DUT-OMRON dataset.
Sensors 21 01753 g017
Figure 18. Salient object detection results for motocross-jump video of MIT300 dataset. (top) Input video frame and (bottom) results of saliency map.
Figure 18. Salient object detection results for motocross-jump video of MIT300 dataset. (top) Input video frame and (bottom) results of saliency map.
Sensors 21 01753 g018
Table 1. The experimental results of the ECSSD dataset.
Table 1. The experimental results of the ECSSD dataset.
MethodNumber of ParametersECSSD Dataset
MAEF-MeasureS-MeasureE-Measure
ELD28.37 M0.07960.81020.8380.881
DS134.27 M0.12160.82550.8200.874
DCL66.25 M0.14950.82930.8630.885
Amulet33.16 M0.05880.86840.8930.901
DGRL126.35 M0.04190.90630.9030.917
AFNet21.08 M0.04220.90850.9130.918
NRA-Net56.42 M0.04890.91260.8980.907
Table 2. The experimental results of HKU-IS and DUT-OMRON datasets.
Table 2. The experimental results of HKU-IS and DUT-OMRON datasets.
MethodHKU-IS DatasetDUT-OMRON Dataset
MAEF-MeasureS-MeasureE-MeasureMAEF-MeasureS-MeasureE-Measure
ELD0.07410.76940.8200.8800.09230.61100.7500.775
DS0.07800.78510.8520.8890.12040.60310.7500.761
DCL0.13590.85330.8600.9130.09710.68370.7640.801
Amulet0.05210.85420.8830.9100.09770.64740.7800.778
DGRL0.03630.88820.8940.9430.06180.73320.8060.848
AFNet0.03640.89040.9050.9420.05740.73820.8260.853
NRA-Net0.04280.89240.8940.9190.07060.74490.8110.836
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Kim, H.; Kwon, S.; Lee, S. NRA-Net—Neg-Region Attention Network for Salient Object Detection with Gaze Tracking. Sensors 2021, 21, 1753. https://doi.org/10.3390/s21051753

AMA Style

Kim H, Kwon S, Lee S. NRA-Net—Neg-Region Attention Network for Salient Object Detection with Gaze Tracking. Sensors. 2021; 21(5):1753. https://doi.org/10.3390/s21051753

Chicago/Turabian Style

Kim, Hoijun, Soonchul Kwon, and Seunghyun Lee. 2021. "NRA-Net—Neg-Region Attention Network for Salient Object Detection with Gaze Tracking" Sensors 21, no. 5: 1753. https://doi.org/10.3390/s21051753

APA Style

Kim, H., Kwon, S., & Lee, S. (2021). NRA-Net—Neg-Region Attention Network for Salient Object Detection with Gaze Tracking. Sensors, 21(5), 1753. https://doi.org/10.3390/s21051753

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop