Next Article in Journal
Morphological Fabrication of Rubber Cutaneous Receptors Embedded in a Stretchable Skin-Mimicking Human Tissue by the Utilization of Hybrid Fluid
Next Article in Special Issue
WPO-Net: Windowed Pose Optimization Network for Monocular Visual Odometry Estimation
Previous Article in Journal
Investigation on Beam Alignment of a Microstrip-Line Butler Matrix and an SIW Butler Matrix for 5G Beamforming Antennas through RF-to-RF Wireless Sensing and 64-QAM Tests
Previous Article in Special Issue
3D Object Detection with SLS-Fusion Network in Foggy Weather Conditions
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Comparison of Bottom-Up Models for Spatial Saliency Predictions in Autonomous Driving

by
Jaime Maldonado
* and
Lino Antoni Giefer
Cognitive Neuroinformatics, University of Bremen, Enrique-Schmidt-Straße 5, 28359 Bremen, Germany
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(20), 6825; https://doi.org/10.3390/s21206825
Submission received: 18 August 2021 / Revised: 6 October 2021 / Accepted: 9 October 2021 / Published: 14 October 2021
(This article belongs to the Special Issue Advanced Computer Vision Techniques for Autonomous Driving)

Abstract

:
Bottom-up saliency models identify the salient regions of an image based on features such as color, intensity and orientation. These models are typically used as predictors of human visual behavior and for computer vision tasks. In this paper, we conduct a systematic evaluation of the saliency maps computed with four selected bottom-up models on images of urban and highway traffic scenes. Saliency both over whole images and on object level is investigated and elaborated in terms of the energy and the entropy of the saliency maps. We identify significant differences with respect to the amount, size and shape-complexity of the salient areas computed by different models. Based on these findings, we analyze the likelihood that object instances fall within the salient areas of an image and investigate the agreement between the segments of traffic participants and the saliency maps of the different models. The overall and object-level analysis provides insights on the distinctive features of salient areas identified by different models, which can be used as selection criteria for prospective applications in autonomous driving such as object detection and tracking.

1. Introduction

Visual attention is the mechanism by which human beings can selectively process salient stimuli. The selection mechanism can be influenced by bottom-up and top-down factors [1]. Bottom-up factors refer to the features of the image (e.g., color, intensity and orientation). Top-down refers to the cognitive factors of the observer which determine whether an object or a region of the visual field is salient. In the context of driving, different regions of a street are salient depending on the task, e.g., whether the driver is looking for a parking spot or just driving forward.
Bottom-up factors have been extensively studied in the literature and many computational models to identify salient regions have been proposed [1]. Depending on the computational mechanisms involved, as well as the features or cues used to detect saliency, different bottom-up models identify salient areas of an image differently (see Section 2). As a result, salient areas identified by different models differ in terms of their size, shape and location. Furthermore, an area identified as salient by one model can be regarded as non-salient by another model. These differences are illustrated in Figure 1, showing the saliency maps of an image generated by different bottom-up models.
Bottom-up saliency models have been used in computer vision applications including object detection and recognition, robot navigation and localization, and image processing (for a comprehensive list of applications see [1]). Such computer vision tasks can be encountered on applications of Advanced Driver Assistance Systems (ADAS) and highly autonomous driving (HAD). In these applications, an input image is initially processed with a bottom-up model to obtain a saliency map. As a result of the feature extraction conducted by the saliency model, the salient regions indicate the locations of proto-objects, which represent coherent areas that approximate whole, partial or groups of objects in the image. Subsequent processing steps focus on the salient areas. The incorporation of bottom-up models in object detection, object segmentation and object recognition applications enables a faster processing of the image compared to approaches that analyze the full scene to detect the objects by, for example, shifting analysis windows [2]. In the field of bottom-up saliency for object detection, Silva et al. [3] proposed a method to improve detection performance and execution speed. Specifically, saliency was used to prune the search space for objects. They evaluated their approach over a dataset for person detection in different types of scenes including cities and outdoor landscapes. They emphasize the fact that bottom-up saliency can be affected by uncontrolled factors in scenes, such as variations in color, size, illumination and noise of the target objects.
Differences in performance resulting from the use of different bottom-up saliency models have been investigated for object detection [4] and driver gaze prediction applications [5]. Duthon et al. [4] evaluated different bottom-up saliency models to test their applicability for object detection in autonomous vehicles. The experimental results show that bottom-up saliency on its own is not sufficient for robust detection in the road context. In this case, detection is not reliable because not all relevant objects (e.g., cars, bicycles or pedestrians) are necessarily salient. Nevertheless, they emphasize that bottom-up saliency can facilitate detection when used as a pre-processing step. They also conclude that factors such as the task and the context (e.g., type of landscape, point of view) can influence the performance and that the selection of a saliency model should consider the target application.
Deng et al. [5] evaluated the use of different state-of-the-art bottom-up saliency models for traffic saliency detection. The experimental results obtained from the comparison of the predicted saliency and human gaze data indicate that bottom-up saliency models cannot be directly applied to predict the drivers’ allocation of visual attention. They show that this limitation can be resolved by adding a top-down control to modulate the bottom-up prediction.
Typically, saliency models are evaluated using benchmark datasets for object detection [6] and human gaze allocation during free-viewing tasks [7]. Benchmark datasets aim at testing the applicability of saliency models over images with diverse characteristics: natural indoor and outdoor scenes, artificial stimuli (e.g., patterns), and a wide variety of objects and settings. On the contrary, context- and task-specific evaluations reveal how the models perform over images of particular characteristics and/or attention tasks. This is exemplified in the evaluation performed by Deng et al., where saliency models that are successful at predicting gaze allocation during free-viewing of images across different categories fail to predict where drivers look at while observing driving scenes [5].
To use a bottom-up saliency model in a particular autonomous driving application, it is important to consider the characteristics of the images registered by the frontal camera of a vehicle. Frontal camera images provide panoramic views of the road ahead in which the low-level features of the image, such as color, contrast and object size, might differ a lot depending on the landscape (e.g., urban or rural roads), traffic conditions (e.g., light or heavy traffic), time of day (e.g., day or night) and weather conditions. Previous evaluations of bottom-up saliency models in driving applications have been focused on task performance (i.e., driver gaze prediction or object detection). However, those studies provide no insights on how the distinctive features of the saliency predictions produced by different models influence the outcome.
In this paper, we conduct a systematic evaluation of the saliency maps computed with different bottom-up models on urban and highway scenes. Our goal is to compare the size and shape of the salient areas identified by different models and the extent to which traffic participants in the image fall within them. For this purpose, we compute the saliency of the images available in the KITTI semantic instance segmentation dataset [8]. The evaluation is based on energy and entropy features which represent the size and shape-complexity of the salient areas. In addition, we assess the agreement between the salient areas and the segments corresponding to traffic participants in the images. The features used to assess the saliency maps enable us to perform a comprehensive quantitative and qualitative comparison of different models. It is important to note that we consider the saliency computation as the initial step of a computer vision processing pipeline. Therefore, the evaluation features and criteria are not restricted to the outcome of a particular computer vision task, such as object segmentation or recognition, which depend on how the identified salient locations are processed. In this sense, our analysis describes and compares the salient areas identified in a prospective initial processing step with respect to the traffic participants in the picture. To the best of our knowledge, no other studies have been devoted to compare the size and shape complexity of salient areas and their relation to object instances in the image. The analysis aims to provide insights on how distinct salient areas are identified by different models, both over the whole image as well as on the object level. We encounter significant differences between the models which indicate the extent to which object instances (such as cars or pedestrians) fall within the salient areas.
The rest of the paper is organized as follows. In Section 2, we provide details about different types of bottom-up saliency models and describe the main characteristics of the ones that our study is based on. Section 3 focuses on details about the dataset we employ, and the computation of saliency maps followed by a thorough evaluation of those based on their energy, entropy and the agreement between the salient areas and the segments corresponding to traffic participants in Section 4. Based on the evaluation, we discuss how our results provide selection criteria to choose a particular saliency model for prospective autonomous driving applications in Section 5.

2. Bottom-Up Saliency Models

Based on their computational mechanisms, bottom-up models can be classified into four categories [9]: (1) Rarity/Contrast-Based, which compute center–surround contrast and/or local rarity/contrast based on image features, (2) Spectral Analysis Models, which are based on the frequency spectrum of the image, (3) Learning-Based Models, in which machine learning models are trained using low, middle and high-level features and/or eye tracking data, and (4) Salient Object-Detection Models, which aim to segment boundaries of salient objects by highlighting overall foreground regions.
In general, a saliency model takes an image I ( x , y ) with pixel coordinates x and y as input and outputs a saliency map S ( x , y ) representing the conspicuity or saliency at every location in the image by a scalar quantity [10]. Thus, saliency maps are typically displayed as intensity images, where the intensity of each pixel represents its probability of belonging to salient regions or objects. Although models belonging to the first three categories usually identify sparse blob-like salient regions aimed to predict gaze fixations, salient object-detection models often generate smooth connected areas [6].
In this work, four bottom-up models are evaluated for comparison covering three of the above-mentioned categories. In the rarity/contrast-based category we consider the IttiKoch model [10], which is based on feature integration theory, and the Graph-Based Visual Saliency (GBVS) [11], which is based on graph theory [2]. Both models perform low-level feature extraction and integration. In the spectral analysis category, we include the Spectral Residual (SR) model [12]. Finally, the Boolean Map Saliency (BMS) model [9] is considered for the salient object-detection category. The selected models are representative of each category and are frequently cited in the literature. Furthermore, three of them (IttiKoch, GBVS and SR) have been analyzed in previous studies in the context of traffic saliency detection [5] and two of them (GBVS and SR) in the context of object detection for autonomous vehicles [4]. To conduct a fair comparison, we exclude learning-based approaches as they depend on the quality of the training data and the learning capabilities of the machine learning model. In the following, we describe the main characteristics of the selected models.

2.1. IttiKoch: Rarity and Center—Surround Contrast Model

The model proposed by Itti et al. [10] is regarded as the baseline saliency model [2] and is referred to as the Itti or IttiKoch model in the literature. The model computes feature maps from an image’s intensity, color and orientation, which represent the feature at every location of the image by a scalar quantity [10]. These low-level features are known to attract human visual attention [2]. These feature maps are analyzed in different scales (i.e., in different resolutions) to account for objects and locations of different sizes.
The effect of the intensity, color and orientation on the saliency of a pixel region depends on the contrast with its surroundings. Thus, the edges of objects in the image, as well as regions that locally stand out from their surroundings, are highlighted by means of a center–surround operation [2]. The resulting saliency map represents local saliency over the entire image [10]. Itti et al. pointed out the applicability of the identified salient image locations for subsequent computer vision tasks, such as object detection [10].

2.2. Graph-Based Visual Saliency (GBVS): Rarity-Based Model

The GBVS model proposed by Harel et al. [11] extracts intensity, color and orientation features maps such as the IttiKoch model but omits the center–surround process for orientation maps and reduces the number of scales with which the image is processed [2]. As a result, the GBVS model groups sparse edges into integral regions that stand out from their surroundings [2]. GBVS produces high-saliency values in the center of the image plane (center bias) and is regarded to be robust with respect to differences in the sizes of salient regions [11].
Harel et al. argue that the center bias of the model is well suited for predicting human gaze allocation based on two observations: (1) everyday life head motion often results in gazing straight ahead and (2) the motif of photographs is typically located in the center [11].

2.3. Spectral Residual (SR): Spectral Analysis Model

The SR model proposed by Hou and Zhang [12] aims to simulate the behavior of pre-attentive visual search, in which low-level features such as orientation, edges, or intensities stand out automatically. In the SR model, the spatial frequency content of an image represents novel and redundant information, where a peak in the frequency spectrum is considered novel information. Based on this assumption, SR approximates the salient parts of an image by removing the statistical redundant components. The saliency map is computed based on the spectral residual of the image in the frequency domain, based on the hypothesis that saliency is the residual difference between the spectrum and the characteristic spectrum of natural images [2].
Since SR removes the statistical redundant components of an image, the saliency map highlights the non-trivial regions of the scene and suggests the positions of proto-objects [12]. SR is regarded by its authors as a general-purpose saliency detection system, well suited for object-detection applications since the saliency computation does not rely on features, categories, or other prior knowledge about the objects [12]. In contrast to the IttiKoch and GBVS models, the saliency computation of the SR model does not need to compute color, intensity and orientation feature maps or to analyze the image at different scales.

2.4. Boolean Map Saliency (BMS): Enclosure-Based Figure-Ground Segregation

The BMS model proposed by Zhang and Sclaroff [9] detects salient regions with closed outer contours based on the surroundedness (enclosure) cue for figure-ground segregation. As a result of the surroundedness cue, the model does not assign high-saliency values to high-contrast boundary areas typical of natural images (e.g., the boundary between the trees and the sky). Thus, compared to other models, BMS is less responsive to the edges and cluttered areas in the background [9]. Another effect of saliency detection based on the surroundedness cue is that the model highlights the interior regions of salient objects [9]. BMS can identify salient regions of different sizes due to the scale-invariant nature of the surroundedness cue and thus does not require the image to be processed at different scales [9].
BMS was proposed as a saliency detection model for eye fixation prediction but has been shown to be useful for salient object detection [9], and for the detection of proto-objects [6] outperforming several salient object-detection models [6]. Furthermore, it was ranked as the best non-neural network model on the MIT300 Benchmark [7].
To demonstrate the qualitative differences between the output generated by different models, we provide an example of the saliency maps computed for a video sequence of the KITTI object tracking dataset in Supplementary Materials.

3. Materials and Methods

We analyze the saliency maps obtained from semantically annotated images from urban and highway scenes, where we compute four maps from each image with the IttiKoch, GBVS, SR and BMS models. From each saliency map, we compute two features that describe the amount and shape-complexity of the salient areas. Subsequently, we assess the agreement between the salient areas and the segments corresponding to the traffic participants by means of an error measure. Finally, we compute the proportion of salient pixels for different saliency thresholds. This enables us to compare the size of the salient areas identified by different models both over the whole image, as well as on the object instance level. The statistical analysis of the differences between bottom-up models was performed with R [13] and the package PMCMR for pairwise comparisons [14].

3.1. Dataset and Computation of Saliency Maps

For the evaluation and comparison of the saliency maps produced by the selected models, we use the KITTI semantic instance segmentation dataset [8]. We choose this dataset because of the variety contained in the images with respect to classes of traffic participants (pedestrians and different types of vehicles), their position and size, traffic situations (from a single to several vehicles) and environments (urban, rural and highway scenarios). Specifically, we compute the saliency of the set of 200 training images, and we use the pixel-level semantic instance segmentation annotations to assess the agreement between the salient areas and segmented traffic participants in the image. We use the term traffic participants to refer to the following object categories annotated in the dataset: car, truck, person, bicycle, rider, bus, train, motorcycle, caravan, and trailer.
Saliency maps of the IttiKoch, GBVS and BMS models are computed using Pysaliency [15], a Python package for saliency modeling. The saliency maps of the SR model are computed using the OpenCV implementation (library version 4.0.1) [16]. We compute the SR with a resolution parameter of 64 pixels, which provides a good estimation of the scale of normal visual conditions [12]. An example of the saliency maps computed with the selected models is shown in Figure 1. The saliency maps produced by the IttiKoch and the BMS models are normalized such that 0 S ( x , y ) 1 .

3.2. Description of the Amount and Shape-Complexity of Saliency Maps Based on Energy and Entropy Features

The saliency maps produced by different models are compared in terms of their energy and entropy, as defined by Perrin et al. [17], which represent the amount and shape-complexity of the salient areas in a saliency map. Based on the observation that salient areas identified with different models differ in their shape, size and saliency level, we use the energy and entropy features in this paper to provide an objective comparison between the saliency maps obtained with different bottom-up models. Our use of the energy and entropy features differs from the approach in [17], where these measures are used to assess human saliency maps (i.e., saliency maps computed from human gaze data) for the evaluation of content-wise biases of images acquired by unmanned aerial vehicles.
The energy E ( x , y ) of a saliency map S ( x , y ) , defined in (1), is the sum of the vertical and horizontal gradient absolute magnitudes [17]. A Sobel filter [18] of kernel size 5 is used as derivative operator, following the implementation in [17].
E ( x , y ) = S ( x , y ) x + S ( x , y ) y
As energy feature, we take the mean of E ( x , y ) over all pixels of the image [17]. Following the interpretation of Perrin et al., high mean energy indicates that the saliency map contains several salient regions or shape-wise complex areas of interest, whereas low energy indicates more simple-shaped salient regions [17]. To illustrate this interpretation, the saliency maps with the maximum and minimum mean energy are shown in Figure 2.
The Shannon entropy [19] H of a saliency map is defined in (2), where p ( x , y ) is the probability of a pixel to be salient.
H = x , y p ( x , y ) log ( p ( x , y ) )
A saliency map S ( x , y ) is converted into a probability distribution using a soft-max function:
p ( x , y ) = exp S ( x , y ) x , y exp S ( x , y )
such that 0 p ( x , y ) 1 and x , y p ( x , y ) = 1 . Entropy has been used as feature for the evaluation of images and saliency maps [20,21], where larger entropy values indicate the presence of a complicated texture or structure [17,21]. Following the interpretation of Perrin et al. [17], a high value of entropy indicates that the saliency map contains a lot of information (i.e., “it is likely that saliency is complex”) and a low entropy value indicates a single zone of salience. This interpretation is illustrated with the saliency maps with the maximum and minimum entropy in Figure 3.

3.3. Evaluation of the Agreement between the Salient Areas and the Traffic Participants

Different measures have been proposed to measure the agreement between saliency predictions and annotated ground-truth. Two approaches are distinguished in the literature [6]: (1) metrics based on the overlap between salient areas identified by a model and the ground-truth (i.e., a marked region in the image), and (2) measures of the accuracy of the saliency maps with respect to the boundaries of the ground-truth regions.
To measure the agreement between a saliency map and ground-truth regions the Mean Absolute Error (MAE) is recommended over other overlap-based measures as it considers both true positives (i.e., the pixels correctly marked as salient) and the true negatives (i.e., the pixels correctly marked as non-salient) in a single measure [6,22]. (For a review of other standard and universally agreed measures in the context of object detection see [6,22], and in the context of human gaze prediction see [23].) (For a review of other standard and universally agreed measures in the context of object detection see [6,22], and in the context of human gaze prediction see [23].) (For a review of other standard and universally agreed measures in the context of object detection see [6,22], and in the context of human gaze prediction see [23].) The MAE is defined in (4) for continuous saliency map S ( x , y ) normalized to the range [0, 1] and the binary ground-truth G ( x , y ) (i.e., 1 within a marked region and 0 outside), where w and h represent the width and the height of the saliency map, respectively.
M A E = 1 w × h x , y S ( x , y ) G ( x , y )
The MAE provides an overall measure of the quality of a saliency map [22]. Smaller MAE values correspond to a better agreement between the salient regions and the objects in the image. Therefore, we compute the MAE between the saliency maps obtained with different bottom-up models and the binary ground-truth masks containing the traffic participants in the image.

3.4. Overall Proportion of Salient Areas

Saliency maps computed with different models differ in the size of the areas identified as salient. We analyze how much of the image is salient by thresholding the saliency maps. As the threshold increases, the number of pixels in the salient area decreases. We quantify this using (5) by computing the proportion P S t h of salient pixels for each of the following saliency thresholds t h : 0.25, 0.5 and 0.75, where n is the number of pixels in the salient area and N is the total number of pixels in the image. An example of how the proportion of salient pixels decreases for different saliency thresholds is shown in Figure 4.
P S t h = n N

3.5. Object Instance Saliency

We analyze the extent to which object instances are contained within the salient area of the image for different saliency thresholds. As the threshold increases, the number of object pixels within the salient area decreases. In (6), we compute the proportion of salient pixels within the object instance O P S t h for each of the following saliency thresholds t h : 0.25, 0.5 and 0.75, where n s a l i e n t i n o b j e c t is the number of object pixels within the salient area and N o b j e c t the total number of object instance pixels.
O P S t h = n s a l i e n t i n o b j e c t N o b j e c t
To account for cases in which an object instance is partially within the salient area, the object instance is regarded as salient if O P S t h > 0.5 . Thus, as the saliency threshold increases, fewer object instances are salient. An example of the number of salient instances for different saliency thresholds is shown in Figure 4.

4. Results

4.1. Energy and Entropy Comparison of Saliency Maps

The distribution of the mean energy values obtained with the selected models is shown in Figure 5a. There are significant differences in the mean energy of the saliency maps computed with different models under a repeated measures ANOVA test [24] F ( 3 , 796 ) = 4.632 , p < 0.001 . A post-hoc Tukey test [25] (see Table 1) revealed that all the pairwise differences are significant ( p < 0.001 ). These results indicate that the number of salient regions and their shape complexity differ across models (cf. Figure 1). Although BMS produces more simple-shaped salient areas, SR produces a larger number of complex-shaped salient areas (cf. Figure 2).
The distribution of the entropy values obtained with the selected models is shown in Figure 5b. There are significant differences in the entropy of the saliency maps computed with different models under a Friedman test [26] χ 2 ( 3 ) = 352.4 , p < 0.001 . A post-hoc Nemenyi pairwise test [27] (see Table 2) revealed that except for the pair BMS-IttiKoch, all the pairwise differences are significant ( p < 0.001 ). These results indicate that SR produces the most complex saliency maps in terms of the number of salient areas. In contrast, the entropy values indicate that the saliency produced by the GBVS model tends to be concentrated in a single zone.

4.2. MAE between Saliency Maps and Traffic Participants

The distribution of MAE values obtained with the selected models is shown in Figure 6. There are significant differences under a Friedman test: χ 2 ( 3 ) = 379.75 , p < 0.0001 . A post-hoc Nemenyi pairwise test (see Table 3) reveals that except for the pair BMS-GBVS, all the pairwise differences are significant ( p < 0.001 ). These results indicate that the best agreement (i.e., smaller MAE values) between the segments of the traffic participants and the saliency maps is obtained with the IttiKoch and SR models.

4.3. Comparison of the Proportion of Salient Areas

The proportion of salient pixels over different thresholds and saliency models is shown in Figure 7. There are significant differences under a Friedman test: χ 2 ( 11 ) = 1909.3 , p < 0.001 . The results obtained with a post-hoc Nemenyi pairwise test for the combinations of threshold and model are shown in Table 4. Note that cross-threshold pairs lack of practical meaning. Therefore, they are not interpreted in the analysis. We encounter significant differences ( p < 0.05 ), except for the following pairs: BMS-GBVS and IttiKoch-SR in the 0.25 threshold, BMS-IttiKoch in the 0.5 threshold, and BMS-IttiKoch in the 0.75 threshold. These results indicate that for the 0.5 and 0.75 thresholds the SR and GBVS models produce the smallest and largest proportion of salient pixels, respectively. Based on these results, it is expected that the number of objects within the salient areas differs across the models. Therefore, in the following section, we perform a thresholding analysis to determine the extent to which object instances are contained within the salient areas of an image.

4.4. Comparison of Object Instance Saliency

The percentage of salient object instances with respect to the total number of object instances in the dataset for different saliency thresholds is shown in Table 5. The table indicates that for all the objects, the largest percentage of salient instances is identified with the GBVS model for all the saliency thresholds. The results shown for the rider, bus, train, motorcycle, caravan, and trailer should be interpreted with caution due to the low number of instances with respect to the total number of images in the dataset.

4.5. Qualitative Analysis

Based on the quantitative results shown in the previous sections we perform a qualitative analysis of the saliency maps obtained with different models. This analysis aims to illustrate the cases in which the saliency models show the best and worst agreement with respect to the segments of traffic participants, highlighting the advantages and shortcomings of each model.

4.5.1. Qualitative Analysis BMS Saliency Maps

The top 5 best and worst agreements between the segments of traffic participants and the saliency maps computed with the BMS model are shown in Figure 8. The saliency maps with the best agreements show that while smaller traffic participants in the background fall within a uniform salient area, the mid-size vehicles in the foreground generate high-saliency blobs. The simple-shaped high-saliency areas can be attributed to the closed contours of the vehicles. It is important to note that even in the pictures with small MAE values, elements such as lights (image 000104_10.png ) and traffic signs (images 000038_10.png and 000043_10.png) generate areas of higher salience compared to the ones located over the traffic participants.
The saliency maps with the worst agreements show that even though smaller traffic participants in the background fall within salient areas, elements in the background such as tree branches and the grassy field form closed contours which are identified by the model as salient. Such spurious salient areas lead to a large MAE. The image with the train (000144_10.png) shows another way in which closed contours can generate large MAE values. The closed shapes formed by the train’s windows and its white side generate high-saliency blobs, which result in a non-uniform saliency distribution over the ground-truth segment.

4.5.2. Qualitative Analysis of GBVS Saliency Maps

The top 5 best and worst agreements between the segments of traffic participants and the saliency maps computed with the GBVS model are shown in Figure 9. The saliency maps with the best agreements show that the model produces a good match for mid- and small-size traffic participants located in the central-horizontal regions of the image. This result can be attributed to the center bias of the model. The examples also illustrate how the model groups sparse edges into regions (images 000101_10.png and 000128_10.png), by which small-sized vehicles in the background fall into high-saliency blobs. The saliency maps with the worst agreements show that large segments of traffic participants located at the extreme sides of the image, such as parked vehicles or vehicles approaching from the sides, fall out of the salient areas resulting in large MAE values. This can also be attributed to the center bias of the model.

4.5.3. Qualitative Analysis of IttiKoch Saliency Maps

The top 5 best and worst agreements between the segments of traffic participants and the saliency maps computed with the IttiKoch model are shown in Figure 10. The saliency maps with the best agreements show that images with mid-sized traffic participants which locally stand out from their surroundings produce small MAE values. The saliency maps with the worst agreements show that as in the case of GBVS, large segments of traffic participants located at the extreme sides of the image fall out of the salient areas resulting in large MAE values. In this case, the center–surround contrast operation highlights only small portions of the vehicles, which results in separated salient areas.

4.5.4. Qualitative Analysis of SR Saliency Maps

The top 5 best and worst agreements between the segments of traffic participants and the saliency maps computed with the SR model are shown in Figure 11. The saliency maps with the best agreements show that mid- and small-sized traffic participants located in the central-horizontal regions of the image fall within the salient regions. The saliency maps with the worst agreements show that as in the case of GBVS and IttiKoch, large segments of traffic participants located at the extreme sides of the image produce large MAE values. In these examples, the parked vehicles produce small salient areas of complex shapes corresponding to the boundaries of the vehicles. In addition to this, background elements such as trees and buildings produce spurious high-saliency areas.

5. Discussion

In this paper, we conducted a systematic evaluation of the saliency maps computed with different bottom-up models. The number of salient areas and their shape-complexity was analyzed by comparing the mean energy of the saliency maps. Additionally, whether saliency is distributed over different areas or concentrated within a single zone was analyzed by comparing the entropy of the saliency maps.
Regarding the shape complexity of the salient areas, the BMS and SR models produced simple-blob-like and complex-shaped areas, respectively. The entropy values revealed that although saliency produced by the SR model is distributed over several areas, the saliency produced by the GBVS model tends to be concentrated in a single zone.
The analysis of the MAE values revealed that the best agreements between the salient areas and the segments of the traffic participants are obtained by the IttiKoch model, followed by SR. A qualitative analysis showed that segments of traffic participants located at the extreme sides of the image, which occur when there are parked vehicles or when vehicles approach from the side, result in large MAE values with the GBVS, IttiKoch and SR models. The analysis also showed that background elements, such as tree branches and buildings, produce high-saliency regions with the BMS and SR models, which results in a reduced agreement between the salient areas and the traffic participants.
The insights about the shape and the number of salient areas were complemented by analyzing their size, quantified as the proportion of salient pixels. The comparison of this proportion over different saliency thresholds revealed that the SR and GBVS produce smaller and larger salient regions, respectively.
Given the differences between models with respect to the amount, shape and size of the salient areas, we analyzed the extent to which object instances are contained within the salient areas of an image. The analysis indicates that for the car, truck, person, and bicycle objects the largest percentage of salient instances is identified with the GBVS model across different saliency thresholds.
The analysis suggests that due to the smaller size and the shape complexity of the SR saliency maps fewer object instances fall within the salient areas of an image. In contrast, the larger size and concentration of saliency over one area of the GBVS model increases the likelihood of object instances being contained within the salient areas of an image. In this respect, it is important to recall that the salient areas of SR maps show a better agreement with the segments of traffic participants in contrast with the worse agreement obtained with GBVS. Altogether, these results constitute a trade-off between coverage and agreement for the SR and GBVS models.
It is important to note that saliency of particular object categories can be improved by means of adding a top-down prior component such as the vanishing point of the road [5] or the horizon line of the image [28]. Based on our qualitative analysis, a top-down prior could be applied to reduce the saliency assigned to trees and buildings and to increase saliency to traffic participants on the side of the image. Furthermore, the saliency of particular object categories or image areas can also be improved by combining the predictions of different bottom-up models.
The suitability of a particular model depends on the application. Although for object segmentation it is desirable that the high-saliency areas correspond to the object’s contours, for object detection large portions of the object should be located within salient areas. Furthermore, in object-detection applications for autonomous vehicles is highly important that all traffic participants fall within a predicted map because otherwise they become invisible and could lead to dangerous situations. Our evaluation method and the features used to characterize the saliency maps provide selection criteria that can be applied to different computer vision and prediction of driver behavior tasks. For example, SR saliency maps might be appropriate for a segmentation pipeline due to the large number of complex-shaped salient areas, as quantified by the entropy and energy features. On the other hand, the large and simple-shaped salient areas with center bias produced by the GBVS model can be used to prune the surveillance area of an algorithm.
To extend the insights obtained from our quantitative and qualitative results, further evaluations need to be conducted on other datasets including various light, weather and traffic conditions, which might introduce a large variability in the images registered by the frontal camera of a vehicle. However, it is important to note that such an evaluation is challenging to conduct with full control of factors to achieve a fair comparison. For example, a fair comparison between different rain intensities would require constant background and constant, or at least highly comparable, positioning of traffic participants and other elements in the scene. Furthermore, future work will be focused on the combination of bottom-up and top-down models to obtain reliable prior information for object detection and tracking algorithms such as [29]. In this way, we assume to reduce both the computational effort and the number of false detections resulting in more accurate environmental perception for autonomous driving scenarios.

Supplementary Materials

The following are available online at https://www.mdpi.com/article/10.3390/s21206825/s1, Video S1: Saliency maps over a sequence in a traffic scene. The video illustrates the BMS, IttiKoch, GBVS and SR computed over each frame from a sequence of the KITTI object tracking dataset (sequence id: 0006).

Author Contributions

Conceptualization, J.M. and L.A.G.; methodology, J.M. and L.A.G.; software, J.M.; validation, J.M. and L.A.G.; formal analysis, J.M.; investigation, J.M. and L.A.G.; resources, J.M.; data curation, J.M.; writing—original draft preparation, J.M.; writing—review and editing, L.A.G.; visualization, J.M.; supervision, L.A.G.; project administration, J.M. and L.A.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Continental AG.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Acknowledgments

We kindly thank Continental for their great cooperation within PRORETA 5, a joint research project of the University of Bremen, TU Darmstadt, TU Iasi and Continental to investigate future concepts for autonomous driving systems.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

Abbreviations

The following abbreviations are used in this manuscript:
ADASAdvanced Driver Assistance System
ANOVAAnalysis of Variance
BMSBoolean Map Saliency
GBVSGraph-Based Visual Saliency
HADHighly Autonomous Driving
MAEMean Absolute Error
SRSpectral Residual

References

  1. Borji, A.; Itti, L. State-of-the-Art in Visual Attention Modeling. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 185–207. [Google Scholar] [CrossRef] [PubMed]
  2. Zhang, L.; Lin, W. Selective Visual Attention: Computational Models and Applications; John Wiley & Sons: Singapore, 2013. [Google Scholar]
  3. Silva, G.; Schnitman, L.; Oliveira, L. Multi-Scale Spectral Residual Analysis to Speed up Image Object Detection. In Proceedings of the 2012 25th SIBGRAPI Conference on Graphics, Patterns and Images, Ouro Preto, Brazil, 22–25 August 2012. [Google Scholar] [CrossRef]
  4. Duthon, P.; Quinton, J.C.; Colomb, M. Visual saliency on the road: Model and database dependent detection. In Proceedings of the RFIA 2016, Clermont-Ferrand, France, 27 June–1 July 2016. [Google Scholar]
  5. Deng, T.; Yang, K.; Li, Y.; Yan, H. Where Does the Driver Look? Top-Down-Based Saliency Detection in a Traffic Driving Environment. IEEE Trans. Intell. Transp. Syst. 2016, 17, 2051–2062. [Google Scholar] [CrossRef]
  6. Borji, A.; Cheng, M.M.; Jiang, H.; Li, J. Salient Object Detection: A Benchmark. IEEE Trans. Image Process. 2015, 24, 5706–5722. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Bylinskii, Z.; Recasens, A.; Borji, A.; Oliva, A.; Torralba, A.; Durand, F. Where Should Saliency Models Look Next? In Computer Vision –ECCV 2016; Springer International Publishing: Cham, Switzerland, 2016; pp. 809–824. [Google Scholar] [CrossRef]
  8. Alhaija, H.; Mustikovela, S.; Mescheder, L.; Geiger, A.; Rother, C. Augmented Reality Meets Computer Vision: Efficient Data Generation for Urban Driving Scenes. Int. J. Comput. Vis. (IJCV) 2018, 126, 961–972. [Google Scholar] [CrossRef] [Green Version]
  9. Zhang, J.; Sclaroff, S. Exploiting Surroundedness for Saliency Detection: A Boolean Map Approach. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 38, 889–902. [Google Scholar] [CrossRef] [PubMed]
  10. Itti, L.; Koch, C.; Niebur, E. A model of saliency-based visual attention for rapid scene analysis. IEEE Trans. Pattern Anal. Mach. Intell. 1998, 20, 1254–1259. [Google Scholar] [CrossRef] [Green Version]
  11. Harel, J.; Koch, C.; Perona, P. Graph-Based Visual Saliency. In Proceedings of the 19th International Conference on Neural Information Processing Systems, NIPS’06, Cambridge, MA, USA, 4–7 December 2006; MIT Press: Cambridge, MA, USA, 2006; pp. 545–552. [Google Scholar]
  12. Hou, X.; Zhang, L. Saliency Detection: A Spectral Residual Approach. In Proceedings of the 2007 IEEE Conference on Computer Vision and Pattern Recognition, Minneapolis, MN, USA, 17–22 June 2007. [Google Scholar] [CrossRef]
  13. R Core Team. R: A Language and Environment for Statistical Computing; R Foundation for Statistical Computing: Vienna, Austria, 2021. [Google Scholar]
  14. Pohlert, T. The Pairwise Multiple Comparison of Mean Ranks Package (PMCMR). R Package. 2014. Available online: https://mran.microsoft.com/snapshot/2015-03-21/web/packages/PMCMR/vignettes/PMCMR.pdf (accessed on 13 October 2021).
  15. Kummerer, M. Pysaliency. Available online: https://github.com/matthias-k/pysaliency (accessed on 13 October 2021).
  16. Bradski, G. The OpenCV Library. Dr. Dobb’S J. Softw. Tools 2000, 25, 120–123. [Google Scholar]
  17. Perrin, A.F.; Zhang, L.; Le Meur, O. Inferring Visual Biases in UAV Videos from Eye Movements. Drones 2020, 4, 31. [Google Scholar] [CrossRef]
  18. Duda, R.O.; Hart, P.E.; Stork, D.G. Pattern Classification and Scene Analysis; Wiley: New York, NY, USA, 1973; Volume 3. [Google Scholar]
  19. Shannon, C.E. A mathematical theory of communication. Bell Syst. Tech. J. 1948, 27, 379–423. [Google Scholar] [CrossRef] [Green Version]
  20. Duan, H.; Wang, X. Visual Attention Model Based on Statistical Properties of Neuron Responses. Sci. Rep. 2015, 5, 8873. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  21. Zeng, J.; Fu, X.; Leng, L.; Wang, C. Image Inpainting Algorithm Based on Saliency Map and Gray Entropy. Arab. J. Sci. Eng. 2018, 44, 3549–3558. [Google Scholar] [CrossRef]
  22. Borji, A.; Cheng, M.M.; Hou, Q.; Jiang, H.; Li, J. Salient object detection: A survey. Comput. Vis. Media 2019, 5, 117–150. [Google Scholar] [CrossRef] [Green Version]
  23. Bylinskii, Z.; Judd, T.; Oliva, A.; Torralba, A.; Durand, F. What Do Different Evaluation Metrics Tell Us about Saliency Models? IEEE Trans. Pattern Anal. Mach. Intell. 2019, 41, 740–757. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  24. Fisher, R.A. Statistical methods for research workers. In Breakthroughs in Statistics; Springer: New York, NY, USA, 1992; pp. 66–70. [Google Scholar]
  25. Tukey, J.W. Comparing individual means in the analysis of variance. Biometrics 1949, 5, 99–114. [Google Scholar] [CrossRef] [PubMed]
  26. Friedman, M. The use of ranks to avoid the assumption of normality implicit in the analysis of variance. J. Am. Stat. Assoc. 1937, 32, 675–701. [Google Scholar] [CrossRef]
  27. Nemenyi, P.B. Distribution-Free Multiple Comparisons. Ph.D. Thesis, Princeton University, Princeton, NJ, USA, 1963. [Google Scholar]
  28. Le Meur, O. Predicting saliency using two contextual priors: The dominant depth and the horizon line. In Proceedings of the 2011 IEEE International Conference on Multimedia and Expo, Barcelona, Spain, 11–15 July 2011. [Google Scholar] [CrossRef] [Green Version]
  29. Giefer, L.A.; Khamsehashari, R.; Schill, K. Evaluation of Measurement Space Representations of Deep Multi-Modal Object Detection for Extended Object Tracking in Autonomous Driving. In Proceedings of the 2020 IEEE 3rd Connected and Automated Vehicles Symposium (CAVS), Victoria, Canada, 18 November–16 December 2020; pp. 1–6. [Google Scholar]
Figure 1. Example of an image, its semantic segmentation and the corresponding saliency maps obtained with different bottom-up models. The example illustrates how salient areas identified with different models differ on their shape, size, location and saliency level. The models are described in Section 2.
Figure 1. Example of an image, its semantic segmentation and the corresponding saliency maps obtained with different bottom-up models. The example illustrates how salient areas identified with different models differ on their shape, size, location and saliency level. The models are described in Section 2.
Sensors 21 06825 g001
Figure 2. Images and their corresponding saliency maps with the maximum and minimum mean energy. SR produces several salient regions with complex shapes. In contrast, BMS produces blob-like salient areas.
Figure 2. Images and their corresponding saliency maps with the maximum and minimum mean energy. SR produces several salient regions with complex shapes. In contrast, BMS produces blob-like salient areas.
Sensors 21 06825 g002
Figure 3. Images and the corresponding saliency maps with the maximum and minimum entropy. SR produces complex areas of saliency (cf. SR saliency map in Figure 2). In contrast, the higher saliency values produced by the GBVS model appear to be concentrated in a single zone.
Figure 3. Images and the corresponding saliency maps with the maximum and minimum entropy. SR produces complex areas of saliency (cf. SR saliency map in Figure 2). In contrast, the higher saliency values produced by the GBVS model appear to be concentrated in a single zone.
Sensors 21 06825 g003
Figure 4. Example of the threshold analysis for a given segmented image. The top image shows the segmented image and the number n of object instances. Below the corresponding saliency map computed with the GBVS model is displayed. Subsequently, the proportion of salient pixels and the number of object instances in the salient area for different saliency thresholds are shown.
Figure 4. Example of the threshold analysis for a given segmented image. The top image shows the segmented image and the number n of object instances. Below the corresponding saliency map computed with the GBVS model is displayed. Subsequently, the proportion of salient pixels and the number of object instances in the salient area for different saliency thresholds are shown.
Sensors 21 06825 g004
Figure 5. Mean energy and entropy of the saliency maps computed with different models.
Figure 5. Mean energy and entropy of the saliency maps computed with different models.
Sensors 21 06825 g005
Figure 6. MAE between the saliency maps computed with different saliency models and the segments containing traffic participants.
Figure 6. MAE between the saliency maps computed with different saliency models and the segments containing traffic participants.
Sensors 21 06825 g006
Figure 7. (a) Proportion of salient pixels over different saliency thresholds. (b) Detailed view over the scale of each threshold.
Figure 7. (a) Proportion of salient pixels over different saliency thresholds. (b) Detailed view over the scale of each threshold.
Sensors 21 06825 g007
Figure 8. Top 5 best and worst agreements between the segments of traffic participants and the saliency maps computed with the BMS model. Smaller MAE values correspond to a better agreement.
Figure 8. Top 5 best and worst agreements between the segments of traffic participants and the saliency maps computed with the BMS model. Smaller MAE values correspond to a better agreement.
Sensors 21 06825 g008
Figure 9. Top 5 best and worst agreements between the segments of traffic participants and the saliency maps computed with the GBVS model. Smaller MAE values correspond to a better agreement.
Figure 9. Top 5 best and worst agreements between the segments of traffic participants and the saliency maps computed with the GBVS model. Smaller MAE values correspond to a better agreement.
Sensors 21 06825 g009
Figure 10. Top 5 best and worst agreements between the segments of traffic participants and the saliency maps computed with the IttiKoch model. Smaller MAE values correspond to a better agreement.
Figure 10. Top 5 best and worst agreements between the segments of traffic participants and the saliency maps computed with the IttiKoch model. Smaller MAE values correspond to a better agreement.
Sensors 21 06825 g010
Figure 11. Top 5 best and worst agreements between the segments of traffic participants and the saliency maps computed with the SR model. Smaller MAE values correspond to a better agreement.
Figure 11. Top 5 best and worst agreements between the segments of traffic participants and the saliency maps computed with the SR model. Smaller MAE values correspond to a better agreement.
Sensors 21 06825 g011
Table 1. Tukey multiple comparisons of means for energy. 95% family-wise confidence level.
Table 1. Tukey multiple comparisons of means for energy. 95% family-wise confidence level.
Contrastdifflwruprp adj
IttiKoch-BMS0.240.210.27<0.001
SR-BMS0.360.330.39<0.001
GBVS-BMS0.070.040.10<0.001
SR-IttiKoch0.120.090.15<0.001
GBVS-IttiKoch 0.17 0.20 0.14 <0.001
GBVS-SR 0.29 0.32 0.27 <0.001
Table 2. p-Values obtained with a Nemenyi pairwise test for entropy.
Table 2. p-Values obtained with a Nemenyi pairwise test for entropy.
BMSIttiKochSR
IttiKoch0.08
SR<0.001<0.001
GBVS<0.001<0.001<0.001
Table 3. Nemenyi pairwise test for MAE.
Table 3. Nemenyi pairwise test for MAE.
BMSIttiKochSR
IttiKoch<0.001
SR<0.001<0.001
GBVS0.16<0.001<0.001
Table 4. p-Values obtained with a Nemenyi pairwise test for proportion of salient pixels over different saliency thresholds.
Table 4. p-Values obtained with a Nemenyi pairwise test for proportion of salient pixels over different saliency thresholds.
Threshold0.250.50.75
ModelBMSGBVSIttiKochSRBMSGBVSIttiKochSRBMSGBVSIttiKoch
0.25GBVS0.78
IttiKoch<0.001<0.001
SR’<0.001<0.0010.81
0.5BMS<0.001<0.001<0.001<0.001
GBVS<0.001<0.0010.961.00<0.001
IttiKoch<0.001<0.001<0.001<0.0010.79<0.001
SR<0.001<0.001<0.001<0.001<0.001<0.0010.02
0.75BMS<0.001<0.001<0.001<0.001<0.001<0.001<0.001<0.001
GBVS<0.001<0.001<0.001<0.0010.03<0.0010.910.72<0.001
IttiKoch<0.001<0.001<0.001<0.001<0.001<0.001<0.001<0.0011.00<0.001
SR<0.001<0.001<0.001<0.001<0.001<0.001<0.001<0.001<0.001<0.001<0.001
Table 5. Percentage of salient objects with respect to the total number of instances over different saliency thresholds. The column n shows the total number of object instances in the dataset. The saliency model with the largest percentage of salient instances within each threshold is emphasized.
Table 5. Percentage of salient objects with respect to the total number of instances over different saliency thresholds. The column n shows the total number of object instances in the dataset. The saliency model with the largest percentage of salient instances within each threshold is emphasized.
Threshold 0.25Threshold 0.5Threshold 0.75
objectnBMSGBVSIttiKochSRBMSGBVSIttiKochSRBMSGBVSIttiKochSR
car81072.887.712.839.420.468.62.15.82.625.40.40.0
truck10170.395.026.740.623.887.15.02.05.043.60.00.0
person10070.086.011.026.029.053.00.04.06.018.00.00.0
bicycle4353.576.70.016.311.651.20.00.00.011.60.00.0
rider2969.086.23.420.717.269.00.06.90.020.70.00.0
bus1952.6100.010.542.126.363.20.05.310.531.65.30.0
train1872.288.922.216.711.172.25.65.60.016.70.00.0
motorcycle887.587.50.062.525.075.00.012.50.037.50.00.0
caravan7100.0100.028.657.128.6100.00.014.314.342.90.00.0
trailer5100.060.020.040.00.040.00.00.00.020.00.00.0
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Maldonado, J.; Giefer, L.A. A Comparison of Bottom-Up Models for Spatial Saliency Predictions in Autonomous Driving. Sensors 2021, 21, 6825. https://doi.org/10.3390/s21206825

AMA Style

Maldonado J, Giefer LA. A Comparison of Bottom-Up Models for Spatial Saliency Predictions in Autonomous Driving. Sensors. 2021; 21(20):6825. https://doi.org/10.3390/s21206825

Chicago/Turabian Style

Maldonado, Jaime, and Lino Antoni Giefer. 2021. "A Comparison of Bottom-Up Models for Spatial Saliency Predictions in Autonomous Driving" Sensors 21, no. 20: 6825. https://doi.org/10.3390/s21206825

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop