Next Article in Journal
A Blockchain-Based Secure Multi-Party Computation Scheme with Multi-Key Fully Homomorphic Proxy Re-Encryption
Next Article in Special Issue
Deep Learning-Based Multiple Droplet Contamination Detector for Vision Systems Using a You Only Look Once Algorithm
Previous Article in Journal
CBIR-DSS: Business Decision Oriented Content-Based Recommendation Model for E-Commerce
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Why Does the Automation Say One Thing but Does Something Else? Effect of the Feedback Consistency and the Timing of Error on Trust in Automated Driving

1
VEDECOM Institute, 78000 Versailles, France
2
Laboratoire d’Etude des Mécanismes Cognitifs (EA 3082), University Lyon 2, 69007 Lyon, France
3
Institut de Démographie de l’Université Paris 1 (IDUP), 75013 Paris, France
4
Institut Universitaire de France, 75005 Paris, France
*
Author to whom correspondence should be addressed.
Information 2022, 13(10), 480; https://doi.org/10.3390/info13100480
Submission received: 7 July 2022 / Revised: 23 September 2022 / Accepted: 4 October 2022 / Published: 6 October 2022

Abstract

:
Driving automation deeply modifies the role of the human operator behind the steering wheel. Trust is required for drivers to engage in such automation, and this trust also seems to be a determinant of drivers’ behaviors during automated drives. On the one hand, first experiences with automation, either positive or not, are essential for drivers to calibrate their level of trust. On the other hand, an automation that provides feedback about its own level of capability to handle a specific driving situation may also help drivers to calibrate their level of trust. The reported experiment was undertaken to examine how the combination of these two effects will impact the driver trust calibration process. Four groups of drivers were randomly created. Each experienced either an early (i.e., directly after the beginning of the drive) or a late (i.e., directly before the end of it) critical situation that was poorly handled by the automation. In addition, they experienced either a consistent continuous feedback (i.e., that always correctly informed them about the situation), or an inconsistent one (i.e., that sometimes indicated dangers when there were none) during an automated drive in a driving simulator. Results showed the early- and poorly-handled critical situation had an enduring negative effect on drivers’ trust development compared to drivers who did not experience it. While being correctly understood, inconsistent feedback did not have an effect on trust during properly managed situations. These results suggest that the performance of the automation has the most severe influence on trust, and the automation’s feedback does not necessarily have the ability to influence drivers’ trust calibration during automated driving.

1. Introduction

Driving automation is ongoing and the higher levels of automated driving (level 4 and 5, [1]) should allow drivers to reduce the cognitive load related to driving and liberate time to engage in other activities. However, these benefits imply that drivers would be willing to delegate the driving task to a Highly Automated Driving (HAD) system that will be able to perform the dynamic driving task instead of them in specific environments, corresponding to level 4 [1]. Going from a fully manual task to an increasingly more automated one requires the drivers to trust automation [2]. This specific type of trust (i.e., Trust in Automation, TiA) has been shown to strongly influence and predict automation use and performance [3,4,5,6]. TiA is a central concept when one considers human–automation interaction from the human perspective. Indeed, TiA is a psychological construct that encapsulates human attitudes toward automation and directly influences reliance on automation.
Theoretical models have been developed to better understand TiA. A previous literature review exploring suitable models [7] led to the selection of one factor-centered model, the structure of which served as a basis to study trust-influencing factors. Hoff and Bashir [8] proposed a model where trust is decomposed into three layers and described the factors influencing each of these trust layers. The first layer, dispositional trust, is linked to the human operator and evolves slowly during one’s lifetime. The second layer, situational trust, is mainly linked to the context of interaction between the individual and the automation. It changes from one situation to another. The last layer, learned trust, is composed of two parts. The initial learned trust refers to the previous knowledge and expectations the operator has about the considered automation. It changes after each interaction when the new experience is processed by the operator. Finally, the dynamic learned trust depends on the automation’s performance and features, and it evolves dynamically during the interaction. In the driving context, Trust in Automated Driving (TiAD) strongly relies on TiA and seems to play a major role in HAD acceptance, adoption [9,10], and proper use [11]. It therefore appears crucial to study the implications of this TiAD and the aspects that influence its dynamic calibration during HAD use. Here, the focus was set on dynamic learned trust to investigate how TiAD evolves when drivers actually interact with the HAD system.
Previous experiments showed drivers’ TiAD was negatively influenced by critical situations (i.e., a driving situation that requires the automation to perform an urgent maneuver in order to preserve passengers’ safety) occurring during drives [12,13,14]. Hoff and Bashir [8] reported the timing of error, which is related to the automation performance, had a major influence on the operator’s trust development. Critical situations that occur early during the interaction with automation have a greater negative influence than errors occurring after a positive experience with the automation [15,16]. In the driving context, errors may encompass a variety of automation malfunctions, such as unexpected Take-Over Requests (TORs) that decrease TiAD [12] or poorly managed situations [13].
In addition, studies have shown drivers sometimes had trouble to understand why the system managed the situations poorly [13]. Hoff and Bashir [8] also reported the transparency of the feedback provided by the automation was an influential factor related to its design features. Transparent feedback has an indirect positive influence on TiA, while less transparent feedback may confuse operators.
The transparency of the feedback can be described as the ability of individuals to correctly understand and evaluate the inner processes that lead an automation to behave in a certain manner [17]. Feedback that allows drivers to better estimate the HAD performance and malfunctions seems to promote a well-calibrated TiAD [12]. Moreover, continuous feedback indicating HAD uncertainty in various situations (e.g., in foggy weather) also supports TiAD calibration [18,19]. Understanding the ongoing performance of HAD therefore seems crucial for drivers to have trust in it.
Moreover, drivers’ initial learned level of trust seemed to play a role in their further perception of critical situations during drives. Trustful drivers trusted the HAD system more during similar critical situations, compared to distrustful drivers [13,14]. In these previous experiments, drivers were recruited depending on their either high or low initial learned level of TiAD. This methodology allowed us to obtain two homogeneous groups and display strong results on drivers’ TiAD evolution patterns, depending on situations encountered under HAD. Distrustful participants’ level of trust increased more during the first minutes of interaction, compared with trustful participants. Moreover, distrustful participants’ level of trust varied with a greater amplitude depending on the encountered critical situations. Participants’ level of trust also impacted their visual behaviors during drives: people with a higher level of trust monitored the driving environment less, while people with a lower level of trust monitored it more.
The present experiment aims to evaluate both the influence of perceived understanding on TiAD and the dynamic evolution of TiAD depending on the timing of a critical situation poorly managed by the HAD system. These two effects are examined depending on drivers’ initial learned level of trust, in a driving simulator, for the driver’s very first experience with such automation.
The timing of the critical situation is expected to influence the dynamics of drivers’ trust evolution. Drivers who experience an early critical situation are expected to have a lower level of trust in the HAD system, and a slower trust increase during the rest of the drive, with more gazes directed toward the driving environment (H1). Inconsistent feedback is expected to slow the TiAD increase and the perceived understanding of the HAD system and increase the number of gazes directed toward the driving environment and the automation (H2). Finally, drivers who have a low initial learned level of TiAD are expected to have a lower level of trust during the drive (H3). Low initial learned level of TiAD was also expected to have more impact with an early critical situation (H4) or an inconsistent feedback (H5) compared with drivers who have a high initial learned level of TiAD.

2. Materials and Methods

2.1. Participants

Sixty-one drivers participated in this experiment (29 females, M = 41.97 years old, SD = 11.25, min = 21, max = 64). They all had normal or corrected-to-normal vision and audition, to ensure they would correctly perceive the automation feedback (because of the feedback nature, people with color vision deficiency were not included in this study). Participants had all carried a valid driver’s license for a minimum of three years (M = 22.3, SD = 11.4, min = 3, max = 46). They reported that they drove regularly: 37 (60.7%) drove every day, 18 (29.5) at least once a week, five (8.2%) at least once a month, and one (1.6%) drove less frequently. Drivers’ self-evaluated level of experience with advanced automated driving assistance systems was inquired. Eleven (18.0%) declared they had advanced knowledge of AD, 20 (32.8%) had intermediate knowledge, 21 (34.4%) basic knowledge, and nine (14.8%) declared they had no knowledge at all regarding HAD. Drivers also completed a trust scale form [13,14] to measure their initial learned level of TiAD.

2.2. Apparatus

The study was conducted in a static driving simulator equipped with three 2520 × 1440 mm panels, providing a 200° horizontal field of view. Three 7″ 16:9 LCD screens displayed a rear view of the driving environment, in place of real car mirrors. A 10″ 16:9 LCD screen was used to display the dashboard. A 10.1″ touchscreen was placed to the right of the steering wheel as the Human–Machine Interface (HMI). On the left sideband of the HMI, different pictograms were displayed to indicate the state of the HAD system and any relevant vehicle maneuvers, duplicating the information from the dashboard. The rest of the HMI displayed a tablet with an Android™ emulator with video games and Internet access (Figure 1). A colored RGB LED strip was positioned all around the edges of the cockpit and around the HMI to ensure the colored information would always be present in the drivers’ field of view. The color of the LED strip varied depending on the state of the HAD system (Figure 2). The RGB LED strip was 4.1 m long and was composed of approximately 60 LEDs per meter, at an intensity of four to six lumens. The used RGB colors were white (RGB (255, 255, 255) for manual driving), blue (RGB (0, 0, 255) for available HAD), green (RGB (0, 255, 0) for engaged HAD in normal conditions), orange (RGB (255, 150, 0) when the HAD detected a potential problem, but was able to manage it), and red (RGB (255, 0, 0) for TORs). The SCANeR™ Studio 1.8 software (AV Simulation, France, https://www.avsimulation.fr, accessed on 6 July 2022), was used to run the driving simulation. The eye-tracking system was composed of four 60 Hz cameras recording data through the Smart-Eye Pro 6.2 software.

2.3. Procedure

Participants arrived at the VEDECOM Institute, signed a consent form, and completed a trust scale form [13] to evaluate their initial learned level of trust (Table 1, Qi in Figure 3). The driving simulator was explained to ensure all participants correctly understood what each screen was displaying. Instructions were then given to participants. They were told they would experience a 35 min highly automated drive on the highway. The pictograms and colors indicating the several HAD states were explained, then a five-minute training was completed to allow drivers to familiarize themselves with the simulator and the HMI. Participants were encouraged to behave in the same way they would in a real vehicle. They were not required to do any particular task, and they were free to engage in any non-driving-related activity they felt appropriate.
Participants were then randomly assigned to one of the four combinations of the two between-subject variables: the timing of the critical situation (early vs. late) and the feedback consistency (consistent vs. inconsistent), which will be detailed later. They were not aware of this.
Participants then started the experiment. They answered a first single trust item that served as a baseline after their training experience. One minute after the beginning of the drive, drivers who were assigned to the early critical situation condition experienced a missing lane marking situation for 350 m. This event was poorly managed by the vehicle which zigzagged on the road with a maximum lateral lane shift of 1.2 m, until lane markings were found again (see Video S1: Critical situation (zigzag)). Drivers who were assigned to the other condition experienced a normal, monotonous highway driving during this time. After that, all drivers experienced three correctly managed situations, each spaced by several minutes of monotonous highway driving. First, a roadworks area forced the ego vehicle to decelerate (from 128 km/h to 90 km/h in five seconds) and change lanes to avoid this area (see Video S2: Critical situation (roadwork area)). Then, a slow truck forced the ego vehicle to decelerate (from 128 km/h to 80 km/h in four seconds) because the left lane was congested, until the truck had exited the highway. Finally, a deer crossing the road forced the ego vehicle to decelerate (from 128 km/h to 80 km/h in five seconds). Each of these situations was replicated from a previous experience that gave insights regarding situations that have a positive or a negative effect on drivers’ level of TiAD [14]. Nine minutes after the crossing deer situation, drivers who were assigned to the late critical situation condition experienced the portion of the road with no lane marking, while the other drivers had a monotonous drive time. Last, all drivers experienced an unplanned TOR, indicated by a red color of the LEDs, an urgent sound, and a relevant pictogram.
During the drive, participants assigned to the inconsistent feedback condition experienced four false alerts during monotonous driving and two missing events during the last two correctly managed situations (the slow truck and the crossing deer situations). During the false alerts, the LED became orange, and a sound and a pictogram indicated that the automation had detected a situation on the road while there was nothing particular at these points. During the missing events, the feedback failed to detect the situation (i.e., green LED color instead of orange). All these situations were correctly detected for the consistent feedback group.
After each situation (both critical situations, CS in Figure 3, and correctly managed situations, S1, S2, and S3 in Figure 3) and after each moment of feedback conditions (both consistent and inconsistent; see ◊ in Figure 3), a single trust item inquiring drivers’ current level of TiAD was displayed on the HMI (i.e., “How much do you trust the automated driving system?”). Another item inquiring drivers’ perceived understanding of the HAD system behavior was also displayed (i.e., “How much did you understand the automated driving system reactions”). After the end of the drive, participants complated a 10-item trust scale form (the same that was used to assess their initial learned level of TiAD) to evaluate their final level of TiAD. They were finally debriefed, thanked, and rewarded.

2.4. Data Analysis

Dependent variables included subjective trust, perceived understanding of the HAD system, and visual strategies.
Single trust items and perceived understanding items, both ranging from zero to 100 were used. The 10-item trust scale was scored according to [14], and participants were divided into two groups (i.e., referred as Trustful and Distrustful as compared to the rest of the sample) based on a median-split.
Visual strategies data, or time spent looking towards specific Areas of Interest (AOI) were processed in percentages according to ISO 15007 recommendations (International Organization for Standardization, 2020). AOI were defined as “Road”, “Rear Mirrors”, “Landscape”, “Dashboard”, “HMI”, “Android tablet”, “Phone”, and “Other” (for all glances directed somewhere else than the previously established AOI). “Road” and “Rear Mirrors” were grouped in the category “Road”, “Dashboard” and “HMI” were grouped in the category “Dashboard”, and “Android tablet” and “Phone” were grouped in the category “Non-Driving-Related Activities” for analysis.
Data were processed using R [20] and ggplot2 [21]. ANOVAs were run using the Greenhouse–Geisser correction.

2.5. Design

The following analyses were performed using a mixed experimental design. Pre-examination-confirmed data followed a normal distribution, without outliers. Between-subject variables were the initial learned level of trust: Trustful vs. Distrustful, the feedback consistency: Consistent vs. Inconsistent, and the timing of the critical situation: Early vs. Late. Within-subject variables were the color of the feedback: Green, Orange, or Red, and Time, which took different values depending on the analyses. In the ANOVAs analyses, Time was ten measurement points. In the linear models, Time was each passing minute from the start to the end of the experiment.

3. Results

3.1. Trust and Perceived Understanding Questionnaires

The median-split on the drivers’ initial learned level of trust score resulted in two groups, which were subdivided into eight groups depending on the randomly assigned independent variables (i.e., feedback consistency and timing of error). Groups specificities are given in Table 2. Table 3 presents Bartlett’s t-test that shows no evidence to claim that the groups’ variances are not equal.
A repeated measure ANOVA on single trust items showed the initial learned level of trust had an impact on the further level of trust during the drive, F(1, 59) = 13.5, p < 0.001, ηp2 = 0.187. Trustful drivers declared a higher trust in the automated driving system during the whole drive. Another repeated measure ANOVA was then run on the feedback consistency and the timing of the critical situation (Figure 4). The results showed a main effect of the Time, F(4.86, 273.7) = 21.3, p < 0.001, ηp2 = 0.272, and an interaction between the Time and the timing of the critical situation, F(4.86, 273.7) = 2.27, p < 0.05, ηp2 = 0.038. No other effects were significant. This result indicates drivers’ trust varied during the drive, and this variation was different depending on whether drivers experienced an early or a late critical situation. Trust increased during the drive among drivers who experienced a late critical situation compared with drivers who experienced an early one. Trust dropped during the critical situation (regardless of if it was early or late) and during the TOR, compared to the other trust measurements.
Finally, a repeated measure ANOVA was run with trust scores and perceived understanding scores to investigate whether the feedback consistency was correctly perceived by participants. It again exhibited a main effect of the Time, F(4.98, 598.0) = 27.0, p < 0.001, ηp2 = 0.184, and an interaction between the Time and the type of measurement (i.e., trust vs. perceived understanding), F(4.98, 598.0) = 3.70, p < 0.01, ηp2 = 0.030. This result indicates the two measures evolved differently over time. Trust increased during the drive while perceived understanding was lower in the case of inconsistent feedback, among drivers who were assigned to the inconsistent feedback condition (Figure 5).
Two linear models investigating, respectively, trust and perceived understanding were constructed to explore the effects of initial learned level of trust, age, gender, and Time (in minutes) during the drive. They also explored the several combinations of timing of the critical situation and feedback consistency. Moreover, the color of the feedback (green, orange, or red) at each measurement point was tested, alongside the timing of the critical situation.
The results (Table 4) indicated the initial learned level of trust had a positive influence on further trust evolution and perceived understanding of the HAD system. Each initial point of trust (on a scale from one to 100) had a strong probability (p < 0.001) to increase the further trust measurements of 0.46 points, and the other perceived understanding measurements of 0.40 points. Time also increased the trust measurements by 0.33 points for each passing minute but did not seem to influence perceived understanding. Age did not seem to have any effect at all. Finally, all things being equal (i.e., neutralizing the effects of the initial learned level of trust, of the time, and of the age), males seemed to have a lower level of trust and lower self-assessed level of understanding of the automation than females.
The results then confirmed the effect of the timing of the critical situation: drivers who had consistent feedback rated 7.65 points more at each trust measurement when they experienced a late critical situation, compared to the group that experienced an early critical situation. This result was not found for drivers who had inconsistent feedback.
Finally, the model indicated an effect of the feedback consistency and the color of the feedback on drivers’ perceived understanding of the HAD. Inconsistent orange feedback during monotonous automated driving had a strong negative impact on perceived understanding, regardless of the timing of the critical situation. Drivers who experienced an early critical situation rated their perceived understanding 16.47 points lower than the reference situation (i.e., when the feedback was green and during monotonous automated driving). Drivers who experienced a late critical situation rated their perceived understanding 24.92 points lower than this same reference situation, suggesting that the last impression they had from the HAD was the most impactful. Nevertheless, no such effects were found on trust measurements. This result supports the idea that the mismatch between the HAD feedback and the situations was correctly perceived by drivers who experienced inconsistent feedback, but it had no main effect on their level of trust. Contrarily, the orange feedback during the critical situations and the red feedback during the TOR had a negative influence on both trust and perceived understanding, regardless of the timing of the critical situation.

3.2. Visual Behaviour

Drivers’ visual behavior was then investigated during the several types of situations of the drive (i.e., HAD, correctly managed situations, critical situations, and TOR) and depending on the color of the feedback (Figure 6). Chi-squared tests were run to test whether the gaze frequency towards the several AOI were different during these combinations of colors and situations. The Chi-squared tests revealed that drivers’ gaze frequency during HAD and when the feedback was green (Green/HAD) was different from all the other combinations of feedback color and situation (p < 0.001, Table 5). A difference was also observed when feedback was orange during correctly managed situations (Orange/Correct Situations) compared to the orange feedback while everything was normal (p < 0.05), the orange feedback during the critical situation (p < 0.05), and finally the red feedback during the TOR (p < 0.05, Table 5). These results confirm drivers correctly perceived feedback and monitored the driving environment accordingly. Drivers disengaged from on-board activities and increased their road monitoring when feedback indicated a potential problem.

4. Discussion

The present experiment aimed to investigate (a) the effect of the timing of a critical situation on trust construction, and (b) the effect of the feedback consistency on this trust evolution, depending on the drivers’ initial learned level of trust.

4.1. Effect of the Timing of a Critical Situation

The timing of the critical situation had an influence on further trust evolution among drivers. Participants who experienced an early critical situation presented a trust increase at the next trust measurement, six minutes after that critical situation but no further increase after subsequent correctly managed situations. In comparison, drivers who experienced a late critical situation presented a trust progression during the previous correctly managed situations, confirming H1 and suggesting early critical situations may be a determinant in the drivers’ further trust calibration process. This result is consistent with previous work that investigated other types of automations in the general TiA theory [15,16], and with previous results focused on TiAD [14]. Another experiment that showed a malfunction happening in the middle of the drive also led to a strong decrease in TiAD, which was recovered during the following minutes at the same level as before the malfunction [12]. In this last case, the malfunction did not prevent a further trust increase. It therefore seems malfunctions or critical situations decrease TiAD in the moment, but do not have a remaining negative effect when drivers were given the possibility to see the HAD system perform well before the malfunction. When drivers do not have the time to experience a properly functioning HAD system before the malfunction, TiAD progression seems affected for the rest of the experiment.
Otherwise, debriefings revealed most people did not understand why the ego vehicle zigzagged during the critical situation. When they were told about the missing lane marking, they were surprised to learn that HAD systems need this information to navigate properly. It therefore seems crucial for future study to inquire what drivers need to know about HAD functioning in order to ensure a proper fundamental knowledge of these technologies is spread in the drivers’ population.

4.2. Effect of the Feedback Consistency

Previous studies showed that a HAD system displaying a level of uncertainty regarding its own capabilities may help drivers to calibrate their level of trust, particularly in situations where the HAD system is less reliable [18,19]. Moreover, other studies have suggested that using anthropomorphized feedback [22,23,24], or punctual messages explaining the reasons behind a specific HAD system behavior [25] may help drivers to calibrate their level of trust.
In the present study, drivers who experienced a consistent feedback had a stable trust during the drive. Comparatively, drivers who had an inconsistent feedback seemed to have seen a light increase in trust. Their first experience with the inconsistent feedback induced a lower level of trust that subsequently increased during the drive, although they indeed had trouble to understand this inconsistent feedback. This result invalidates H2 and suggests drivers’ trust is not particularly sensitive to the feedback consistency, at least under the conditions of the reported experiment. This finding is at odds with existing literature on feedback importance [26,27]. Further investigations carried out on a larger sample of participants would be required to confirm this result. Such investigations would benefit from a detailed analysis of the tasks and visual behaviors engaged by people behind the wheel, including under real life automated driving conditions. In addition, debriefings showed that drivers frequently report they understood the HAD feedback (e.g., because there was a situation they did not perceive), even when there was nothing to understand. It is possible that drivers rationalize the HAD feedback a posteriori. In short, trust increases when drivers are given the opportunity to see the automation performs well, even when its features are not fully understood. An inconsistent feedback does not seem to have a negative influence on trust, and the performance of the automation appears to have the ultimate ability to influence trust. Future studies may also combine continuous and punctual feedback to fill the gap between the previous studies’ results and the present ones.
Otherwise, a priori information related to an HAD system’s reliability has been shown to influence the development of TiAD, either positively or negatively, before any interaction and during a video-based experimentation where participants experienced the same videos [28]. In the present study, participants had no particular a priori information, but experienced different types of feedback consistency that may have influenced the level of HAD reliability perceived by drivers. This consistency of the feedback did not seem to have a positive effect on TiAD, suggesting drivers’ mental model and expectations may have a stronger influence on trust than actual HAD’s feedback reliability. Of course, this conclusion is not irrevocable. Even if the colored lights used were able to attract visual attention efficiently and the color coding was very easy to interpret, a large variety of design options are conceivable and other designs may result in different effects.
Moreover, studies reported that missed detections from a lane departure warning system did not lead to significantly poorer manual driving performance [29,30]. In the present study, the vehicle’s missed detections (i.e., green during correctly managed situations) are also likely to have a small effect on drivers’ perceptions, because these situations were well-handled by the HAD system. On the other hand, a false positive (i.e., orange during monotonous automated driving) seems to disturb participants’ TiAD at the first occurrence, but not after they saw the HAD system correctly detect the further situations on-road. This suggests that the drivers’ “buffer” regarding false warning is quickly reset and does not load more discrepancies while the HAD system performs well. Additionally, previous studies showed these missed detections and false positives may have cumulative effects in case of several successive occurrences [29,30]. Future studies may explore if these cumulative effects also exist for TiAD. Nevertheless, it seems both ADAS and HAD do not need be entirely reliable for drivers to trust them.

4.3. Effect of the Initial Learned Level of Trust

The initial learned level of trust seemed to have an enduring effect on drivers’ level of trust during their first experiences with the HAD system, confirming H3, as reported in previous studies for drivers [13,14] and for passengers [31]. Nevertheless, no interaction between the initial learned level of trust and the time was found, contrasting with these previous results. The method that was used to obtain the two groups (i.e., Trustful and Distrustful) was different and may have led to smaller inter-individual differences between groups. In addition, the provided feedback may have had a stabilizing effect on drivers’ TiAD, which would explain such results. Alternatively, the critical situation and the unexpected TOR may have participated to reduce the trust increase among distrustful drivers. Future studies are therefore needed in order to better understand the relation between the initial learned level of trust and other HAD-related specificities, such as the feedback consistency and the timing of critical situations, that may have interacted with it. Finally, no interactions between the initial learned level of trust and the timing of error, or the feedback consistency, were found, invalidating H4 and H5. This information may indicate the level of initial learned trust has a stronger effect than HAD’s features and design (related to the dynamic learned trust, [8]) during drivers’ first interaction with such automations.

4.4. Limitations and Perspectives

The proposed experimental design regarding the early or late critical situation allowed us to investigate the trust development process during a 35-min drive. Nevertheless, the late critical situation was still relatively early, considering the driver’s total experience time with a vehicle during its lifetime. Future studies may investigate TiAD construction and evolution over longer periods of time, with a first critical situation appearing after several positive interactions with an HAD system. Next, participants in the inconsistent feedback condition experienced several false positives between the situations that were correctly managed by the HAD system. Future studies may extend the present results by testing the dynamic evolution of TiAD when drivers are confronted with a succession of several false warnings without any other correct situations. Last, the current experiment used a median-split method to create the two groups based on the initial learned level of trust (i.e., Trustful and Distrustful). This approach provided a better vision of the general level of trust in the drivers’ population compared to other grouping method based on pre-defined thresholds [13]). Nevertheless, the two groups were less representative of strongly trustful or strongly distrustful people, and the reported effects therefore must be carefully considered.

5. Conclusions

The present study suggests the provided automation’s feedback has a much lower influence than the actual HAD performance, when considering drivers’ trust formation process during HAD. Hoff and Bashir’s [8] factor linked with the automation performance (i.e., the timing of malfunction) is more influential than the factor linked with the design features (i.e., the feedback consistency), confirming TiAD is mainly dependent of automation’s ongoing performance. Nevertheless, it seems the initial learned trust is still preponderant during the first interaction with an HAD system, compared with factors related to the dynamic trust. Other studies investigating the trust formation and calibration process over longer durations (with, i.e., longitudinal studies on cohorts of drivers), could bring more information about the moment of transition when the dynamic learned trust gains more influence than the initial learned trust. Moreover, this information must be carefully considered when attempting to guide the drivers’ trust calibration process. Modifying the vehicle’s behavior in specific situations seems to be a more effective way to encourage drivers to recalibrate their trust than using feedback.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/info13100480/s1, Video S1: Critical situation (zigzag); Video S2: Critical situation (roadwork area).

Author Contributions

Conceptualization, J.B.M., M.B. and J.N.; methodology, J.B.M., M.B., J.N.; software, J.B.M.; validation, J.B.M., M.B., J.N. and R.B.; formal analysis, J.B.M. and R.B.; investigation, J.B.M.; resources, M.B. and J.N.; data curation, J.B.M.; writing—original draft preparation, J.B.M.; writing—review and editing, J.B.M., M.B. and J.N.; visualization, J.B.M. and R.B.; supervision, M.B. and J.N.; funding acquisition, M.B. and J.N. All authors have read and agreed to the published version of the manuscript.

Funding

This study was funded by the VEDECOM Institute.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki and approved by the Ethics Committee of VEDECOM Institute (2020).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Society of Automotive Engineers (SAE) International. Taxonomy and Definitions for Terms Related to Driving Automation Systems for On-Road Motor Vehicles; SAE International: Warrendale, PA, USA, 2021. [Google Scholar] [CrossRef]
  2. Navarro, J. A state of science on highly automated driving. Theor. Issues Ergon. Sci. 2019, 20, 366–396. [Google Scholar] [CrossRef]
  3. Dzindolet, M.T.; Peterson, S.A.; Pomranky, R.A.; Pierce, L.G.; Beck, H.P. The role of trust in automation reliance. Int. J. Hum. Comput. Stud. 2003, 58, 697–718. [Google Scholar] [CrossRef]
  4. Lee, J.; Moray, N. Trust, control strategies and allocation of function in human-machine systems. Ergonomics 1992, 35, 1243–1270. [Google Scholar] [CrossRef] [PubMed]
  5. Lee, J.D.; Moray, N. Trust, self-confidence, and operators’ adaptation to automation. Int. J. Hum. Comput. Stud. 1994, 40, 153–184. [Google Scholar] [CrossRef] [Green Version]
  6. Lee, J.D.; See, K.A. Trust in automation: Designing for appropriate reliance. Hum. Factors 2004, 46, 50–80. [Google Scholar] [CrossRef] [Green Version]
  7. Manchon, J.B.; Bueno, M.; Navarro, J. From manual to automated driving: How does trust evolve? Theor. Issues Ergon. Sci. 2020, 22, 528–554. [Google Scholar] [CrossRef]
  8. Hoff, K.A.; Bashir, M. Trust in automation: Integrating empirical evidence on factors that influence trust. Hum. Factors 2015, 57, 407–434. [Google Scholar] [CrossRef]
  9. Liu, P.; Yang, R.; Xu, Z. Public acceptance of fully automated driving: Effects of social trust and risk/benefit perceptions: Public acceptance of fully automated driving. Risk Anal. 2019, 39, 326–341. [Google Scholar] [CrossRef]
  10. Molnar, L.J.; Ryan, L.H.; Pradhan, A.K.; Eby, D.W.; Louis, R.M.S.; Zakrajsek, J.S. Understanding trust and acceptance of automated vehicles: An exploratory simulator study of transfer of control between automated and manual driving. Transp. Res. Part F Traffic Psychol. Behav. 2018, 58, 319–328. [Google Scholar] [CrossRef]
  11. Kundinger, T.; Wintersberger, P.; Riener, A. (Over)Trust in Automated Driving: The sleeping Pill of Tomorrow? In Proceedings of the Extended Abstracts of the 2019 CHI Conference on Human Factors in Computing Systems, Glasgow, Scotland, UK, 4–9 May 2019; pp. 1–6. [Google Scholar] [CrossRef]
  12. Kraus, J.; Scholz, D.; Stiegemeier, D.; Baumann, M. The more you know: Trust dynamics and calibration in highly automated driving and the effects of take-overs, system malfunction, and system transparency. Hum. Factors 2020, 62, 718–736. [Google Scholar] [CrossRef]
  13. Manchon, J.B.; Bueno, M.; Navarro, J. Calibration of Trust in Automated Driving: A matter of initial level of trust and automated driving style? Hum. Factors 2021. [Google Scholar] [CrossRef] [PubMed]
  14. Manchon, J.B.; Bueno, M.; Navarro, J. How the Initial Level of Trust in Automated Driving Impacts Drivers’ Behaviour and Early Trust Construction. Transp. Res. Part F Traffic Psychol. Behav. 2022, 86, 281–291. [Google Scholar] [CrossRef]
  15. Manzey, D.; Reichenbach, J.; Onnasch, L. Human Performance Consequences of Automated Decision Aids: The Impact of Degree of Automation and System Experience. J. Cogn. Eng. Decis. Mak. 2012, 6, 57–87. [Google Scholar] [CrossRef] [Green Version]
  16. Sanchez, J. Factors That Affect Trust and Reliance on an Automated Aid. Ph.D. Thesis, Georgia Institute of Technology, Atlanta, GA, USA, 2006. Available online: https://core.ac.uk/display/4682674 (accessed on 2 March 2022).
  17. Seong, Y.; Bisantz, A.M. The impact of cognitive feedback on judgment performance and trust with decision aids. Int. J. Ind. Ergon. 2008, 38, 608–625. [Google Scholar] [CrossRef]
  18. Beller, J.; Heesen, M.; Vollrath, M. Improving the driver–automation interaction: An approach using automation uncertainty. Hum. Factors 2013, 55, 1130–1141. [Google Scholar] [CrossRef]
  19. Helldin, T.; Falkman, G.; Riveiro, M.; Davidsson, S. Presenting system uncertainty in automotive UIs for supporting trust calibration in autonomous driving. In Proceedings of the 5th International Conference on Automotive User Interfaces and Interactive Vehicular Application—AutomotiveUI ’13, Eindhoven, The Netherlands, 27–30 October 2013; pp. 210–217. [Google Scholar] [CrossRef]
  20. R Core Team. R: A Language and Environment for Statistical Computing; R Foundation for Statistical Computing: Vienna, Austria, 2017; Available online: https://www.R-project.org/ (accessed on 12 January 2022).
  21. Wickham, H. Ggplot2: Elegant Graphics for Data Analysis; Springer: New York, NY, USA, 2016; Available online: http://ggplot2.tidyverse.org (accessed on 12 January 2022).
  22. Forster, Y.; Naujoks, F.; Neukum, A. Increasing Anthropomorphism and Trust in Automated Driving Functions by Adding Speech Output. In Proceedings of the 2017 IEEE Intelligent Vehicles Symposium (IV), Los Angeles, CA, USA, 11–14 June 2017; pp. 365–372. [Google Scholar] [CrossRef]
  23. Häuslschmid, R.; von Buelow, M.; Pfleging, B.; Butz, A. Supporting Trust in Autonomous Driving. In Proceedings of the 22nd International Conference on Intelligent User Interfaces—IUI ’17, Limassol, Cyprus, 13–16 March 2017; pp. 319–329. [Google Scholar] [CrossRef]
  24. Niu, D.; Terken, J.; Eggen, B. Anthropomorphizing information to enhance trust in autonomous vehicles. Hum. Factors Ergon. Manuf. Serv. Ind. 2018, 28, 352–359. [Google Scholar] [CrossRef]
  25. Koo, J.; Kwac, J.; Ju, W.; Steinert, M.; Leifer, L.; Nass, C. Why did my car just do that? Explaining semi-autonomous driving actions to improve driver understanding, trust, and performance. Int. J. Interact. Des. Manuf. IJIDeM 2015, 9, 269–275. [Google Scholar] [CrossRef]
  26. Norman, D.A.; Broadbent, D.E.; Baddeley, A.D.; Reason, J. The ‘problem’ with automation: Inappropriate feedback and interaction, not ‘over-automation’. Philos. Trans. R. Soc. London. B Biol. Sci. 1990, 327, 585–593. [Google Scholar] [CrossRef]
  27. Wagenaar, W.A.; Keren, G.B. Does The Expert Know? The Reliability of Predictions and Confidence Ratings of Experts. In Intelligent Decision Support in Process Environments; Springer: Berlin/Heidelberg, Germany, 1986; pp. 87–103. [Google Scholar] [CrossRef]
  28. Kraus, J.M.; Forster, Y.; Hergeth, S.; Baumann, M. Two routes to trust calibration: Effects of reliability and brand information on Trust in Automation. Int. J. Mob. Hum. Comput. Interact. 2019, 11, 17. [Google Scholar] [CrossRef]
  29. Navarro, J.; Deniel, J.; Yousfi, E.; Jallais, C.; Bueno, M.; Fort, A. Does False and Missed Lane Departure Warnings Impact Driving Performances Differently? Int. J. Hum. Comput. Interact. 2019, 35, 1292–1302. [Google Scholar] [CrossRef]
  30. Navarro, J.; Deniel, J.; Yousfi, E.; Jallais, C.; Bueno, M.; Fort, A. Influence of lane departure warnings onset and reliability on car drivers’ behaviors. Appl. Ergon. 2017, 59, 123–131. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  31. Hartwich, F.; Hollander, C.; Johannmeyer, D.; Krems, J.F. Improving passenger experience and Trust in Automated Vehicles through user-adaptive HMIs: “The more the better” does not apply to everyone. Front. Hum. Dyn. 2021, 3, 669030. [Google Scholar] [CrossRef]
Figure 1. Close-up view of the apparatus (here, the LED stripe is green).
Figure 1. Close-up view of the apparatus (here, the LED stripe is green).
Information 13 00480 g001
Figure 2. Wide view of the apparatus (the LED stripe is green).
Figure 2. Wide view of the apparatus (the LED stripe is green).
Information 13 00480 g002
Figure 3. Procedure depending on the assigned conditions.
Figure 3. Procedure depending on the assigned conditions.
Information 13 00480 g003
Figure 4. Evolution of drivers’ level of trust during the drive, depending on the timing of the critical situation and the feedback consistency (error bars: standard error).
Figure 4. Evolution of drivers’ level of trust during the drive, depending on the timing of the critical situation and the feedback consistency (error bars: standard error).
Information 13 00480 g004
Figure 5. Trust and understanding scores for drivers who were assigned to the inconsistent feedback condition. Grey crosses indicate the moment of feedback inconsistencies (error bars: standard error).
Figure 5. Trust and understanding scores for drivers who were assigned to the inconsistent feedback condition. Grey crosses indicate the moment of feedback inconsistencies (error bars: standard error).
Information 13 00480 g005
Figure 6. Proportion of drivers’ gazes towards different area of interest. ***: p < 0.001; *: p < 0.05.
Figure 6. Proportion of drivers’ gazes towards different area of interest. ***: p < 0.001; *: p < 0.05.
Information 13 00480 g006
Table 1. Initial learned trust scale form [13]. * Answer was inverted for scoring.
Table 1. Initial learned trust scale form [13]. * Answer was inverted for scoring.
Items
1I would feel safe in an automated vehicle.
2The automated driving system provides me with more safety compared to manual driving.
3 *I would rather keep manual control of my vehicle than delegate it to the automated driving system on every occasion.
4I would trust the automated driving system decisions.
5I would trust the automated driving system capacities to manage complex driving situations.
6If the weather conditions were bad (e.g., fog, glare, rain), I would delegate the driving task to the automated driving system.
7Rather than monitoring the driving environment, I could focus on other activities confidently.
8If driving was boring for me, I would rather delegate it to the automated driving system than do it myself.
9I would delegate the driving to the automated driving system if I was tired.
Table 2. Participants’ initial and final level of trust. Q1 and Q3 are the first and third quartiles.
Table 2. Participants’ initial and final level of trust. Q1 and Q3 are the first and third quartiles.
Initial Level of Learned TrustFeedback ConsistencyTiming of ErrorTrust
Measurement
NAge (Mean)Trust
MeanMin.Q1MedianQ3Max.
TrustfulConsistentEarlyInitial845.269.86067.5697182
Final81.25075.58291.5100
TrustfulConsistentLateInitial841.1755871.5738090
Final9472808790.5100
TrustfulInconsistentEarlyInitial836.574.55868.57482.590
Final80.26671838794
TrustfulInconsistentLateInitial943.372.95866708290
Final8054688286100
DistrustfulConsistentEarlyInitial740.741.41042445052
Final523034546876
DistrustfulConsistentLateInitial742.645.13045464952
Final69.752616475100
DistrustfulInconsistentEarlyInitial741.647.13641465556
Final65.14649588094
DistrustfulInconsistentLateInitial744.739.12431384756
Final69.438577274100
Table 3. Bartlett’s t-tests for participants’ initial level of trust.
Table 3. Bartlett’s t-tests for participants’ initial level of trust.
Feedback
Consistency
Timing of ErrorInitial Level of Learned TrustBartlett’s
df
Bartlett’s
t
Bartlett’s p
ConsistentEarlyTrustful13.790.052
Distrustful
ConsistentLateTrustful10.5950.441
Distrustful
InconsistentEarlyTrustful10.3390.560
Distrustful
InconsistentLateTrustful10.0020.967
Distrustful
Table 4. Linear models of trust and perceived understanding. ***: p < 0.001; **: p < 0.01; *: p < 0.05; ns: not significant.
Table 4. Linear models of trust and perceived understanding. ***: p < 0.001; **: p < 0.01; *: p < 0.05; ns: not significant.
TrustPerceived
Understanding
(Intercept)40.50***51.57***
Timing of Error and Feedback Consistency
      Early critical situation and consistent feedbackRef. Ref.
      Early critical situation and inconsistent feedback−4.50 ns
      Late critical situation and consistent feedback7.65*ns
      Late critical situation and inconsistent feedback−1.24 ns
Feedback Color and Timing of Error
      Green/HADRef. Ref.
      Orange/HAD and early critical situation8.56 −16.47***
      Green/Correct situations and early critical situation9.23 −2.23
      Orange/Correct situations and early critical situation1.39 −0.67
      Orange/Critical situation and early critical situation−17.33**−21.00***
      Red/TOR and early critical situation−20.22***−18.43***
      Orange/HAD and late critical situation2.11 −24.92***
      Green/Correct situations and late critical situation7.05 0.31
      Orange/Correct situations and late critical situation2.41 1.73
      Orange/Critical situation and late critical situation−25.87***−21.75***
      Red/TOR and late critical situation−25.71***−20.88***
Initial learned level of trust0.46***0.40***
Time (in minutes)0.33*0.10
Age−0.00 0.08
Male−3.43*−4.34*
Adjusted R2 0.2998
Table 5. Chi-squared results for the several comparisons.
Table 5. Chi-squared results for the several comparisons.
Comparisonsχ2dfnp
Green/HADOrange/HAD32.23361<0.001***
Green/Correct Situations40.36361<0.001***
Orange/Correct Situations40.82361<0.001***
Orange/Critical Situation50.53361<0.001***
Red/TOR31.97361<0.001***
Orange/HADGreen/Correct Situations1.48361>0.1
Orange/Correct Situations4.62361>0.1
Orange/Critical Situation8.34361<0.05*
Red/TOR0.902361>0.1
Green/Correct SituationsOrange/Correct Situations1.87361>0.1
Orange/Critical Situation5.06361>0.1
Red/TOR3.5361>0.1
Orange/Correct SituationsOrange/Critical Situation10.23361<0.05*
Red/TOR8.85361<0.05*
Orange/Critical SituationRed/TOR7.67361>0.1
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Manchon, J.B.; Beaufort, R.; Bueno, M.; Navarro, J. Why Does the Automation Say One Thing but Does Something Else? Effect of the Feedback Consistency and the Timing of Error on Trust in Automated Driving. Information 2022, 13, 480. https://doi.org/10.3390/info13100480

AMA Style

Manchon JB, Beaufort R, Bueno M, Navarro J. Why Does the Automation Say One Thing but Does Something Else? Effect of the Feedback Consistency and the Timing of Error on Trust in Automated Driving. Information. 2022; 13(10):480. https://doi.org/10.3390/info13100480

Chicago/Turabian Style

Manchon, J. B., Romane Beaufort, Mercedes Bueno, and Jordan Navarro. 2022. "Why Does the Automation Say One Thing but Does Something Else? Effect of the Feedback Consistency and the Timing of Error on Trust in Automated Driving" Information 13, no. 10: 480. https://doi.org/10.3390/info13100480

APA Style

Manchon, J. B., Beaufort, R., Bueno, M., & Navarro, J. (2022). Why Does the Automation Say One Thing but Does Something Else? Effect of the Feedback Consistency and the Timing of Error on Trust in Automated Driving. Information, 13(10), 480. https://doi.org/10.3390/info13100480

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop