Skip to Content
SensorsSensors
  • Article
  • Open Access

20 May 2022

Design of Proactive Interaction for In-Vehicle Robots Based on Transparency

,
,
,
,
,
and
1
Car Interaction Design Lab, College of Arts and Media, Tongji University, Shanghai 201804, China
2
Shenzhen Research Institute, Sun Yat-Sen University, Shenzhen 518057, China
3
Nanchang Research Institute, Sun Yat-Sen University, Nanchang 330224, China
4
Ulm University, 89081 Ulm, Baden-Württemberg, Germany
This article belongs to the Topic Human–Machine Interaction

Abstract

Based on the transparency theory, this study investigates the appropriate amount of transparency information expressed by the in-vehicle robot under two channels of voice and visual in a proactive interaction scenario. The experiments are to test and evaluate different transparency levels and combinations of information in different channels of the in-vehicle robot, based on a driving simulator to collect subjective and objective data, which focuses on users’ safety, usability, trust, and emotion dimensions under driving conditions. The results show that appropriate transparency expression is able to improve drivers’ driving control and subjective evaluation and that drivers need a different amount of transparency information in different types of tasks.

1. Introduction

With the rapid development of intelligent vehicles, drivers’ requirements of more intelligent assistances from the cockpit have increased. More vehicles are equipped with virtual image voice assistants or vehicle robots with a physical entity, etc. These in-vehicle intelligent assistants enhance the intelligence level of the cockpit and can execute diverse tasks. The interaction between human and in-vehicle robots is considered as an integration of a complex social and technical system [1], which needs an advanced model to improve safety and trust in autonomous vehicles [2].
Anthropomorphism and proactivity have been widely studied for the future in-vehicle robots. A study by Waytz et al. [3] showed that a more anthropomorphic cockpit can increase human trust and is perceived to have more human-like mental abilities. It also showed that an anthropomorphic robot’s voice response can increase trust, pleasure, and dominance of the situation compared to mechanical voice response [4]. However, there are still concerns about communication barriers for such robots. The accuracy and validity of the output produced by intelligent systems can be problematic because it is difficult for the operator to interpret the output [5]. Part of the reason is that humans have limits to understand the proactivity of robots [6]. The study showed that people are more receptive to the support provided by robots with moderate proactivity than those with high or low proactivity [7]. Another reason is that the robot interaction design does not match human cognition.
An important condition for a robot to be able to interact fluently with humans is that the two can share a common cognitive framework [8] and form a coherent mental expectation during the interaction without adding additional learning costs. Therefore, in order to promote a common mental model between human operators and automated systems, in-vehicle robots should be designed according to human cognition to improve communication efficiency and trust. Human cognitive architectures have also been increasingly applied to the intelligent architecture of robots in recent years [9]. Some studies have highlighted system transparency as a potential barrier to trust and technology acceptance [10]. Lee and See [11] also suggested that autonomous systems should communicate their capabilities and limitations as much as possible in order for humans to develop appropriate trust and dependence. It is important to make the state of the autonomous system transparent to humans under conditions consistent with human cognition. Chen and colleagues [12] proposed the situation awareness-based agent transparency theory (SAT) to help establish a common mental model between human and machine, which further generates assistance to the human decision-making process and enhances task performance.
Using cognitive theory to design in-vehicle robots to provide appropriate expression to humans has not been widely studied. This paper applies situation awareness-based agent transparency theory to design in-vehicle robots. The objective of this paper is to explore the appropriate amount of transparency information that can assist human decision making by conducting experiments. Because the inappropriate amount of information while driving, especially in critical situations can affect safety, usability, workload, and trust, the following hypotheses are made.
  • The expressed in-vehicle robot’s information to drivers needs to be selective.
  • The appropriate amount of transparency information for different driving situations needs to be determined.
In order to test these hypotheses, in-vehicle robot transparency design has been made at three levels of perception, comprehension and projection based on SAT theory, including channels of voice and visual in the design. Then, experimental evaluation of the transparency design hypothesis has been carried out in selected driving scenarios. After analyzing the experimental results, conclusions of transparency design of in-vehicle robot proactive interaction in different scenarios have been made.
The contribution of this paper is that it gives results on how an appropriate level of transparency expression affects drivers’ driving control and subjective measurement. It also identifies that drivers need a different amount of transparency information in different types of scenarios. Results can provide guidance to help design in-vehicle robot interface with appropriate transparency.

3. Transparency Design of Proactive Interaction

According to the theory of situation awareness, the human cognitive process includes three levels: perception, comprehension, and projection [30]. In the human–robot interaction process, the necessity of transparency in the three levels is reflected in the human in robot. At the same time, anthropomorphism and proactivity require robots to present multiple channels such as voice and visual.
In the design of in-vehicle human–robot interaction, it is necessary to consider the appropriateness of the amount of information in different channels and different SAT levels. The impact on several aspects needs to be examined such as driving safety, usability, and emotion. The research framework is shown in Figure 1. We apply SAT theory to conduct analysis to determine which SAT level the information is in. Combined with human–robot interface levels [14], the transparency design assumptions for the in-vehicle robot were carried out based on the proactive interaction scenario of the in-vehicle robot. The design assumptions are then evaluated through experiments, using a driving simulator.
Figure 1. Research framework.

3.1. Human–Robot Interface Levels

In order to promote appropriate transparency, Lyons argues that an opportunity to foster transparency between the human and the robot is from the human–robot interface. The human–robot interface includes three levels: informational, communicative, and physical [14]. Each level covers a portion of what transparency design needs to consider. The levels in SAT theory are distinct from the levels of the human–robot interface. The SAT levels focus on the interaction process, where the human–computer interaction process can be well divided into three stages. Meanwhile, human–robot interface levels function as a guidance on the specific design to determine the existence and the amount of specific information.
For specific designs, interface features at the informational level need to be considered to avoid too much information or non-intuitive displays, which may confuse and frustrate the user of the robotic system. Interface features at the communication level need to be considered to avoid the robot’s inappropriate responses, which affects user trust and performance. Interface features at the physical level may include the robot’s emotional expression, effectively describing the robot’s emotional state.
In transparency design assumptions, the amount of information and information intuitiveness should be considered at the informational level. At the communication level, communication smoothness and response timeliness should be considered. At the physical level, emotional expressions should be considered.

3.2. Transparency Design Assumptions

Human–robot transparency design requires identifying certain interaction timing. The identification enables the design of in-vehicle robots to present the driver with the appropriate level of transparency to facilitate the driver’s understanding. The in-vehicle robots give appropriate information that conforms with driver’s cognition, thus resulting in transparency of the entire interaction process. We have explored the possible design combination and excluded the designs that were clearly unreasonable for each interaction timing in a given scenario. For example, considering communication smoothness and response timeliness, the information can be neither too much or too little. Then, the design pattern for each transparency level emerged, forming the transparency design assumption.
Transparency was first analyzed from the perspective of information intuitiveness. From design exploration we conducted before the experiment, we found that when users were given information at the comprehension and perception levels, they can accurately identify the existence of the projection level. However, when the information at the projection level existed alone, users defaulted to the existence of perception and comprehension levels of information. It showed that subjects were unable to cognitively recognize low-level transparency information in the presence of high-level transparency information. When analyzed at the physical level, emotional expression providing information at the perception and comprehension levels may affect subjects’ emotions. Therefore, perception and comprehension levels of information should still be taken into account in the transparency design assumptions.
In the proactive interaction condition of the in-vehicle robot, it has more information that needs to be shared with people, and therefore, it requires a higher transparency of expression. At the perception level, considering emotional expression from the physical level, voice cues can increase the robot’s anthropomorphism and enhance human emotions in the corresponding scenarios. At the comprehension level, considering the informational level aspect, the robot needs to covey the information it perceived to people, and how it comprehends the information needs to be transparent in the voice channel. At the perception and comprehension levels, information in the visual channel, the information intuitiveness, and task fluency are considered, and the comprehension level information is chosen to be retained. At the projection level, considering information intuitiveness, the voice channel gives easier understanding compared to the visual channel, so that the projection information is designed to be expressed through the voice channel. Therefore, the design of the projection level is assumed to include the voice channel information. It is expected that the visual channel projection level information can play an auxiliary role to the voice channel information to strengthen the robot’s expressiveness. Incorporating the communication smoothness into the comprehensive consideration, the presence or absence of visual channel SAT3 level information can be examined. The design assumptions are listed in Table 1 below, where the question mark represents its inability to analyze the necessity of its existence from the design perspective and the need for experimental verification.
Table 1. Transparency design assumptions in proactive interaction condition of the in-vehicle robot.
To summarize the above research approach, the transparency design assumes that the human–robot communication can be more understood by each other and interaction can be more efficient. What kind of design pattern can reach the above goal needs to be determined. Our research approach is first to use the SAT model to analyze the stages, splitting the entire human–robot interaction process into three stages. After the stages were analyzed, we adopt the human–robot interface levels of informational, communicative, and physical guidance. We then further expand human–robot interface levels into the amount of information, information intuitiveness, communication smoothness, response timeliness, and emotional expression for each stage to analyze the information under voice and visual channels and SAT levels.

4. Experiment

4.1. Participants

Thirty subjects (25 males and 5 females), with the age ranging from 22 to 40 years ( M = 28.2 ,   S D = 5.83 ), were selected for the experiment. All subjects had proficient driving experiences, and 16 of them drove 2–3 days per week and 14 subjects drove 4 days per week and above. Subjects were recruited through an online screening process, and their driving experience ranged from 1 to 10 years ( M = 5.8 ,   S D = 2.9 ). Of these, 21 subjects had never used an in-vehicle robot before, nine participants had experienced one once before, and there were no existing in-vehicle robot users. Therefore, the participants were all regarded as novice in-vehicle robot users. The independent variable of the experiment was the degree of transparency, and between-group experiments were designed. The age and gender distribution of subjects were adjusted according to the number of experimental groups in the task so that the demographic attributes of the subjects were as balanced as possible.

4.2. Design of Experiment

The experiment focuses on the proactive interaction scenarios of in-vehicle robots. After conducting real car research and interviews, we collected the scenarios which were used frequently and more representative. The final experimental tasks for proactive interaction with the in-vehicle robot were identified as telephone and speeding. The telephone task required the driver and the in-vehicle robot to complete a non-driving task together. The speeding task required the driver and the in-vehicle robot to complete a critical task together. The experiment used between-subjects design. The aim of the experiment is to explore the relationship between the change in transparency levels, the information quantity inside each level and the driving behavior. Two tasks with different scenarios also enable a comparative exploration of the optimal information quantity design for different scenarios. In the two tasks, the SAT Level 1 was a perception to speeding or incoming call. The SAT Level 2 was a comprehension of the situation mentioned above. The SAT Level 3 was a projection on the driver’s nearest action, expressing as a suggestion or question.
In the telephone task, the experimenter simulated a phone call from the subject’s friend Wang. Then, the robot took the initiative to alert and ask the person whether to answer it. The specific experimental group design is as follows:
  • Group 1 contained SAT Level 2 and SAT Level 3 messages. The robot told the driver ‘Wang is calling you’, with a pleasant expression. Then, the robot asked the driver ‘Would you like to answer’, with a phoning expression.
  • Group 2 contained SAT Level 1, SAT Level 2 and SAT Level 3 messages. SAT Level 1 voice channel message ‘Ring’ was added.
The set up for the telephone task is shown in Figure 2. In total, 12 males and 3 females were assigned to experimental group 1; 13 males and 2 females were assigned to experimental group 2.
Figure 2. Telephone task experimental group setup. The gray choices represent information that appeared at a certain level in the experimental group, and the white choices represent that there is no information in that level.
In the speeding task, the subject was asked to drive in the left lane at a speed of 30 km/h and then accelerated to 70 km/h. Once the speed was above 60 km/h, the robot took the initiative to remind the driver of the speed limit. The specific experimental group design is as follows:
  • Group 1 contained SAT Level 2 and SAT Level 3 messages. The robot told the driver ‘The speed limit ahead is 60 km/h, you have exceeded the speed limit’, with a fear expression. Then, the robot suggested the driver ‘Slow down please’.
  • Group 2 also contained SAT Level 2 and SAT Level 3 messages. Based on having all the information in the experimental group 1, a speed limit expression, containing an SAT Level 3 visual channel message, was added.
  • Group 3 contained SAT Level 1, SAT Level 2, and SAT Level 3 messages. Based on having all the information in the experimental group 1, an SAT Level 1 voice channel message ‘Oops’, which expressed robot perception to the driver, was added.
The experimental group setup for the speeding task is shown in Figure 3. In total, 8 males and 2 females were assigned to experimental group 1 and experimental group 2; 9 males and 1 female were assigned to experimental group 3.
Figure 3. Speeding task experimental group setup. The gray choices represent information that appeared at a certain level in the experimental group, and the white choices represents that there is no information in that level.

4.3. Measurement

The behavior of the in-vehicle robots will draw the attention of the drivers, even if robots perform decisions that conform with the drivers’ cognition and give them more helpful information. Appropriate design of human–robot interaction strategies based on human cognitive factors can help compensate for human limitations to achieve safety [27]. Therefore, it is necessary to examine the multi-channel transparency design assumptions of in-vehicle robots to ensure driving safety, to improve task execution efficiency, and to enhance drivers’ trust in the robot.
Harbluk et al. [33] showed that the driver’s visual behavior and vehicle control changed when performing tasks with different cognitive requirements. The driver’s visual behavior data and vehicle data were collected while participants were performing tasks. The visual behavior data included the total saccades time (times) and total fixation time (seconds), which were extracted from recorded videos. The vehicle data included vehicle speed (km/h) and driveway offset (dm), which measured subjects’ driving control in the vertical and horizontal directions, respectively. In terms of subjective data, post-task questionnaires were used to make subjects score subjectively on the usability, trust, workload and affective dimensions using a Likert scale. Usability was measured using the After-Scenario Scale (ASQ) [34], which combined three dimensions of ease of task completion, the time required to complete tasks, and satisfaction with support information to produce an evaluation. The trust scores in the study were based on the model of trust in vehicle automation proposed by Muir [35], using a post-task trust scale with a comprehensive analysis of three trust dimensions: predictability, dependability, and faith. It was found that human reliance on automation was influenced by workload [36]. We used the Driving Activity Load Index (DALI) scale [37], which is a scale that concerns multi-channel information and includes the effort of attention, visual demand, auditory demand, temporal demand, interference, and situational stress. The DALI scale is more fit for dynamic driving conditions. The SAM scale [38] designed by Bradley et al. was used to measure the emotional state of the person in terms of pleasure, arousal, and dominance. In the comparison of the analysis of the result between specific experimental groups, priority was given to safety, followed by usability, trust, and workload ratings. Emotional ratings were also taken into account as a secondary evaluation.

4.4. Apparatus and Materials

The experimental environment was built based on the driving simulator system independently developed by the Car Interaction Design Lab of Tongji University. The simulator was used as the main equipment of the experiment. The scene was developed using Unity software to simulate the real driving environment. The scene used in the experiment was a two-lane straight road with a large number of oncoming cars in the opposite lane. The simulator was equipped with monitoring equipment that automatically collected vehicle data during each simulation. The robot was fixed in a suitable position (see Figure 4). The location of the robot is determined by three factors: first, previous work by Williams [24] and others showed that in-vehicle robots were fixed in a position above the center screen. Second, we conducted a real vehicle study on nomi, the in-vehicle robot of Nio, measuring the relative position of the in-vehicle robot to the center of the steering wheel in three-dimensional space. Third, we placed the robot on our simulator and conducted a small test on the pre-fixed robot to adjust its position to make it closer to the real driving environment (see Figure 5). The vehicle robot had three degrees of freedom and was controlled by servos that can raise and lower its head and rotate toward the driver. The robot’s face screen displayed features and colorful auxiliary graphics to express expressions, which was in-depth explored and designed in our previous research [39]. The expression design was shown in Figure 6, in which the color and brush strokes were adjusted due to the confidentiality of company cooperation. The in-vehicle robot was accompanied by an interactive simulation program to control the robot’s movements and expressions. The interactive simulation program can record the robot’s behavioral data. Since the program requires the control of the experimenter, factors such as the experimenter’s reaction time may pose a threat to the validity of the experiment. The subject’s basic information form and scale were used to collect subjective data. Two cameras were used to record the visual behavior of the subjects. After the videos were recorded, the user’s saccades of the robot were manually checked frame by frame. The number of saccades time and fixation time were recorded with a minimum frequency unit of 0.042 s (1/24 s per frame). Similar simulator construction and data acquisition methods also appeared in other contemporary studies [40].
Figure 4. Robot for experiment.
Figure 5. Experiment environment.
Figure 6. Expression design set used on robot’s face screen [39].

4.5. Procedure

The experimental consent forms were signed before the experiment, and the subject was registered for basic information including driving frequency as well as in-vehicle robot understanding and experience. Before conducting the experiment, the subject was introduced with the purpose of the experiment and the main tasks. The subject started with a driving in the simulator for about five minutes to learn and became familiar with the simulator. Then, the subject interacted with the in-vehicle robot in a simulation. After the subject was familiar with the simulator and the in-vehicle robot, the experiment started, and the experimenter began data recordings.
The tester described the driving task requirements and subtask requirements to the subject. After the subject confirmed, the tester issued the command to start the task. Subjects completed the telephone task first and the speeding task second. In each task, subjects were assigned to one of the groups. In the telephone task, the subject was asked to keep driving in the left lane at a speed of 30 km/h. After the subject kept driving stable, the experimenter simulated a phone call from the subject’s friend Wang. Then, the robot took the initiative to alert and ask the person whether to answer it. In the speeding task, the subject was asked to drive in the left lane at a speed of 30 km/h and then accelerated to 70 km/h. Once the speed reached 70 km/h, the robot took the initiative to remind the speed limit. After each task was completed, the subject completed the subjective questionnaires and scales. Then, the tester conducted the interview regarding the task. Each task lasted approximately five to ten minutes. After tasks were completed, subjects were interviewed by the tester regarding the in-vehicle robot generally.

5. Results

T-tests were conducted in the telephone task and one-way analysis of variance (ANOVA) with post hoc pairwise comparison was conducted in the speeding task. A summary of the experimental results in separate tasks can be seen in the following sections.

5.1. Telephone Task Results

As shown in Figure 7, the mean value of the standard deviation of driveway offset for experimental group 1 was 0.18, and the mean value of the standard deviation of driveway offset for experimental group 2 was 0.77. There was a significant difference between the two experimental groups ( p = 0.00 < 0.01 ).
Figure 7. Standard deviation of driveway offset in telephone task. ** indicates significant difference (p < 0.01).
As shown in Figure 8, the mean saccades time was 3 s for experimental group 1 and 5 s for experimental group 2. There was a significant difference between the two experimental groups ( p = 0.02 < 0.05 ) . The mean fixation time was 1.671 s for experimental group 1 and 3.893 s for experimental group 2. There was a highly significant difference between the two experimental groups ( p = 0.00 < 0.01 ) . Subjects in experimental group 1 had significantly fewer saccades time and fixation time than subjects in experimental group 2.
Figure 8. Saccades time and fixation time in telephone task. * indicates significant difference (p < 0.05). ** indicates significant difference (p < 0.01).
As shown in Figure 9, it can be concluded from the usability scores that the experimental group 2 subjects rated the usability of the robot lower ( p = 0.049 ) . In terms of details, the time required to complete tasks was significantly different between experimental groups ( p = 0.01 ) , while ease of task completion ( p = 0.374 ) and satisfaction with support information ( p = 0.262 ) was not significantly different.
Figure 9. Usability in telephone task. * indicates significant difference (p < 0.05).
As shown in Figure 10, it was concluded from the PAD scores that experimental group 2 enabled subjects to obtain more positive emotions. The subjects were more pleasant ( p = 0.001 ) and aroused ( p = 0.001 ) than experimental group 1. There was no significant difference in dominance ( p = 0.202 ) .
Figure 10. Emotion score of telephone task. ** indicates significant difference (p < 0.01).
As shown in Figure 11, it can be concluded from the workload scores that in experimental group 2, subjects had a lower workload than in experimental group 1 ( p = 0.002 ) . Judging from the DALI detail scores, the main reasons were lower interference ( p = 0.003 ) and lower situational stress ( p = 0.002 ) .
Figure 11. Workload in telephone task. ** indicates significant difference (p < 0.01).

5.2. Speeding Task Results

As shown in Figure 12, the mean value of the standard deviation of driveway offset was 0.95 for experimental group 1, 1.54 for experimental group 2, and 1.55 for experimental group 3. The results of the one-way ANOVA performed on the data from the three groups indicated that there was a significant difference in the standard deviation of driveway offset among the three experimental groups ( F ( 2 , 32 ) = 6.906 ,   p = 0.00 < 0.01 ) ; post hoc tests revealed that the standard deviation of driveway offset was significantly lower in experimental group 1 ( M = 0.97 ,   S D = 0.51 ) than in experimental group 2 ( M = 1.60 ,   S D = 0.58 ) and experimental group 3 ( M = 1.55 ,   S D = 0.28 ) , with no significant differences between experimental group 2 and group 3.
Figure 12. Standard deviation of driveway offset in speeding task. * indicates significant difference (p < 0.05). ** indicates significant difference (p < 0.01).
As shown in Figure 13, the mean saccades time was 3.7 s for experimental group 1, 1.3 s for experimental group 2, and 1.4 s for experimental group 3. One-way ANOVA results of the data from the three groups showed that there was a significant difference in the saccades time in the three experimental groups, ( F ( 2 , 18 ) = 9.479 ,   p = 0.00 < 0.01 ) ; post hoc tests revealed that the saccades time in experimental group 1 ( M = 3.67 ,   S D = 1.73 ) was significantly higher than that in experimental group 2 ( M = 1.25 , S D = 0.50 ) and experimental group 3 ( M = 1.38 , S D = 0.52 ) , with no significant difference between experimental group 2 and experimental group 3. The mean fixation time of experimental group 1 was 2.389 s, the mean fixation time of experimental group 2 was 0.540 s, and the mean fixation time of experimental group 3 was 0.834 s. A one-way ANOVA of the data from the three groups showed that there was a significant difference in the fixation time of the three experimental groups, ( F ( 2 , 20 ) = 7.245 ,   p = 0.00 < 0.01 ) ; post hoc tests revealed that the fixation time of experimental group 1 ( M = 2.39 , S D = 1.41 ) was significantly higher than those of experimental group 2 ( M = 0.54 , S D = 0.30 ) and experimental group 3 ( M = 0.83 , S D = 0.50 ) , and there was no significant difference between experimental group 2 and experimental group 3.
Figure 13. Saccades time and fixation time in speeding task. *** indicates significant difference (p < 0.001).
The results of the behavior data showed that the subjects in experimental group 1 were significantly higher than those in experimental group 2 and experimental group 3 in terms of saccades time and fixation time.
As shown in Figure 14, significant differences were shown between experimental groups in terms of workload score means ( F ( 2 , 55 ) = 10.408 ,   p = 0.000 < 0.001 ) and also in terms of each detailed dimension. Post hoc test analysis yielded that the workload of experimental group 2 was significantly higher than those of experimental group 1 ( p = 0.000 < 0.001 ) and experimental group 3 ( p = 0.002 < 0.01 ) , while experimental group 1 and experimental group 3 did not show significant differences in both workload score means and workload score details.
Figure 14. Workload in speeding task. ** indicates significant difference (p < 0.01). *** indicates significant difference (p < 0.001).
As shown in Figure 15, it can be concluded from the PAD scores that there were significant differences among three experimental groups in the arousal ( F ( 2 , 38 ) = 3.430 ,   p = 0.043 < 0.05 ) and dominance ( F ( 2 , 42 ) = 5.945 ,   p = 0.005 < 0.01 ) dimensions. Post hoc test analysis yielded significantly higher arousal ( p = 0.036 < 0.05 ) and significantly higher dominance ( p = 0.004 < 0.01 ) in experimental group 1 than in experimental group 2.
Figure 15. Emotion score in speeding task. * indicates significant difference (p < 0.05). ** indicates significant difference (p < 0.01).

6. Discussion

In the telephone task, the design with SAT1 perception level, the voice channel information (hereafter referred to as the ringing group) in group 2 enabled subjects to obtain more positive emotions and reduced the workload of the subjects. This was probably because the added information enhanced the anthropomorphism of the robot. However, subjects had significantly worse horizontal control of the vehicle, and the saccades time and fixation time increased significantly. This suggests that the subjects were overly distracted by the robot and had reduced concentration on the driving task, which may cause danger. As for the subjective data, subjects rated the usability of the ringing group lower, and significant differences were found mainly in the aspect of time required to complete tasks. It can be obtained that task disfluency due to lengthy speech triggered subjects’ lower usability ratings of the robot. In addition, the increased information at the expense of time spent did not improve usability ratings in terms of ease of task completion and satisfaction of support information. Overall, the existence of SAT1 perception level provide higher transparency. However, it does not suggest better cooperation between human and robot.
In the speeding task, the design with SAT2 comprehension level presenting voice and visual channel information as well as SAT3 projection level showing voice channel information had better information transparency. In terms of safety, group 1 had the best data performance in terms of vehicle horizontal control. Adding the SAT3 projection level visual channel information or SAT1 level voice channel information would make the driver’s horizontal control of the vehicle worse. From the subjective data, having SAT3 level visual channel information resulted in significantly higher workload and lower arousal and dominance. The speed limit expression at the projection level caused subjects’ greater workload and did not help in driving, indicating that it was an unnecessary distraction for the driver. The presence or absence of SAT1 level with voice channel information (‘oops’) did not significantly affect workload. Therefore, in the speeding task, the design of the in-vehicle robot can omit the SAT1 stage information, which is not very helpful for the task, and the complicated SAT3 level visual information can be similarly excluded to increase safety and reduce the workload.
In summary, we analyzed the appropriate amount of transparency information for in-vehicle robots in a proactive interaction scenario. The experimental groups with SAT1 performed poorly in both scenarios. In order to improve the effectiveness of message delivery and help drivers concentrate, designers should carefully consider increasing voice channel information at the SAT1 perception level. The design that performs better in telephone tasks has SAT2 comprehension and SAT3 projection levels of both voice and visual channel information, which is different from the better design for speeding tasks. In the speeding task, the SAT3 projection level and visual information make the workload of subjects significantly higher, and the driving safety is compromised. Analyzing the task characteristics, we can conclude that the telephone task belongs to function, while the speeding task belongs to critical scenarios. The driver’s concentration level on driving tasks in the two scenarios is different; therefore, the driver needs different robot transparency.
The advantage of higher transparency for the in-vehicle robot is that it can help drivers judge the situation, reduce drivers’ need to reprocess the information provided by the robot, and increase the drivers’ decision-making confidence. However, higher transparency also has its own disadvantage. In the case of insufficient attention and mental resources, too much information is not only not fully received by the human but also affects the driver’s driving performance, especially when implicit information such as robot expressions that require extra mental resources from the driver appears. These results suggest that the ‘highly transparent’ assumption made by previous work needs careful consideration. In critical scenarios, it is recommended to design more direct voice information without adding extra workload. In non-critical driving scenarios, designers can consider giving robots a more transparent response to drivers while maintaining task fluency.
The experiments were based on transparency design assumptions, and they yielded preliminary results for the exploration of suitable transparency. We are able to draw enlightening conclusions from different measurements, and the purposes of the experiments are essentially achieved. However, due to the experimental limitations, it was not possible to exhaust the possible design assumptions, and there may be more appropriate transparency designs for specific scenarios.
The relationship between saccades time and driving safety was also noted in both scenarios. In the non-critical scenario, an increase in saccades time and fixation time on the robot undermines driving safety. In the critical scenario, on the contrary, an increase in saccades time and fixation time of the robot improves driving control. The reason for this result derives from the difference in the type of tasks that the driver and the in-vehicle robot achieve together in the two scenarios. Taking workload into comprehensive analysis, in the case of joint driving tasks such as the speeding task, the robot’s information alerts can share the workload for the driver and obtain better driving performance. However, when the robot is completing non-driving tasks such as picking up the phone, the increase in saccades time and fixation time indicates that the driver is more involved in non-driving tasks and does not concentrate on driving tasks, so the driver tends to have poor control of the vehicle. Further analysis revealed a tendency for the reduced self-reported workload when drivers were more involved in non-driving tasks. This may be due to the higher workload of the driving task compared to the non-driving task, which is also able to corroborate with the multi-resource prediction of dual-task interference in the multi-resource theory (MRT) proposed by Wickens [41].

7. Conclusions and Future Work

In this paper, we migrate SAT theory into an in-vehicle robot design method and conduct a proactive interaction design of in-vehicle robots based on transparency design assumptions. The results of the experiment show that the driver’s driving control, behavior, and self-reporting results produce differences under different in-vehicle robot transparency. Overall, anthropomorphic information of perception level, though bringing a better emotional experience to the driver, can cause poorer driving performance. In the critical scenario, because of the need to ensure timely and effective information, the driver can accept lower information transparency of the in-vehicle robot than in the non-critical scenario, retaining only key information. The conclusions can be distilled into design guidance: in the in-vehicle scenario, information transparency design assumptions can be made for both voice and visual channels based on SAT theory to help designers arrive at interaction solutions with more appropriate transparency. The guidance can also give a methodology to solve the transparency challenge that lies in future research of human-centered shared control [42]. Following the guidance, this paper verifies that in the proactive interaction scenario of in-vehicle robots, non-critical scenarios require comprehension and projection level information, while the projection level information can be reduced in critical scenarios.
At this stage, the multi-channel interaction design is not separated from transparency. The multi-channel content is only used as an independent variable of transparency change to serve the transparency design. In terms of future work, we will go further to establish the design model integrating multimodal and transparency, concerning input and output modalities [43], and then add subsequent experiments to validate it. A study shows that EEG-based measurements can be a powerful tool for studying driver behavior [44]. It is also able to measure the effect of multimodal information on cognition very accurately [45]. Psychological precision measurement instruments such as EEG, ECG, or eye movements have not been included in the scope of measurement methods in this paper, and further work needs to consider bringing such measurement tools methods into the experiment.

Author Contributions

Conceptualization, F.Y. (Fang You); Project administration, J.W.; Methodology, Y.L.; Software, C.W.; Visualization, Y.W.; Writing—original draft, T.Y.; Writing—review & editing, T.Y. and F.Y. (Fei Yan). All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Social Science Fund of China (No. 19FYSB040); CES-Kingfar Excellent Young Scholar Joint Research Funding (No.202002JG26); China Scholarship Council Foundation (2020-1509); Shenzhen Collaborative Innovation Project: International Science and Technology Cooperation (No. GHZ20190823164803756); Innovative talents of Nanchang “Double Hundred Plan” (introduced category); Tongji University Excellent Experimental Teaching Program (0600104096).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Data sharing is not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Moniz, A.B.; Krings, B.-J. Robots Working with Humans or Humans Working with Robots? Searching for Social Dimensions in New Human-Robot Interaction in Industry. Societies 2016, 6, 23. [Google Scholar] [CrossRef] [Green Version]
  2. Mokhtarzadeh, A.A.; Yangqing, Z.J. Human-robot interaction and self-driving cars safety integration of dispositif networks. In Proceedings of the IEEE International Conference on Intelligence and Safety for Robotics (ISR), Shenyang, China, 24–27 August 2018; pp. 494–499. [Google Scholar]
  3. Waytz, A.; Heafner, J.; Epley, N. The mind in the machine: Anthropomorphism increases trust in an autonomous vehicle. J. Exp. Soc. Psychol. 2014, 52, 113–117. [Google Scholar] [CrossRef]
  4. Large, D.R.; Harrington, K.; Burnett, G.; Luton, J.; Thomas, P.; Bennett, P. To please in a pod: Employing an anthropomorphic agent-interlocutor to enhance trust and user experience in an autonomous, self-driving vehicle. In Proceedings of the 11th International Conference on Automotive User Interfaces and Interactive Vehicular Applications, Utrecht, The Netherlands, 22–25 September 2019; pp. 49–59. [Google Scholar]
  5. Linegang, M.P.; Stoner, H.A.; Patterson, M.J.; Seppelt, B.D.; Hoffman, J.D.; Crittendon, Z.B.; Lee, J.D. Human-automation collaboration in dynamic mission planning: A challenge requiring an ecological approach. In Proceedings of the Human Factors and Ergonomics Society Annual Meeting, San Francisco, CA, USA, 1 October 2006; Volume 50, pp. 2482–2486. [Google Scholar]
  6. Miller, G.A. The magical number seven plus or minus two: Some limits on our capacity for processing information. Psychol. Rev. 1956, 63, 81–97. [Google Scholar] [CrossRef] [Green Version]
  7. Peng, Z.; Kwon, Y.; Lu, J.; Wu, Z.; Ma, X. Design and evaluation of service robot’s proactivity in decision-making support process. In Proceedings of the CHI Conference on Human Factors in Computing Systems, Glasgow, Scotland, 2 May 2019; pp. 1–13. [Google Scholar]
  8. Chen, J.Y.C.; Barnes, M.J. Human–Agent Teaming for Multirobot Control: A Review of Human Factors Issues. IEEE Trans. Hum.-Mach. Syst. 2014, 44, 13–29. [Google Scholar] [CrossRef]
  9. Cassenti, D.N.; Kelley, T.D.; Yagoda, R.E.; Avery, E. Improvements in robot navigation through operator speech preferences. Paladyn J. Behav. Robot. 2012, 3, 102–111. [Google Scholar] [CrossRef]
  10. Davis, F.D. Perceived Usefulness, Perceived Ease of Use, and User Acceptance of Information Technology. MIS Q. 1989, 13, 319–340. [Google Scholar] [CrossRef] [Green Version]
  11. Lee, J.; See, K.A. Trust in Automation: Designing for Appropriate Reliance. Hum. Factors 2004, 46, 50–80. [Google Scholar] [CrossRef]
  12. Chen, J.Y.; Procci, K.; Boyce, M.; Wright, J.; Garcia, A.; Barnes, M. Situation Awareness-Based Agent Transparency; ARL-TR-6905; U.S. Army Research Laboratory: Aberdeen, MD, USA, 2014. [Google Scholar]
  13. Mast, M.; Burmester, M.; Fatikow, S.; Festl, L.; Krüger, K.; Kronreif, G. Conceptual HRI Design, Usability Testing, HRI Interaction Patterns, and HRI Principles for Remotely Controlled Service Robots; Deliverable D2.2; European Commission: Bruxelles, Belgium, 2011. [Google Scholar]
  14. Lyons, J. Being transparent about transparency: A model for human-robot interaction. In Proceedings of the AAAI Spring Symposium: Trust and Autonomous Systems, Standford, CA, USA, 25–27 March 2013. [Google Scholar]
  15. Cheverst, K.; Byun, H.E.; Fitton, D.; Sas, C.; Kray, C.; Villar, N.; Kray, C. Exploring Issues of User Model Transparency and Proactive Behaviour in an Office Environment Control System. User Model. User-Adapt. Interact. 2015, 15, 235–273. [Google Scholar] [CrossRef]
  16. Merat, N.; Seppelt, B.; Louw, T.; Engström, J.; Lee, J.D.; Johansson, E.; Green, C.A.; Katazaki, S.; Monk, C.; Itoh, M.; et al. The “Out-of-the-Loop” concept in automated driving: Proposed definition, measures and implications. Cogn. Technol. Work 2018, 21, 87–98. [Google Scholar] [CrossRef] [Green Version]
  17. Rau, P.-L.P.; Li, Y.; Liu, J. Effects of a Social Robot’s Autonomy and Group Orientation on Human Decision-Making. Adv. Hum. Comput. Interact. 2013, 2013, 263721. [Google Scholar] [CrossRef] [Green Version]
  18. Wang, J.; Liu, Y.; Yue, T.; Wang, C.; Mao, J.; Wang, Y.; You, F. Robot Transparency and Anthropomorphic Attribute Effects on Human–Robot Interactions. Sensors 2021, 21, 5722. [Google Scholar] [CrossRef] [PubMed]
  19. Szajna, A.; Stryjski, R.; Woźniak, W.; Chamier-Gliszczyński, N.; Kostrzewski, M. Assessment of Augmented Reality in Manual Wiring Production Process with Use of Mobile AR Glasses. Sensors 2020, 20, 4755. [Google Scholar] [CrossRef]
  20. Williams, K.; Flores, J.A.; Peters, J. affective robot influence on driver adherence to safety, cognitive load reduction and sociability. In Proceedings of the 6th International Conference on Automotive User Interfaces and Interactive Vehicular Applications, Seattle, WA, USA, 17–19 September 2014. [Google Scholar]
  21. Buss, M.; Carton, D.; Gonsior, B.; Kuehnlenz, K.; Landsiedel, C.; Mitsou, N.; de Nijs, R.; Zlotowski, J.; Sosnowski, S.; Strasser, E.; et al. Towards proactive human-robot interaction in human environments. In Proceedings of the 2nd International Conference on Cognitive Infocommunications (CogInfoCom), Budapest, Hungary, 7–9 July 2011; pp. 1–6. [Google Scholar]
  22. Cuevas, H.M.; Fiore, S.M.; Caldwell, B.S.; Strater, L. Augmenting team cognition in human-automation teams performing in complex operational environments. Aviat. Space Environ. Med. 2007, 78 (Suppl. 1), B63–B70. [Google Scholar] [PubMed]
  23. Vasconez, J.P.; Carvajal, D.; Cheein, F.A. On the design of a human-robot interaction strategy for commercial vehicle driving based on human cognitive parameters. Adv. Mech. Eng. 2019, 11, 1687814019862715. [Google Scholar] [CrossRef] [Green Version]
  24. Van Dongen, K.; van Maanen, P.-P. A framework for explaining reliance on decision aids. Int. J. Hum. Comput. Stud. 2013, 71, 410–424. [Google Scholar] [CrossRef]
  25. Park, K.; Im, Y. Ergonomic Guidelines of Head-Up Display User Interface during Semi-Automated Driving. Electronics 2020, 9, 611. [Google Scholar] [CrossRef] [Green Version]
  26. Bader, R.; Siegmund, O.; Woerndl, W. A study on user acceptance of proactive in-vehicle recommender systems. In Proceedings of the 3rd International Conference on Automotive User Interfaces and Interactive Vehicular Applications, Salzburg, Austria, 30 November–2 December 2011; pp. 47–54. [Google Scholar]
  27. Billsus, D.; Hilbert, D.M.; Maynes-Aminzade, D. Improving proactive information systems. In Proceedings of the 10th International Conference on Intelligent User Interfaces, San Diego, CA, USA, 10–13 January 2005; pp. 159–166. [Google Scholar]
  28. Russell, S.J.; Norvig, P. Artificial Intelligence: A Modern Approach, 3rd ed.; Tsinghua University Press: Beijing, China, 2011. [Google Scholar]
  29. Lee, J.; Moray, N. Trust, control strategies and allocation of function in human-machine systems. Ergonomics 1992, 35, 1243–1270. [Google Scholar] [CrossRef]
  30. Endsley, M. Toward a Theory of Situation Awareness in Dynamic Systems. Hum. Factors J. Hum. Factors Ergon. Soc. 1995, 37, 32–64. [Google Scholar] [CrossRef]
  31. Selkowitz, A.R.; Lakhmani, S.G.; Larios, C.N.; Chen, J.Y. Agent transparency and the autonomous squad member. In Proceedings of the Human Factors and Ergonomics Society Annual Meeting 60, Washington, DC, USA, 19–23 September 2016; pp. 1319–1323. [Google Scholar]
  32. Mercado, J.E.; Rupp, M.A.; Chen, J.Y.C.; Barnes, M.J.; Barber, D.; Procci, K. Intelligent Agent Transparency in Human–Agent Teaming for Multi-UxV Management. Hum. Factors J. Hum. Factors Ergon. Soc. 2016, 58, 401–415. [Google Scholar] [CrossRef] [Green Version]
  33. Harbluk, J.L.; Noy, Y.I.; Eizenman, M. The impact of cognitive distraction on driver visual behaviour and vehicle control. In Proceedings of the 81st Annual Meeting of the Transportation Research Board, Washington, DC, USA, 8–12 January 2002. [Google Scholar]
  34. LewisJames, R. Psychometric evaluation of an after-scenario questionnaire for computer usability studies. ACM Sigchi. Bull. 1991, 23, 78–81. [Google Scholar] [CrossRef]
  35. Muir, B.M. Trust in automation. I: Theoretical issues in the study of trust and human intervention in automated systems. Ergonomics 1994, 37, 1905–1922. [Google Scholar] [CrossRef]
  36. Manzey, D.; Reichenbach, J.; Onnasch, L. Human Performance Consequences of Automated Decision Aids. J. Cogn. Eng. Decis. Mak. 2012, 6, 57–87. [Google Scholar] [CrossRef]
  37. Pauzie, A. A method to assess the driver mental workload: The driving activity load index (DALI). IET Intell. Transp. Syst. 2008, 2, 315–322. [Google Scholar] [CrossRef]
  38. Bradley, M.M.; Lang, P.J. Measuring emotion: The self-assessment manikin and the semantic differential. J. Behav. Ther. Exp. Psychiatry 1994, 25, 49–59. [Google Scholar] [CrossRef]
  39. Wang, J.; Wang, Y.; Liu, Y.; Yue, T.; Wang, C.; Yang, W.; Hansen, P.; You, F. Experimental Study on Abstract Expression of Human-Robot Emotional Communication. Symmetry 2021, 13, 1693. [Google Scholar] [CrossRef]
  40. Charissis, V.; Falah, J.; Lagoo, R.; Alfalah, S.F.M.; Khan, S.; Wang, S.; Altarteer, S.; Larbi, K.B.; Drikakis, D. Employing Emerging Technologies to Develop and Evaluate In-Vehicle Intelligent Systems for Driver Support: Infotainment AR HUD Case Study. Appl. Sci. 2021, 11, 1397. [Google Scholar] [CrossRef]
  41. Wickens, C. Multiple Resources and Mental Workload. Hum. Factors J. Hum. Factors Ergon. Soc. 2008, 50, 449–455. [Google Scholar] [CrossRef] [Green Version]
  42. Xing, Y.; Huang, C.; Lv, C. Driver-Automation Collaboration for Autoated Vehicles: A Review of Human-Centered Shared Control. In Proceedings of the 2020 IEEE Intelligent Vehicles Symposium (IV), Las Vegas, NV, USA, 19 October–13 November 2020; pp. 1964–1971. [Google Scholar]
  43. Detjen, H.; Faltaous, S.; Pfleging, B.; Geisler, S.; Schneegass, S. How to Increase Automated Vehicles’ Acceptance through In-Vehicle Interaction Design: A Review. Int. J. Hum.-Comput. Interact. 2021, 37, 308–330. [Google Scholar] [CrossRef]
  44. Di Flumeri, G.; Borghini, G.; Aricò, P.; Sciaraffa, N.; Lanzi, P.; Pozzi, S.; Vignali, V.; Lantieri, C.; Bichicchi, A.; Simone, A.; et al. EEG-Based Mental Workload Neurometric to Evaluate the Impact of Different Traffic and Road Conditions in Real Driving Settings. Front. Hum. Neurosci. 2018, 12, 509. [Google Scholar] [CrossRef] [Green Version]
  45. Marucci, M.; Di Flumeri, G.; Borghini, G.; Sciaraffa, N.; Scandola, M.; Pavone, E.F.; Babiloni, F.; Betti, V.; Aricò, P. The impact of multisensory integration and perceptual load in virtual reality settings on performance, workload and presence. Sci. Rep. 2021, 11, 4831. [Google Scholar] [CrossRef]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.