Next Article in Journal
Deleterious Effect of Air Pollution on Human Microbial Community and Bacterial Flora: A Short Review
Next Article in Special Issue
CEEMDAN-IPSO-LSTM: A Novel Model for Short-Term Passenger Flow Prediction in Urban Rail Transit Systems
Previous Article in Journal
Neurostimulation for Chronic Low Back Pain during Pregnancy: Implications for Child and Mother Safety
Previous Article in Special Issue
An Integrated Multi-Objective Optimization for Dynamic Airport Shuttle Bus Location, Route Design and Departure Frequency Setting Problem
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

IvCDS: An End-to-End Driver Simulator for Personal In-Vehicle Conversational Assistant

1
School of Transportation and Civil Engineering, Nantong University, Nantong 226000, China
2
School of Informatics, University of Edinburgh, Edinburgh EH8 9YL, UK
3
Alibaba Group, Hangzhou 311121, China
4
ADAPT Centre, School of Computing, Dublin City University, D09 DXA0 Dublin, Ireland
5
Faculty of Languages and Translation, Macao Polytechnic University, Macao, China
6
China Electronics Standardization Institute, Beijing 101102, China
7
SFI Centre for Research Training in Machine Learning, School of Computing, Dublin City University, D09 DXA0 Dublin, Ireland
*
Author to whom correspondence should be addressed.
Int. J. Environ. Res. Public Health 2022, 19(23), 15493; https://doi.org/10.3390/ijerph192315493
Submission received: 14 October 2022 / Revised: 9 November 2022 / Accepted: 19 November 2022 / Published: 22 November 2022

Abstract

:
An advanced driver simulator methodology facilitates a well-connected interaction between the environment and drivers. Multiple traffic information environment language processing aims to help drivers accommodate travel demand: safety prewarning, destination navigation, hotel/restaurant reservation, and so on. Task-oriented dialogue systems generally aim to assist human users in achieving these specific goals by a conversation in the form of natural language. The development of current neural network based dialogue systems relies on relevant datasets, such as KVRET. These datasets are generally used for training and evaluating a dialogue agent (e.g., an in-vehicle assistant). Therefore, a simulator for the human user side is necessarily required for assessing an agent system if no real person is involved. We propose a new end-to-end simulator to operate as a human driver that is capable of understanding and responding to assistant utterances. This proposed driver simulator enables one to interact with an in-vehicle assistant like a real person, and the diversity of conversations can be simply controlled by changing the assigned driver profile. Results of our experiment demonstrate that this proposed simulator achieves the best performance on all tasks compared with other models.

1. Introduction

Transportation-related issues are highly related to the road/driver safety and traffic environment, and are attracting growing interests [1,2,3,4]. For example, fatigue driving can significantly raise the possibility of car accidents [5], and is generally influenced by driver-related factors such as the state of sleep and health [6]. The developing intelligent driving techniques are deemed to be capable of providing feasible approaches to addressing such issues [7,8], and recent research shows a propensity of utilizing intelligent vehicle systems (e.g., a driver assistant system or a driver behavior-recognition system). These systems are designed to detect the real intents behind human driver behaviors and adopt relevant measures [9,10], resulting in the improvement of driving safety and efficiency. Despite the prevalence of the application of intelligent driving however, it additionally raises the concern that these newly advanced techniques may even lead to traffic accidents due to fatal errors such as causing driver distraction or taking wrong actions [11].
Meanwhile, because current work mainly focuses on the system side, a driving simulator is an appropriate technique which provides approaches to simulating how real drivers interact with the driving environment (e.g., an intelligent driving system). There exists practical applications of driving simulators which involve a wide range of traffic areas and vehicle techniques, such as advanced driver assistance systems [12], driver education in driving automation [13], assessment of safety at signalized intersection [14], and so on. For example, Simović et al. [15] employ an in-laboratory driving simulator to investigate how different factors can influence the e-bicycle speed perception of human drivers.
Hence, we propose a conversational driver simulator in this paper by leveraging the task of dialogue systems, which can potentially facilitate driver safety by increasing the quality of drivers’ interaction with in-vehicle assistant systems. Generally speaking, dialogue systems aim to generate appropriate responses to user input, and it is an important research direction not only due to its practical applications but also for its direct connection to artificial intelligence (AI) [16,17]. Dialogue systems can be categorized into two classes according to their applications: (1) open-domain dialogue systems that aim at synthesizing human-like conversations with users [18,19,20], and (2) task-oriented dialogue (TOD) systems of which the goal is to help human users to complete certain tasks such as virtual assistant [21,22,23].
In the two main classes of dialogue system, task-oriented dialogue system has been the focus of the research community as it has many valuable applications such as an online booking system, counselling service, and in-vehicle assistant [22,24]. In recent years, significant progress has been made in the development of task-oriented dialogue systems beginning with the deployment of neural models, especially pretrained language models (PLM) [25]. Task-oriented dialogue systems based on neural networks (especially PLMs-based models) trained on relevant dataset have shown superior performance on capturing user’s intent and generating proper response [26,27,28,29]. However, such success heavily relies on the availability of relevant datasets such as MultiWOZ, KVRET and BANKING77 [24,30,31], which usually needs rich human resources to construct because the annotation of user intent and response requires interaction with human annotators. Therefore, such a dataset is expensive to obtain, especially when the amount of required data is large, as neural models typically need more training data to achieve competitive performance. There is no doubt that the large amount of resource needed to build such datasets has undesirably hindered the development of task-oriented systems. To address this problem, we propose to adopt the idea of using a simulator as a human user whose potential applications include evaluating a task-oriented dialogue system or augmenting data to alleviate the shortage of datasets [32,33]. The primary purpose of user simulator is to have a dialogue model acting as the human user, which then can be used to converse with the task-dialogue system and thus automate the evaluation of such systems without requiring human evaluators [34,35,36].
Current work of simulators in the user side generally involves agenda-based simulation, which entirely relies on handcrafted rules [32,37]. Such methods are deemed too costly to be practical and suffer from a lack of generalization. In addition, some research proposes end-to-end user simulators [35,38] as an approach to conducting reinforcement learning, and these simulators are expected to serve TOD systems rather than act as practical standalone simulators.
In this paper, we propose in-vehicle conversational driver simulator (IvCDS), an end-to-end PLM-based simulator that can play the role of human drivers during the interaction with an assistant in the task-oriented dialogue task. First, we process KVRET, a TOD dataset consisting of conversations between a human driver and an in-car assistant, to obtain an appropriate dataset for the training and inference of a driver simulator. The contents in KVRET are re-labeled by using existing information, and the original driver-assistant conversations are converted into the assistant-driver format. Then, we train IvCDS with the training objective of language modelling on processed KVRET. Finally, we investigate the performance of IvCDS on three distinct tasks: NLU, POL, and NLG, which will be introduced in Section 2 on the test set of KVRET. We demonstrate that the proposed IvCDS can outperform existing state-of-the-art (SOTA) models on aforementioned tasks. Furthermore, we conduct ablation studies to investigate the importance of different components of IvCDS.

Paper Structure

To provide a clear picture of this paper, we will briefly introduce the overall paper structure in this section. Section 2 reviews the background and related research of the TOD task, as well as its three subtasks. In addition, Section 3 introduces the approach to processing the dataset we used, and gives the detailed methodology of our driver simulator. Next, Section 4 compares the performances of our driver simulator and other PLM-based baselines, and reports the results of the ablation study. Finally, Section 5 depicts the summary of this paper, and provides the discussion about potential research directions and practical applications in the future.

2. Related Work: Task-Oriented Dialogue Systems

In general, a task-oriented dialogue system [21,22,23,39,40,41] adopts one of the following architectures: a conventional pipeline system or an end-to-end neural model [42]. The pipeline scheme is a modular architecture consisting of four components: natural language understanding (NLU), dialog state tracker (DST), dialog policy (POL), and natural language generation (NLG). Traditionally, models for different components are separately trained and evaluated on task-specific datasets [43]. For example, BANKING77 provides a dataset composed of paired utterances and intents which is suitable for NLU and NLG tasks [31]. Meanwhile, a recent trend is to jointly train pipeline modules in an end-to-end manner instead of combining models that are separately trained. Note that DST is not involved in a user simulator [44], because the user policy is directly generated according to given assistant actions without the belief state.

2.1. Natural Language Understanding

Natural language understanding (NLU) aims to understand the intents and actions of an utterance. It can portray conversational actions as a set of intents and slot values. Intents are expressions of the reason why the speaker issued the sentence, such as queries and notifications, and are the utterance’s slot-values are specific to the task and content mentioned in the utterance. Based on the structure of conversational behaviour, NLU can be divided into two parts, intent detection and slot-value extraction. Its tasks are based on a series of tokens as input to the model, and RNNs and their variants are powerful for handling such sequence modelling and are widely used for intent detection and slot-value extraction [45]. In addition, the recent pretrained model BERT is another popular choice [46]. In case of a driver simulator, NLU focuses on understanding the intents behind the speaking of an assistant.

2.2. Dialog Policy

For a user–agent dialogue, the input to dialog policy (POL) is the current conversation state consisting of slot-value pairs representing the user’s intents [42,47,48,49], and it generates the system actions. A typical approach is to train a dialog policy based on a conversation corpus by using supervised or simulated learning, and then fine tune the model by using reinforcement learning [50,51,52,53], because it would be too costly to be practical if human users (e.g., experts or volunteers) are involved. In terms of a driver simulator, POL aims to generate actions of a driver according to assistant actions.

2.3. Natural Language Generation

Natural language generation (NLG) maps POL-generated dialogue acts to textual sentences, and is often modelled as a conditional language generation task [54,55,56]. It receives a set of behaviours as input, and generates a textual response as output in the form of meaningful and fluent natural language. The generated response is expected to properly understand the user intents and to give appropriate actions from the user perspective. For a driver simulator, NLG operates as a human driver in response to an assistant with appropriate textual utterance.

3. Methodology

In this section, we introduce how we convert the KVRET dataset into the appropriate shape, including data cleaning, labelling and formatting. We additionally provide the methodology of the driver simulator.

3.1. Data

KVRET is a task-oriented dialogue dataset consisting of the conversations between a driver and an in-vehicle assistant [24]. It is a multidomain multiturn dataset which involves three scenarios: calendar scheduling, weather information retrieval, and point-of-interest (POI) navigation. These distinct scenarios aim to enhance the diversity of driver-assistant conversations for promoting the development of an experienced and comprehensive personal vehicular assistant. Similar to other task-oriented dialogue datasets such as MultiWOZ [30] and RiSAWOZ [57], KVRET likewise utilizes the Wizard-of-Oz scheme [58] as the approach of collecting conversations and labels. However, because the original dataset mainly focuses on the assistant side, some labels that are essential for a driver simulator (e.g., driver actions) are unfortunately missing. We need to extract the actions of the driver and assistant in each conversation as an important feature to train the user simulator.

3.1.1. Assistant Actions

A driver simulator is expected to have the ability of understanding the intent and actions behind the utterance from the assistant. Assistant actions are indispensable to a driver simulator for both NLU and POL tasks.
Despite the unavailability of such labels in the original KVRET dataset, we leverage the provided information to label the actions of assistant utterances in the correct form we need. For each conversation, KVRET gives a knowledge base containing a set of key-value pairs which describe all possible actions that an assistant can take. In addition, each assistant utterance in a conversation is accompanied by the <requested> label, which contains all keys that can be used to locate values in the knowledge base. When investigating the the original dataset, we found that the types and manners of actions can vary according to different scenarios, resulting in different extraction strategies. For calendar scheduling and POI navigation, the <requested> label generally involves location, time and events. In this case, the corresponding actions simply depends on the matched items between the knowledge base and the the <requested> label. Figure 1 gives an example of labeling assistant actions on the POI navigation scenario. By matching the keys (i.e., poi_type and distance in <requested>) and values (i.e., 5 miles and chinese restaurant in the assistant utterance) in the knowledge base, we can therefore determine that the assistant action should be [distance] = 5 miles and [poi_type] = chinese restaurant.
For scenarios regarding weather information retrieval however, the direct matching strategy is inappropriate because the knowledge base has a different format. The keys in it are the day of a week (i.e., Monday to Sunday), and the corresponding value of each key contains weather attributes and the minimum and maximum temperatures. For slots that additionally contain the date information, the date will be used to match the key in the knowledge base to determine the actions. Otherwise, each group of temperature information will be divided into three parts: weather attribute, lowest temperature, and highest temperature. The detailed matching process is illustrated in Algorithm 1, where KB is the knowledge base.
Algorithm 1: The processing step of extracting assistant actions for the weather information retrieval scenario.
  • S a set of key-value pairs in <slots> ▹ weather_attribute = blizzard, date = Friday, ⋯
  • S k e y s all keys in S
  • S [ k e y ] the value of key in S
  • K i the i-th item in KB
  • K k e y s all keys in KB                   ▹ monday, tuesday, location,⋯
  • K i [ k e y ] the value of key in i-th item in KB     ▹ "snow, low of 60F, high of 80F", ⋯
  • a the answer of assistant      ▹ On Friday it’s gonna rain in Brentwood, ⋯
  • a c t i o n s the actions of assistant
  • if  l o c a t i o n S k e y s then
  •      a c t i o n s a question about location
  • else
  •     get K i where S [ k e y ] in K i
  •     if  d a t e S k e y s  then
  •          a c t i o n s l o c a t i o n = K i [ l o c a t i o n ] , d a t e = K i [ d a t e ]
  •     else
  •         split K i [ k e y ] into three parts
  •         if  a n y p a r t s a  then
  •             a c t i o n s l o c a t i o n = K i [ l o c a t i o n ] , k e y = K i [ k e y ]
  •         end if
  •     end if
  • end if

3.1.2. Driver Actions

Driver actions indicate the actual objectives beneath the driver utterance. In this experiment, driver actions are necessary for the POL and NLG task. Although no label is available on the driver side because KVRET mainly focuses on the assistant side, the driver actions can be found in the <slot> from the assistant labels which is provided for the NLU task in the original dataset. We take the content of <slot> in each assistant response as the actions of its previous driver utterance. For example in Figure 1, the action of driver utterance “Show me the closest location where I can get Chinese food” are [distance] = closet and [poi_type] = chinese restaurant from the original <slot> label of its subsequent assistant response. We additionally add greeting labels, such as [greeting] = thank for driver utterance “Thank you”.

3.1.3. Driver Profile

For the driver simulator, we use the driver profile to indicate the overall objective of the driver’s interaction with the in-vehicle assistant, including all intents and actions of the driver throughout the entire conversation. For the training and inference in our experiment, each driver profile is comprised of all driver actions in a conversation. The driver profile performs as a pool of potential actions which can help to determine the next actions on the POL task. Meanwhile, for practical application in the future, a driver profile can be generated by extracting potential actions from the knowledge base. By changing the driver profile, a driver simulator is expected to have different behaviours.

3.1.4. Reordering

The original dataset consists of driver–assistant conversation, meaning that a conversation is always started by the driver. We convert conversations into an assistant–driver format as we expect each turn of a conversation can have an input utterance from the assistant to guarantee the exact NLU-POL-NLG structure. Specifically, the driver data at ith turn together with the assistant data at ( i 1 )th turn composes the new ith assistant-driver turn in the processed dataset. In particular, the driver data at the first turn will be accompanied by the assistant data whose utterance and actions are empty, to constitute the new first turn. Additionally, the assistant data at the last turn will be discarded as no corresponding driver data is available. Figure 2 provides an example.

3.2. Model

Figure 3 illustrates the overall process of the driver simulator when conversing with an in-vehicle assistant, with regards to three tasks: NLU, POL, and NLG. In general, the driver simulator first receives an utterance from the assistant, and the assistant action, the driver action and driver utterance are sequentially generated. We will then introduce the means of the driver simulator to handle these tasks in detail.

3.2.1. NLU

The NLU task expects the driver simulator to produce the actions according to a given assistant utterance. Given an assistant utterance A U t at turn t together with the driver profile D P and the dialogue history H t , Equation (1) describes how the simulator generates the corresponding actions on the NLU task as follows:
A A t = NLU M ( [ D P ; H t ; A U t ] ) ,
where A A t means the generated assistant actions, M represents the model of driver simulator, and the dialogue history H t = [ C 1 ; C 2 ; ; C t 1 ] is made up of the conversation contents C i of all previous turns. The conversation content C i at turn i is composed of the concatenation C i = [ A U i ; A A i ; D A i ; D U i ] where A U = assistant utterance, A A = assistant action, D A = driver action D A i and D U = driver utterance.
For example in Figure 3 at turn 2, the assistant says “I have a couple gas stations listed. Which one would you like to know about?”, and the NLU task expects the simulator to generate the action as [poi_type] = gas_station.

3.2.2. Pol

After confirming the actions behind an assistant utterance on the NLU task, POL then requires the driver simulator to take corresponding actions. At turn t, the driver action ( D A ) is generated as described in Equation (2):
D A t = POL M ( [ D P ; H t ; A U t ; A A t ] ) ,
where A A t is the output of Equation (1).
As shown in Figure 3, the driver profile provides all potential actions that a driver can take, and the simulator then generates the action [distance] = shortest distance in response to the assistant action that is generated from previous NLU task after referring to the given driver profile.

3.2.3. Nlg

NLG is the task which translates the generated action into the utterance in the form of natural language. The driver simulator M will produce the driver utterance at turn t according to Equation (3), combining with the outputs from previous two tasks:
D U t = NLG M ( [ D P ; H t ; A U t ; A A t ; D A t ] ) ,
where A A t and D A t are the output of Equation (1) and Equation (2), respectively.
For example, the simulator is capable of converting the corresponding driver action [distance] = shortest distance into the textual sentence “Pick the quickest one to reach, the one in the shortest distance”, as shown in Figure 3.

3.2.4. Training and Inference

IvCDS is based on GPT-2 [59], a unidirectional language model which consists of a transformer decoder and has been pre-trained on large-scale datasets, therefore training IvCDS is actually fine tuning the pretrained GPT-2. Meanwhile, the training objective of IvCDS is the objective of a language model, needing no additional objectives such as next sequence prediction. Given a training sequence S = { w 1 , w 2 , w 3 , , w n } , it is expected to learn next word prediction by maximizing the likelihood of the next token as Equation (4):
L = i = 1 n l o g P ( w i | { w 1 , w 2 , , w i 1 } )
where w i is the i-th word/token in the training sequence S, and P is the probability of a token given all its previous tokens.
Figure 4 illustrates the training of IvCDS. In addition, it requires IvCDS to have the ability of stopping token prediction at a proper time for each task. Hence, we wrap the constituents in the input sequence with different special tokens to indicate the category of a constituent and to instruct IvCDS to distinguish the current task. Specifically, five categories of constituents are used, as shown in Figure 4 with different colors: driver profile, assistant utterance, assistant action, driver action, and driver utterance. The beginning and ending of a constituent are indicated by “[so x x ]” (start of x x ) and “[eo x x ]” (end of x x ), respectively. Special tokens for these constituents are introduced as follows:
driver profile:[sodp]and[eodp]
assistant utterance:[soau]and[eoau]
assistant action:[soaa]and[eoaa]
driver action:[soda]and[eoda]
driver utterance:[sodu]and[eodu].
Note that, despite the six different colors in Figure 4, the dialogue history in fact consists of driver&assistant utterances and actions from previous turns, and we do not use any specific token for it. We therefore describe five categories of constituents. In addition, Figure 5 gives an example input sequence from the processed KVRET training set for the training of IvCDS.
When the training step is completed, we believe that IvCDS achieves the abilities of understanding utterances from the assistant and makes the decision of actions to take using the history of dialogue and pre-assigned driver profile. Subsequently, the trained IvCDS can be evaluated according to the results of its inference on the test set. Figure 6 shows the inference process on different tasks, where the input sequence for each task consists of necessary constituents and the driver simulator is expected to sequentially generating tokens until the termination condition is achieved. For example, once IvCDS outputs the special token “[eoda]”, the token prediction stops during the POL task. In addition, prediction will be terminated as well if the length of generated sequence reaches a predefined maximum, no matter whether the task-specific token is generated.

4. Experiment

To encourage reproduction of our work, we introduce how the experiment of IvCDS is carried out in detail in this section, including the involved baseline models, the hyperparameters of training IvCDS and baselines, and the evaluation methods for different tasks. In addition, we compare the performances of IvCDS and other baselines by reporting their results on three tasks. Furthermore, we conduct the ablation study to investigate the influence of dialogue history and driver profile on IvCDS.

4.1. Experiment Setup

In this experiment, the training and inference entirely depends on one NVIDIA GeForce RTX 3090 graphics card. For the training parameters of IvCDS, we set the learning rate as 5e-5, the batch size as 4, and the number of epochs as 40. The pretrained model and tokenizer follows the standard implement of GPT-2 model from HuggingFace’s Transformers [60] with 32 additional special tokens, including 10 separator tokens like “[sodp]” and “[eoau]”, and 22 action key tokens such as “[poi]” and “[poi_type]” (see Figure 5). In terms of inference, we limit the maximal length of candidate outputs as [ length of input ] + 80 , and task-specific termination tokens are: “[eoaa]” for NLU, `[eoda]” for POL, and “[eodu]” for NLG.

4.1.1. Baseline Models

To investigate the performance of IvCDS, we additionally include eight models as baselines for comparison such as BERT [61], BART [62], ProphetNet [63], PEGASUS [64], T5 [65], etc. Recently, these baselines have successfully achieved state-of-the-art performance on various NLP-related tasks [66,67,68,69,70] whose pretrained model weights are publicly available on HuggingFace’s Transformers [60] as well. Table 1 gives a brief review of these baselines.
The aforementioned baselines are pretrained models and are fine tuned on the processed KVRET training set for each task. The learning rate is likewise 5e-5, and each baseline is trained with at least 20 epochs to minimize the training and validation loss. The batch size varies according to both the parameter size of baselines and payload of the graphics card, ranging from 2 to 32. The action key tokens such as [poi] are added to these baselines as well for fair comparison.

4.1.2. Evaluation Metrics

For NLU and POL tasks, since the outputs are actions in the form of key-value pairs, we use precision, recall, and f-measure for their evaluation. Items, namely key-value pairs, in the outputs for NLU and POL tasks are compared with the items in the references on test set at turn-level. Strict exact match is applied, resulting in penalization of redundancy items.
For the NLG task, the generated sentences are evaluated by word overlap-based evaluation metrics, which are commonly applied for assessing the quality of textual sequences in generative NLP tasks, such as question generation [75] and open-domain dialogue systems [20]. Four metrics as follows are employed:
  • BLEU [76]: The bilingual evaluation understudy (BLEU) is an evaluation metric which assesses the quality of a generated candidate by the precision of the n-grams between it and its corresponding references. In this experiment, we computed the BLEU-4 score, which uses equally weighed 1∼4 grams.
  • GLEU [77]: Google-BLEU (GLEU) is a variety of BLEU. Instead of the standalone precision, GLEU computes the precision and recall of n-gram between a candidate and a reference. The minimum between precision and recall is then uses as the GLEU score, and n is typically chosen as 4.
  • ROUGE-L [78]: ROUGE-L is the widely applied variant of recall-oriented understudy for gisting evaluation (ROUGE), a recall-adapted version of BLEU, wherein L denotes to the longest common subsequence (LCS). It computes the precision and recall using the LCS between a candidate and a reference, rather than n-gram.
  • METEOR [79]: Metric for evaluation of translation with explicit ordering (METEOR) is the evaluation metric that was initially proposed to remedy the known weaknesses of BLEU (e.g., BLEU does not consider recall and performs inaccurately when evaluating at the sentence level). In addition to exact match, METEOR uses other strategies such as synonyms mapping to match the uni-gram between a candidate and a reference, and the METEOR score is computed by the precision and recall.
To achieve a fair comparison, the following preprocessing strategies are applied on raw model outputs before their evaluation: (1) separator tokens such as [eoau] and [eodp], and punctuation are discarded; (2) model-specific special tokens are removed; (3) outputs are lower-cased.

4.2. System Performances

Table 2 reports the results of our proposed driver simulator and other baseline models on the three tasks: NLU, POL, and NLG, where these models are evaluated based on their outputs on the processed KVRET test set. In general, we find that IvCDS successfully outperform other models on all tasks according to most evaluation metrics.
First, IvCDS has the best performance on the NLU task, and the gap of F1 scores between it and the second-ranked model, namely ProphetNet, achieves 4. It denotes that IvCDS properly plays the role of a human driver that can felicitously realize the intent behind the utterances from an in-vehicle assistant. In addition, an interesting observation is that, most baselines can have a relatively high level of recall, despite the low precision and f1 scores. This may imply that, these models tend to generate a large number of predictions which are maximally capable of covering the items in references; nonetheless, most of the predicted items are incorrect. Such a situation is obvious in terms of BERT2BERT, as its recall reaches as high as 87 with an extremely low precision of less than 20.
With respect to the POL task, we find that IvCDS still achieves the highest precision and F1 score, but the recall is slightly lower than BART-large. The gap of F1 scores between IvCDS and the second-ranked Pegasus is more than 9, whereas the gap of recall between IvCDS and BART-large is only about 2. Similar to results in the NLU task, the high-recall low-precision situation appears again on these baseline models. We additionally find that, models, including IvCDS and most baselines, perform better on NLU than POL, whereas the probable reason may be the POL task requires the additional ability of making decision by retrieving useful information from the driver profile. Meanwhile, an opposite trend appears on Pegasus and BigBird, and such models probably do well in understanding structured information like the assistant actions and driver profile, but lack the ability of understanding text in natural language.
For NLG task, IvCDS is the best among all models according to BLEU-4, ROUGE-L and GLEU metrics but slightly worse than BART-large on METEOR scores. Because METEOR is the only metric that relies on uni-grams, which is generally a single token, we think this implies BART-large is prone to produce correct tokens; however, the token order may differ from the reference, resulting in lower scores of evaluation metrics which depend on n-grams ( n > 1 ) or LCS.

4.3. Ablation Experiment

Ablation study is proposed to investigate how important a certain part of a neural network is [80]. In our experiment, we concatenate driver profile and dialogue history, with the task-essential input components (e.g., assistant utterance for NLU) as the input sequence, and we would like to determine the degree to which these two components can have an influence on the performance of IvCDS. Table 3 reports the results of the ablation experiment, where training means whether a model is trained with specific components in training sequences, inference means whether the components are included in input sequences, and H and D P are dialogue history and driver profile, respectively.

4.3.1. Ablated Training & Inference

The first three models fine tune GPT-2 by using the same training objective as described in Section 3.2.4, and the training sequences follow a structure similar to Figure 5 with different combinations of H and D P . The input sequences use the same combination as training when inferring on the test set. Compared with the original IvCDS (O-IvCDS), namely the last model in Table 3. We find that these IvCDS models with ablated training and inference (A-IvCDS T & I ) appear with varying degrees of performance reduction on different tasks.
For the NLU task, the F1 score of these models can still achieve at least 70, and we think this indicates that A-IvCDS T & I somewhat gains the ability of understanding the intents behind assistant utterances without D P and H. In addition, an interesting observation is that, the A-IvCDS T & I without H and D P performs better than those which contain either H or D P . We think the potential explanation could be that, the standalone use of one such component may import extra disturbance items during training, resulting in the model’s failure of utilizing implicit information in the component during inference.
Meanwhile, their POL performance drops to as low as around 30 in terms of F1 scores, meaning that these models failed to make the decision of a driver to respond to the assistant. Moreover, A-IvCDS T & I with D P can have a better performance than those without D P , which denotes that, the driver profile which is expected to operate as the knowledge base, can have a positive influence on the POL task
In terms of NLU, all models seem to perform faultily with an evident decrease of word overlap-based evaluation metric scores, compared to O-IvCDS, revealing that training without the joint use of H and D P will negatively affect models’ ability of producing meaningful and fluent sentences in natural language.

4.3.2. Sole Ablated Inference

The fourth to sixth models in Table 3, called A-IvCDS I , represent the ablated IvCDS models which are in fact O-IvCDS but with ablated inference using varying combinations of H and D P .
For NLU and NLG, we can observe a decrease on metric scores but they are barely less than O-IvCDS, whereas A-IvCDS I with D P even shows an increase on METEOR scores. Among them, the model with neither H nor D P shows the worst performance. Therefore, we think the information in these two components still help to detect assistant intents and generate meaningful responses during inference, although A-IvCDS I has gained corresponding ability during training.
With respect to POL, despite the decrease of scores in general, there exist striking levels of difference of A-IvCDS I with varying component combinations. Models without D P , regardless of the inclusion of H, perform badly at F 1 < 30 , whereas the one with D P achieves F 1 76 . We think this precisely reveals the importance of driver profiles on the POL task, which additionally emphasizes our contribution of proposing the processed KVRET dataset because it can provide the necessary driver profile for a driver simulator.

5. Conclusions and Future Work

In this paper, we focus on proposing a simulator which can act as a human driver having the abilities of assistant intent perception, making decision and generating appropriate and fluent responses. Two main contributions we made in this paper are: (1) we processed the KVRET dataset into the appropriate format for a driver simulator; (2) we proposed IvCDS, a PLM-based driver simulator. We demonstrated that IvCDS successfully achieves the best performance on all three tasks—NLU, POL, and NLG—compared with existing SOTA models. We subsequently carried out an ablation experiment, the results of which show that the joint use of dialogue history and driver profile can positively affect the performance of IvCDS, whereas the latter plays an important role in improving the performance of a driver simulator on the POL task.
For research in the future, we would first like to further investigate the reason why models trained without any dialogue history or driver profile outperform others which contains one of them according the results of ablation study. Because we think this may imply potential noise in our processed dataset, we will focus on filtering out the undiscovered labeling errors in it by utilizing heuristic algorithms or by expert annotators. In addition, a potential direction of improving IvCDS is to increase its recall in POL since it was found to be slightly lower than BART-large. We think error analysis would be useful for achieving such a goal. Moreover, we are encouraged to examine or adapt this driver simulator on more relevant TOD datasets in the future.
In terms of practical applications of driver simulator, we would like to use IvCDS to interact with in-vehicle assistant systems for evaluation, and this can help to find their weaknesses according to the results of interaction. We believe the improvement of the performance of in-vehicle assistant systems can additionally promote the development of intelligent driving techniques. In addition to the importance of scientific research, the driver simulator is important for the education of all road users as well. A driver simulator can helping to solve transportation-based issues in terms of traffic environment and road safety, by practicing their abilities, skills, and knowledge about traffic in a safe environment, for themselves and other traffic participants.

Author Contributions

Methodology, T.J.; Software, P.C. and C.L.; Resources, P.C. and W.B.; Data curation, X.Y.; Writing—original draft, T.J., X.Y. and C.L.; Writing—review & editing, T.J., L.Z. and C.L.; Supervision, W.B.; Project administration, S.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Science Foundation Ireland in the ADAPT Centre for Digital Content Technology at Dublin City University funded under the SFI Research Centres Programme co-funded under the European Regional Development Fund grant number 13/RC/2106_P2 and 13/RC/2106, and by Science Foundation Ireland through the SFI Centre for Research Training in Machine Learning grant number 18/CRT/6183.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Publicly available datasets were analyzed in this study. This data can be found here: https://github.com/TianboJi/ivcds (accessed on 13 October 2022).

Acknowledgments

We would like to express our warm thanks to all editors and reviewers for their help in improving the quality of this manuscript.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Oviedo-Trespalacios, O.; Truelove, V.; Watson, B.; Hinton, J.A. The impact of road advertising signs on driver behaviour and implications for road safety: A critical systematic review. Transp. Res. Part A Policy Pract. 2019, 122, 85–98. [Google Scholar] [CrossRef]
  2. Cvahte Ojsteršek, T.; Topolšek, D. Influence of drivers’ visual and cognitive attention on their perception of changes in the traffic environment. Eur. Transp. Res. Rev. 2019, 11, 45. [Google Scholar] [CrossRef]
  3. Cao, Z.; Ceder, A.A.; Zhang, S. Real-time schedule adjustments for autonomous public transport vehicles. Transp. Res. Part C Emerg. Technol. 2019, 109, 60–78. [Google Scholar] [CrossRef]
  4. Cao, Z.; Ceder, A.A. Autonomous shuttle bus service timetabling and vehicle scheduling using skip-stop tactic. Transp. Res. Part C Emerg. Technol. 2019, 102, 370–395. [Google Scholar] [CrossRef]
  5. Hailin, W.; Hanhui, L.; Zhumei, S. Fatigue Driving Detection System Design Based on Driving Behavior. In Proceedings of the 2010 International Conference on Optoelectronics and Image Processing, Washington, DC, USA, 11–12 November 2010; Volume 1, pp. 549–552. [Google Scholar] [CrossRef]
  6. Stern, H.S.; Blower, D.; Cohen, M.L.; Czeisler, C.A.; Dinges, D.F.; Greenhouse, J.B.; Guo, F.; Hanowski, R.J.; Hartenbaum, N.P.; Krueger, G.P.; et al. Data and methods for studying commercial motor vehicle driver fatigue, highway safety and long-term driver health. Accid. Anal. Prev. 2019, 126, 37–42. [Google Scholar] [CrossRef] [PubMed]
  7. Fagnant, D.J.; Kockelman, K. Preparing a nation for autonomous vehicles: Opportunities, barriers and policy recommendations. Transp. Res. Part A Policy Pract. 2015, 77, 167–181. [Google Scholar] [CrossRef]
  8. Cao, Z.; Zhang, S.; Ceder, A.A. Novel coupling–decoupling strategy for scheduling autonomous public transport vehicles in overcrowded corridors. Appl. Math. Model. 2022, 106, 299–324. [Google Scholar] [CrossRef]
  9. Xing, Y.; Lv, C.; Cao, D.; Velenis, E. Multi-scale driver behavior modeling based on deep spatial-temporal representation for intelligent vehicles. Transp. Res. Part C Emerg. Technol. 2021, 130, 103288. [Google Scholar] [CrossRef]
  10. Hasenjäger, M.; Heckmann, M.; Wersing, H. A Survey of Personalization for Advanced Driver Assistance Systems. IEEE Trans. Intell. Veh. 2020, 5, 335–344. [Google Scholar] [CrossRef]
  11. Yi, D.; Su, J.; Liu, C.; Quddus, M.; Chen, W.H. A machine learning based personalized system for driving state recognition. Transp. Res. Part C Emerg. Technol. 2019, 105, 241–261. [Google Scholar] [CrossRef]
  12. Gouribhatla, R.; Pulugurtha, S.S. Drivers’ behavior when driving vehicles with or without advanced driver assistance systems: A driver simulator-based study. Transp. Res. Interdiscip. Perspect. 2022, 13, 100545. [Google Scholar] [CrossRef]
  13. Feinauer, S.; Schuller, L.; Groh, I.; Huestegge, L.; Petzoldt, T. The potential of gamification for user education in partial and conditional driving automation: A driving simulator study. Transp. Res. Part F Traffic Psychol. Behav. 2022, 90, 252–268. [Google Scholar] [CrossRef]
  14. Yan, W.; Wong, S.; Loo, B.P.; Wu, C.Y.; Huang, H.; Pei, X.; Meng, F. An assessment of the effect of green signal countdown timers on drivers’ behavior and on road safety at intersections, based on driving simulator experiments and naturalistic observation studies. J. Saf. Res. 2022, 82, 1–12. [Google Scholar] [CrossRef] [PubMed]
  15. Simović, S.; Ivanišević, T.; Trifunović, A.; Čičević, S.; Taranović, D. What Affects the E-Bicycle Speed Perception in the Era of Eco-Sustainable Mobility: A Driving Simulator Study. Sustainability 2021, 13, 5252. [Google Scholar] [CrossRef]
  16. Chen, H.; Liu, X.; Yin, D.; Tang, J. A survey on dialogue systems: Recent advances and new frontiers. ACM SIGKDD Explor. Newsl. 2017, 19, 25–35. [Google Scholar] [CrossRef]
  17. Ni, J.; Young, T.; Pandelea, V.; Xue, F.; Cambria, E. Recent advances in deep learning based dialogue systems: A systematic survey. Artif. Intell. Rev. 2022, 1–101. [Google Scholar] [CrossRef]
  18. Adewumi, T.; Liwicki, F.; Liwicki, M. State-of-the-Art in Open-Domain Conversational AI: A Survey. Information 2022, 13, 298. [Google Scholar] [CrossRef]
  19. Kann, K.; Ebrahimi, A.; Koh, J.; Dudy, S.; Roncone, A. Open-domain Dialogue Generation: What We Can Do, Cannot Do, And Should Do Next. In Proceedings of the 4th Workshop on NLP for Conversational AI, Dublin, Ireland, 26 January 2022; Association for Computational Linguistics: Stroudsburg, PA, USA, 2022; pp. 148–165. [Google Scholar] [CrossRef]
  20. Ji, T.; Graham, Y.; Jones, G.; Lyu, C.; Liu, Q. Achieving Reliable Human Assessment of Open-Domain Dialogue Systems. In Volume 1: Long Papers, Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics, Dublin, Ireland, 22–27 May 2022; Association for Computational Linguistics: Stroudsburg, PA, USA, 2022; pp. 6416–6437. [Google Scholar] [CrossRef]
  21. Louvan, S.; Magnini, B. Recent Neural Methods on Slot Filling and Intent Classification for Task-Oriented Dialogue Systems: A Survey. In Proceedings of the 28th International Conference on Computational Linguistics, Barcelona, Spain, 8–13 December 2020; pp. 480–496. [Google Scholar]
  22. Li, C.; Zhang, X.; Chrysostomou, D.; Yang, H. ToD4IR: A Humanised Task-Oriented Dialogue System for Industrial Robots. IEEE Access 2022, 10, 91631–91649. [Google Scholar] [CrossRef]
  23. Wang, W.; Zhang, Z.; Guo, J.; Dai, Y.; Chen, B.; Luo, W. Task-Oriented Dialogue System as Natural Language Generation. In Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval, Madrid, Spain, 11–15 July 2022; pp. 2698–2703. [Google Scholar]
  24. Eric, M.; Krishnan, L.; Charette, F.; Manning, C.D. Key-Value Retrieval Networks for Task-Oriented Dialogue. In Proceedings of the 18th Annual SIGdial Meeting on Discourse and Dialogue, Saarbrücken, Germany, 15–17 August 2017; Association for Computational Linguistics: Stroudsburg, PA, USA, 2017; pp. 37–49. [Google Scholar] [CrossRef] [Green Version]
  25. Budzianowski, P.; Vulić, I. Hello, It’s GPT-2-How Can I Help You? Towards the Use of Pretrained Language Models for Task-Oriented Dialogue Systems. In Proceedings of the 3rd Workshop on Neural Generation and Translation, Hong Kong, China, 4 November 2019; Association for Computational Linguistics: Stroudsburg, PA, USA, 2019; pp. 15–22. [Google Scholar] [CrossRef] [Green Version]
  26. Yin, Y.; Shang, L.; Jiang, X.; Chen, X.; Liu, Q. Dialog state tracking with reinforced data augmentation. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 9474–9481. [Google Scholar]
  27. Madotto, A.; Lin, Z.; Zhou, Z.; Moon, S.; Crook, P.; Liu, B.; Yu, Z.; Cho, E.; Wang, Z. Continual learning in task-oriented dialogue systems. arXiv 2020, arXiv:2012.15504. [Google Scholar]
  28. Mi, F.; Wang, Y.; Li, Y. Cins: Comprehensive instruction for few-shot learning in task-oriented dialog systems. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtually, 22 February–1 March 2022; Volume 36, pp. 11076–11084. [Google Scholar]
  29. Mi, F.; Li, Y.; Zeng, Y.; Zhou, J.; Wang, Y.; Xu, C.; Shang, L.; Jiang, X.; Zhao, S.; Liu, Q. PANGUBOT: Efficient Generative Dialogue Pre-training from Pre-trained Language Model. arXiv 2022, arXiv:2203.17090. [Google Scholar]
  30. Budzianowski, P.; Wen, T.H.; Tseng, B.H.; Casanueva, I.; Ultes, S.; Ramadan, O.; Gašić, M. MultiWOZ—A Large-Scale Multi-Domain Wizard-of-Oz Dataset for Task-Oriented Dialogue Modelling. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Brussels, Belgium, 31 October–4 November 2018; Association for Computational Linguistics: Stroudsburg, PA, USA, 2018; pp. 5016–5026. [Google Scholar] [CrossRef] [Green Version]
  31. Casanueva, I.; Temčinas, T.; Gerz, D.; Henderson, M.; Vulić, I. Efficient Intent Detection with Dual Sentence Encoders. In Proceedings of the 2nd Workshop on Natural Language Processing for Conversational AI, Online, 9 July 2020; Association for Computational Linguistics: Stroudsburg, PA, USA, 2020; pp. 38–45. [Google Scholar] [CrossRef]
  32. Li, X.; Lipton, Z.C.; Dhingra, B.; Li, L.; Gao, J.; Chen, Y.N. A user simulator for task-completion dialogues. arXiv 2016, arXiv:1612.05688. [Google Scholar]
  33. Tseng, B.H.; Dai, Y.; Kreyssig, F.; Byrne, B. Transferable Dialogue Systems and User Simulators. In Volume 1: Long Papers, Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, Virtual Event, 1–6 August 2021; Association for Computational Linguistics: Stroudsburg, PA, USA, 2021; pp. 152–166. [Google Scholar] [CrossRef]
  34. El Asri, L.; He, J.; Suleman, K. A Sequence-to-Sequence Model for User Simulation in Spoken Dialogue Systems. Interspeech 2016 2016, 1151–1155. [Google Scholar] [CrossRef] [Green Version]
  35. Shi, W.; Qian, K.; Wang, X.; Yu, Z. How to Build User Simulators to Train RL-based Dialog Systems. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), Hong Kong, China, 3–7 November 2019; Association for Computational Linguistics: Stroudsburg, PA, USA, 2019; pp. 1990–2000. [Google Scholar] [CrossRef] [Green Version]
  36. Furlan, R.; Gatti, M.; Menè, R.; Shiffer, D.; Marchiori, C.; Levra, A.G.; Saturnino, V.; Brunetta, E.; Dipaola, F. A natural language processing–based virtual patient simulator and intelligent tutoring system for the clinical diagnostic process: Simulator development and case study. JMIR Med. Inform. 2021, 9, e24073. [Google Scholar] [CrossRef] [PubMed]
  37. Schatzmann, J.; Thomson, B.; Weilhammer, K.; Ye, H.; Young, S. Agenda-Based User Simulation for Bootstrapping a POMDP Dialogue System. In Proceedings of the Human Language Technologies 2007: The Conference of the North American Chapter of the Association for Computational Linguistics, Stroudsburg, PA, USA, 23–25 April 2007; Companion Volume, Short Papers. Association for Computational Linguistics: Stroudsburg, PA, USA, 2007; pp. 149–152. [Google Scholar]
  38. Lee, H.; Jo, S.; Kim, H.; Jung, S.; Kim, T. SUMBT+LaRL: End-to-end Neural Task-oriented Dialog System with Reinforcement Learning. arXiv 2020, arXiv:2009.10447. [Google Scholar]
  39. Lipton, Z.; Li, X.; Gao, J.; Li, L.; Ahmed, F.; Deng, L. Bbq-networks: Efficient exploration in deep reinforcement learning for task-oriented dialogue systems. In Proceedings of the AAAI Conference on Artificial Intelligence, Edmonton, AB, Canada, 13–17 November 2018; Volume 32. [Google Scholar]
  40. Wu, Y.; Li, X.; Liu, J.; Gao, J.; Yang, Y. Switch-based active deep dyna-q: Efficient adaptive planning for task-completion dialogue policy learning. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, Hi, USA, 27 January–1 February 2019; Volume 33, pp. 7289–7296. [Google Scholar]
  41. Li, Y.; Yao, K.; Qin, L.; Che, W.; Li, X.; Liu, T. Slot-consistent NLG for task-oriented dialogue systems with iterative rectification network. In Proceedings of the 58th annual meeting of the association for computational linguistics, Online, 5–10 July 2020; pp. 97–106. [Google Scholar]
  42. Gao, J.; Galley, M.; Li, L. Neural Approaches to Conversational AI. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics: Tutorial Abstracts, Melbourne, Australia, 15–20 July 2018; Association for Computational Linguistics: Stroudsburg, PA, USA, 2018; pp. 2–7. [Google Scholar] [CrossRef]
  43. Peng, B.; Li, C.; Zhang, Z.; Zhu, C.; Li, J.; Gao, J. RADDLE: An Evaluation Benchmark and Analysis Platform for Robust Task-oriented Dialog Systems. In Volume 1: Long Papers, Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, Virtual Event, 1–6 August 2021; Association for Computational Linguistics: Stroudsburg, PA, USA, 2021; pp. 4418–4429. [Google Scholar] [CrossRef]
  44. Zhu, Q.; Zhang, Z.; Fang, Y.; Li, X.; Takanobu, R.; Li, J.; Peng, B.; Gao, J.; Zhu, X.; Huang, M. ConvLab-2: An Open-Source Toolkit for Building, Evaluating, and Diagnosing Dialogue Systems. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics: System Demonstrations, Online, 5–10 July 2020; Association for Computational Linguistics: Stroudsburg, PA, USA, 2020; pp. 142–149. [Google Scholar] [CrossRef]
  45. Yao, K.; Zweig, G.; Hwang, M.Y.; Shi, Y.; Yu, D. Recurrent neural networks for language understanding. In Proceedings of the Interspeech, Lyon, France, 25–29 August 2013; pp. 2524–2528. [Google Scholar]
  46. Chen, Q.; Zhuo, Z.; Wang, W. Bert for joint intent classification and slot filling. arXiv 2019, arXiv:1902.10909. [Google Scholar]
  47. Kreyssig, F.; Casanueva, I.; Budzianowski, P.; Gašić, M. Neural User Simulation for Corpus-based Policy Optimisation of Spoken Dialogue Systems. In Proceedings of the 19th Annual SIGdial Meeting on Discourse and Dialogue, Melbourne, Australia, 12–14 July 2018; Association for Computational Linguistics: Stroudsburg, PA, USA, 2018; pp. 60–69. [Google Scholar] [CrossRef] [Green Version]
  48. Zhang, Z.; Li, X.; Gao, J.; Chen, E. Budgeted Policy Learning for Task-Oriented Dialogue Systems. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, Florence, Italy, 28 July–2 August 2019; Association for Computational Linguistics: Stroudsburg, PA, USA, 2019; pp. 3742–3751. [Google Scholar] [CrossRef] [Green Version]
  49. Wu, C.S.; Hoi, S.C.; Socher, R.; Xiong, C. TOD-BERT: Pre-trained Natural Language Understanding for Task-Oriented Dialogue. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), Online, 16–20 November 2020; Association for Computational Linguistics: Stroudsburg, PA, USA, 2020; pp. 917–929. [Google Scholar] [CrossRef]
  50. Mo, K.; Zhang, Y.; Li, S.; Li, J.; Yang, Q. Personalizing a dialogue system with transfer reinforcement learning. In Proceedings of the AAAI Conference on Artificial Intelligence, Edmonton, AB, Canada, 13–17 November 2018; Volume 32. [Google Scholar]
  51. Liu, B.; Tur, G.; Hakkani-Tur, D.; Shah, P.; Heck, L. Dialogue learning with human teaching and feedback in end-to-end trainable task-oriented dialogue systems. arXiv 2018, arXiv:1804.06512. [Google Scholar]
  52. Shah, P.; Hakkani-Tur, D.; Liu, B.; Tür, G. Bootstrapping a neural conversational agent with dialogue self-play, crowdsourcing and on-line reinforcement learning. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, New Orleans, LA, USA, 1–6 June 2018; Volume 3, (Industry Papers). pp. 41–51. [Google Scholar]
  53. Sun, W.; Zhang, S.; Balog, K.; Ren, Z.; Ren, P.; Chen, Z.; de Rijke, M. Simulating user satisfaction for the evaluation of task-oriented dialogue systems. In Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval, Virtual Event, 11–15 July 2021; pp. 2499–2506. [Google Scholar]
  54. Xu, P.; Hu, Q. An End-to-end Approach for Handling Unknown Slot Values in Dialogue State Tracking. In Volume 1: Long Papers, Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics, Melbourne, Australia, 15–20 July 2018; Association for Computational Linguistics: Stroudsburg, PA, USA, 2018; pp. 1448–1457. [Google Scholar] [CrossRef]
  55. Peng, B.; Li, C.; Li, J.; Shayandeh, S.; Liden, L.; Gao, J. Soloist: Few-shot task-oriented dialog with a single pretrained auto-regressive model. arXiv 2020, arXiv:2005.05298. [Google Scholar]
  56. Peng, B.; Zhu, C.; Li, C.; Li, X.; Li, J.; Zeng, M.; Gao, J. Few-shot Natural Language Generation for Task-Oriented Dialog. In Proceedings of the Findings of the Association for Computational Linguistics: EMNLP 2020, Online Event, 16–20 November 2020; Association for Computational Linguistics: Stroudsburg, PA, USA, 2020; pp. 172–182. [Google Scholar] [CrossRef]
  57. Quan, J.; Zhang, S.; Cao, Q.; Li, Z.; Xiong, D. RiSAWOZ: A Large-Scale Multi-Domain Wizard-of-Oz Dataset with Rich Semantic Annotations for Task-Oriented Dialogue Modeling. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), Online, 16–20 November 2020; Association for Computational Linguistics: Stroudsburg, PA, USA, 2020; pp. 930–940. [Google Scholar] [CrossRef]
  58. Wen, T.H.; Vandyke, D.; Mrkšić, N.; Gašić, M.; Rojas-Barahona, L.M.; Su, P.H.; Ultes, S.; Young, S. A Network-based End-to-End Trainable Task-oriented Dialogue System. In Volume 1, Long Papers, Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics, Online, 19–23 April 2017; Association for Computational Linguistics: Stroudsburg, PA, USA, 2017; pp. 438–449. [Google Scholar]
  59. Radford, A.; Wu, J.; Child, R.; Luan, D.; Amodei, D.; Sutskever, I. Language Models are Unsupervised Multitask Learners. OpenAI Blog 2019, 1, 9. [Google Scholar]
  60. Wolf, T.; Debut, L.; Sanh, V.; Chaumond, J.; Delangue, C.; Moi, A.; Cistac, P.; Rault, T.; Louf, R.; Funtowicz, M.; et al. Transformers: State-of-the-Art Natural Language Processing. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, Online, 16–20 November 2020; Association for Computational Linguistics: Stroudsburg, PA, USA, 2020; pp. 38–45. [Google Scholar] [CrossRef]
  61. Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In Volume 1 (Long and Short Papers), Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Minneapolis, MN, USA, 2–7 June 2019; Association for Computational Linguistics: Stroudsburg, PA, USA, 2019; pp. 4171–4186. [Google Scholar] [CrossRef]
  62. Lewis, M.; Liu, Y.; Goyal, N.; Ghazvininejad, M.; Mohamed, A.; Levy, O.; Stoyanov, V.; Zettlemoyer, L. BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Online, 5–10 July 2020; Association for Computational Linguistics: Stroudsburg, PA, USA, 2020; pp. 7871–7880. [Google Scholar] [CrossRef]
  63. Qi, W.; Yan, Y.; Gong, Y.; Liu, D.; Duan, N.; Chen, J.; Zhang, R.; Zhou, M. ProphetNet: Predicting Future N-gram for Sequence-to-SequencePre-training. In Proceedings of the Findings of the Association for Computational Linguistics: EMNLP 2020, Online Event, 16–20 November 2020; Association for Computational Linguistics: Stroudsburg, PA, USA, 2020; pp. 2401–2410. [Google Scholar] [CrossRef]
  64. Zhang, J.; Zhao, Y.; Saleh, M.; Liu, P.J. PEGASUS: Pre-Training with Extracted Gap-Sentences for Abstractive Summarization. In Proceedings of the 37th International Conference on Machine Learning, Virtual, 13–18 July 2020. [Google Scholar]
  65. Raffel, C.; Shazeer, N.; Roberts, A.; Lee, K.; Narang, S.; Matena, M.; Zhou, Y.; Li, W.; Liu, P.J. Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer. J. Mach. Learn. Res. 2020, 21, 1–67. [Google Scholar]
  66. Qiu, X.; Sun, T.; Xu, Y.; Shao, Y.; Dai, N.; Huang, X. Pre-trained models for natural language processing: A survey. Sci. China Technol. Sci. 2020, 63, 1872–1897. [Google Scholar] [CrossRef]
  67. Lyu, C.; Shang, L.; Graham, Y.; Foster, J.; Jiang, X.; Liu, Q. Improving Unsupervised Question Answering via Summarization-Informed Question Generation. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, Punta Cana, Dominican Republic, 10–11 November 2021; Association for Computational Linguistics: Stroudsburg, PA, USA, 2021; pp. 4134–4148. [Google Scholar] [CrossRef]
  68. Lai, H.; Toral, A.; Nissim, M. Thank you BART! Rewarding Pre-Trained Models Improves Formality Style Transfer. In Volume 2: Short Papers, Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, Online, 1–6 August 2021; Association for Computational Linguistics: Stroudsburg, PA, USA, 2021; pp. 484–494. [Google Scholar] [CrossRef]
  69. Lyu, C.; Foster, J.; Graham, Y. Extending the Scope of Out-of-Domain: Examining QA models in multiple subdomains. In Proceedings of the Third Workshop on Insights from Negative Results in NLP, Dublin, Ireland, 26 May 2022; Association for Computational Linguistics: Stroudsburg, PA, USA, 2022; pp. 24–37. [Google Scholar] [CrossRef]
  70. Zhou, Y.; Portet, F.; Ringeval, F. Effectiveness of French Language Models on Abstractive Dialogue Summarization Task. In Proceedings of the Thirteenth Language Resources and Evaluation Conference, Marseille, France, 20–25 June 2022; European Language Resources Association: Paris, France, 2022; pp. 3571–3581. [Google Scholar]
  71. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.U.; Polosukhin, I. Attention is All you Need. In Advances in Neural Information Processing Systems 30 (NIPS 2017), Proceedings of the Advances in Neural Information Processing Systems, Lake Tahoe, NV, USA, 5–8 December 2013; Guyon, I., Luxburg, U.V., Bengio, S., Wallach, H., Fergus, R., Vishwanathan, S., Garnett, R., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2017; Volume 30, p. 30. [Google Scholar]
  72. Zaheer, M.; Guruganesh, G.; Dubey, K.A.; Ainslie, J.; Alberti, C.; Ontanon, S.; Pham, P.; Ravula, A.; Wang, Q.; Yang, L.; et al. Big bird: Transformers for longer sequences. In Proceedings of the Proceedings of the 34th International Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 6–12 December 2020. [Google Scholar]
  73. Roller, S.; Dinan, E.; Goyal, N.; Ju, D.; Williamson, M.; Liu, Y.; Xu, J.; Ott, M.; Smith, E.M.; Boureau, Y.L.; et al. Recipes for Building an Open-Domain Chatbot. In Main Volume, Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics, Online, 19–23 April 2021; Association for Computational Linguistics: Stroudsburg, PA, USA, 2021; pp. 300–325. [Google Scholar] [CrossRef]
  74. Rothe, S.; Narayan, S.; Severyn, A. Leveraging Pre-trained Checkpoints for Sequence Generation Tasks. Trans. Assoc. Comput. Linguist. 2020, 8, 264–280. [Google Scholar] [CrossRef]
  75. Ji, T.; Lyu, C.; Cao, Z.; Cheng, P. Multi-Hop Question Generation Using Hierarchical Encoding-Decoding and Context Switch Mechanism. Entropy 2021, 23, 1449. [Google Scholar] [CrossRef] [PubMed]
  76. Papineni, K.; Roukos, S.; Ward, T.; Zhu, W.J. Bleu: A Method for Automatic Evaluation of Machine Translation. In Proceedings of the 40th Annual Meeting of the Association for Computational Linguistics, Philadelphia, PA, USA, 6–12 July 2002; Association for Computational Linguistics: Stroudsburg, PA, USA, 2002; pp. 311–318. [Google Scholar] [CrossRef]
  77. Wu, Y.; Schuster, M.; Chen, Z.; Le, Q.V.; Norouzi, M.; Macherey, W.; Krikun, M.; Cao, Y.; Gao, Q.; Macherey, K.; et al. Google’s Neural Machine Translation System: Bridging the Gap between Human and Machine Translation. arXiv 2016, arXiv:1609.08144. [Google Scholar]
  78. Lin, C.Y. ROUGE: A Package for Automatic Evaluation of Summaries. In Proceedings of the Text Summarization Branches Out, Barcelona, Spain, 25–26 July 2004; Association for Computational Linguistics: Stroudsburg, PA, USA, 2004; pp. 74–81. [Google Scholar]
  79. Denkowski, M.; Lavie, A. Meteor 1.3: Automatic metric for reliable optimization and evaluation of machine translation systems. In Proceedings of the Sixth Workshop on Statistical Machine Translation, Online, 10–11 November 2021; Association for Computational Linguistics: Stroudsburg, PA, USA, 2011; pp. 85–91. [Google Scholar]
  80. Meyes, R.; Lu, M.; de Puiseau, C.W.; Meisen, T. Ablation Studies in Artificial Neural Networks. arXiv 2019, arXiv:1901.08644. [Google Scholar]
Figure 1. The overview process of labeling the driver/assistant actions using the given utterances and original labels in the KVRET dataset.
Figure 1. The overview process of labeling the driver/assistant actions using the given utterances and original labels in the KVRET dataset.
Ijerph 19 15493 g001
Figure 2. An example where an original conversation is converted into the assistant-driver format.
Figure 2. An example where an original conversation is converted into the assistant-driver format.
Ijerph 19 15493 g002
Figure 3. The overall structure of how the driver simulator interact with an in-vehicle assistant.
Figure 3. The overall structure of how the driver simulator interact with an in-vehicle assistant.
Ijerph 19 15493 g003
Figure 4. The process of training the simulator which learns to generate the next token according to previous tokens.
Figure 4. The process of training the simulator which learns to generate the next token according to previous tokens.
Ijerph 19 15493 g004
Figure 5. An example from the dataset where colored contents indicates different categories of constituents.
Figure 5. An example from the dataset where colored contents indicates different categories of constituents.
Ijerph 19 15493 g005
Figure 6. The process of inferring on the three tasks after the model training is completed.
Figure 6. The process of inferring on the three tasks after the model training is completed.
Ijerph 19 15493 g006
Table 1. Involved baseline models in this experiment and their details.
Table 1. Involved baseline models in this experiment and their details.
ModelDescription
BART [62]An approach which utilizes a transformer-based [71] denoising auto-regressive decoder combined with a transformer-based bidirectional encoder to pretrain a seq2seq model. In this experiment, BART-large and BART-base are examined.
BigBird [72]A mechanism which uses sparse attention to enable transformer-based models to handle long sequences.
Blenderbot [73]A transformer-based model trained on large scale data to carry out high-quality conversations.
Encoder-Decoder [74]A framework for constructing a seq2seq model which respectively takes two pretrained models as its encoder and decoder. In this experiment, we examine BERT2BERT, a model using a BERT-based [61] encoder and a BERT-based decoder.
PEGASUS [64]A transformer-based seq2seq model for the abstractive summarization task which is pretrained using extracted gap-sentences on large scale datasets.
ProphetNet [63]A self-supervised model with future n-gram prediction and n-stream self-attention mechanisms that is pretrained on large scale text corpora for both question generation and text summarization tasks.
T5 [65]A large-scale pretrained transformer-based model which utilizes transfer learning techniques to convert text-related language tasks into a text-to-text format.
Table 2. The performance of IvCDS and baselines on three tasks, where P = precision, R = recall, F1 = F1 score, the models are sorted by the F1 score on the NLU task, and a highlighted score indicates that its corresponding model outperforms other models according to the evaluation metric.
Table 2. The performance of IvCDS and baselines on three tasks, where P = precision, R = recall, F1 = F1 score, the models are sorted by the F1 score on the NLU task, and a highlighted score indicates that its corresponding model outperforms other models according to the evaluation metric.
ModelNLUPOLNLG
PRF1PRF1BLEU-4ROUGE-LMETEORGLEU
IvCDS92.8791.7192.2987.2386.9887.1028.1949.2947.8827.80
ProphetNet88.6587.8688.2564.1861.5162.8111.2330.2933.6312.69
BART-large76.5986.8581.4066.2689.2676.0620.3740.8649.7522.50
BART-base68.1889.3877.3546.8388.0161.1419.5340.0449.4521.77
Pegasus41.3388.6256.3772.3184.3377.866.3926.5337.977.77
T5-large39.1967.1349.4937.0941.9739.387.8026.5539.359.39
BigBird48.3743.2445.6681.9272.7377.0520.1637.1339.6020.01
Blenderbot29.9776.8643.1232.6258.8041.962.9813.5227.023.62
BERT2BERT19.4187.2331.7621.2868.3832.464.4521.3229.805.35
Table 3. The results of ablation experiment on the performances of IvCDS on three tasks according to various components in training/inference input sequences, where H is dialogue history, D P is driver profile, and highlighted numbers indicate the best model on that evaluation metric.
Table 3. The results of ablation experiment on the performances of IvCDS on three tasks according to various components in training/inference input sequences, where H is dialogue history, D P is driver profile, and highlighted numbers indicate the best model on that evaluation metric.
TrainingInferenceNLUPOLNLG
H DP H DP PRF1PRF1BLEU-4ROUGE-LMETEORGLEU
78.2388.0982.8718.3424.0020.791.4310.1311.672.63
65.2476.7970.5421.4321.5521.490.939.2710.552.14
61.4583.2670.7127.3239.1432.181.5913.4615.662.83
87.3091.4789.3326.4828.2827.3523.4847.4146.9124.14
92.4791.9492.2028.7731.2829.9726.3648.3547.1726.02
88.9491.3690.1373.4779.3076.2727.8248.6748.4327.41
92.8791.7192.2987.2386.9887.1028.1949.2947.8827.80
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Ji, T.; Yin, X.; Cheng, P.; Zhou, L.; Liu, S.; Bao, W.; Lyu, C. IvCDS: An End-to-End Driver Simulator for Personal In-Vehicle Conversational Assistant. Int. J. Environ. Res. Public Health 2022, 19, 15493. https://doi.org/10.3390/ijerph192315493

AMA Style

Ji T, Yin X, Cheng P, Zhou L, Liu S, Bao W, Lyu C. IvCDS: An End-to-End Driver Simulator for Personal In-Vehicle Conversational Assistant. International Journal of Environmental Research and Public Health. 2022; 19(23):15493. https://doi.org/10.3390/ijerph192315493

Chicago/Turabian Style

Ji, Tianbo, Xuanhua Yin, Peng Cheng, Liting Zhou, Siyou Liu, Wei Bao, and Chenyang Lyu. 2022. "IvCDS: An End-to-End Driver Simulator for Personal In-Vehicle Conversational Assistant" International Journal of Environmental Research and Public Health 19, no. 23: 15493. https://doi.org/10.3390/ijerph192315493

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop