1. Introduction
If we are to achieve more beneficial and cohesive personal, business and societal goals, in the face of increasing global challenges such as an increasing sedentary lifestyle for citizens, the promotion of physical environment sustainability, a reduction in poverty, better physical and social living conditions, etc., we need to change our behaviour and consider how to do this more effectively. One plausible approach to achieve this is to consider incentivizing behaviour shifts. An incentive is something that motivates us as society members to select and perform specific actions, usually to achieve beneficial personal, business and societal goals. As our application and focus are on the use of transport as an action, we consider how incentives can beneficially improve the use of transport.
A lack of exercise is a major cause of heart disease [
1]; hence, a good example of a personal transport incentive is to choose how to use human-powered transport between home and work more effectively. For example, in a study in Australia, men who cycled to work were found to be significantly less likely to be overweight and obese compared with those driving to work [
2]. A common use of incentives in business is to facilitate economic gain by an individual or via competition or cooperation within an organization. To help improve the attainment of organization goals, such as increased productivity, prizes of pre-paid travel for MICE (meetings, incentives, conferences and exhibitions) in an exotic location can be offered [
3]. A beneficial transport incentive for society is to promote the use of off-peak travel and hence to lower the peak air quality and high-carbon emissions pertaining to peak-time commuting [
4]. Whilst many aspects of transport use are predictive or deterministic in nature in terms of being circadian or daily, weekly and seasonally driven, some transport use characteristics may be more dynamic and less temporally deterministic, e.g., road traffic congestion can be more effectively modelled if we handle not just the deterministic temporal aspect, but also the non-deterministic temporal aspects [
5]. These time-driven aspects begin already to illustrate the need for a multi-dimensional, context-driven, individualized or personalized, design for travel incentives. We can also consider that a higher daily travel cost for peak-time public transport coupled with over-crowding versus a lower daily cost for off-peak travel with less crowding can be perceived as both a penalizing incentive for using peak travel versus a reward incentive for using off-peak travel.
Hence, one of the challenges when designing incentives for off-peak versus peak travel is how to effectively leverage the use of the temporal context when organizing travel. There are several additional dimensions that need to be considered when designing incentives. Different travellers have different (personal) preference restrictions based on their location context that affect their dominant transport mode use, e.g., numerous studies have demonstrated that living in higher density, mixed transport use neighbourhoods is associated with less car use compared to living in low density, suburban environments [
6]. Hence, offering a global, or the same incentive, to each participant, e.g., to maximize the use of public transport to reach an end destination, by the shortest time and at an affordable cost appears disadvantageous for rural travellers. Often, the goal to reach an end destination is to travel by the shortest time at an affordable cost. If we consider the effectiveness of personalized travel incentives in isolation, and to be static, new secondary-order effects can arise, i.e., the increased demand for the use of inexpensive public travel routes with respect to time and location may lead to additional overcrowding at specific times and locations because the service supply resources cannot meet the demand [
7]. Thus, we need to also consider the cooperative and competitive behaviours of travellers as a collective when proposing travel incentives, else new, second-order less desirable norms could arise such as overcrowding on certain peak travel routes that may violate some travellers’ personal preferences to use these even though societal sustainable travel goals are achieved. The authors in [
3] have noted that for the area of incentivized travel, it is particularly lacking in academic research and remains without a strong theoretical foundation. One promising approach to achieve a balance between multiple goal dimensions taking into account group participants’ behaviour is to model the use of transport using game theory and machine learning methods.
Numerous Intelligent Transport Systems (ITS) have been researched and developed that support that personalized travel plans is advantageous. For example, common personalization preferences are to choose the start and end destination, the transport mode and for route and destination information to be adapted and to be presented in this way according to these preferences [
8,
9]. However, it is challenging to design incentives to shift travellers’ behaviours, for example to use more sustainable transport, away from their preconceived or default transport usage. For example, if by default we offer the same universal or global incentive to shift their use of their habitual travel plan to use more sustainable transport modes, some travellers may feel less inclined to do so because of their temporal, local or personal constraints and preferences. This is a complex problem. We also often tend to have more than one personal transport use goal [
10], so we need to consider how we need to more effectively design multi-dimensional incentives, e.g., ones that use multiple modes and that are context driven, as well as personalized ones with respect to travellers’ preferences and location and the real-time travel situation.
There exists numbers of ITSs designed as part of the smart city initiative to incentivize travellers to use efficient and eco-friendly transportation [
11]. However, when generating incentives for travellers, few ITSs take individual travellers’ preferences and incentive multi-dimensional goals into consideration. Some ITSs simply broadcast traffic signals, state information or non-personalized travel guidance to travellers [
12]. Some systems, such as the Advanced Traveller Information Systems [
13] and Cooperative Multi-agent Transportation Management and Route Guidance System [
14], were proposed to meet the incentive challenge. However, in [
13,
14], the incentives are mostly based on traffic state information, and the authors did not fully take travellers’ preferences combined with traffic state information into consideration to achieve multi-dimensional travel goals. To the best of our knowledge, little or no work has been undertaken within ITS to generate personalized, situation-aware travel incentives that also model and account for social or group behaviour.
The use of machine learning and game theory has been proposed as a promising candidate to solve the computing intelligence and multi-dimensional goal problem [
15,
16]. Following this trend, we designed a Rule-based Incentive Framework (RIF) to intelligently generate context-driven incentives for travellers, while considering the multi-dimensional goals. RIF supports personal incentives through the use of the Iterative Dichotomiser 3 (ID3) algorithm [
15] as its classical machine learning-based decision tree algorithm and evolutionary game theory [
17] as its game theory method. Through the generation of personalized incentives, RIF can intelligently and efficiently re-allocate traffic route capacity over time and space while adhering to any individual traveller’s preferences for transport modality and route, according to multi-dimensional incentive goals. RIF utilizes personal mobility patterns [
4], traffic volume information and traveller’s modality preferences as three of the main sources of real-time inputs for the utility calculations and the decision tree-related processing. In summary, the primary contributions of this paper are to:
Propose a rule-based incentive framework to solve the personalized context-driven incentive problem, whilst achieving personalized travel preference, travel efficiency and eco-friendly global transportation as the multi-dimensional goals.
Apply the decision tree as a machine learning algorithm to help generate personal incentives for travellers in RIF, in part based on the historical travel information.
Apply Population Evolution (PE) theory as a classic evolutionary game theory to address the personal incentive problem in RIF. By PE, the best fit personal incentives are generated for travellers based on the output of the Analytic Hierarchy Process (AHP) and the comparisons of the incentive utilities.
Provide a theoretical and numerical analysis and comparisons of the decision tree and population evolution-based incentive solutions in RIF.
The remainder of this paper is organized as follows. In
Section 2, we give the background and related work. In
Section 3, we describe the problem formulation. In
Section 4 and
Section 5, we present the RIF based on the use of the decision tree and PE to solve the personal incentive problem. In
Section 6, we discuss and analyse the pros and cons of using the decision tree and PE methods. Our simulation results are presented in
Section 7. In
Section 8, we present our conclusions and future work.
3. Problem Formulation
3.1. User Scenario and Incentive
In this paper, we introduce a typical user scenario for RIF to work with, which is shown in
Figure 1. Then, the problem formulation is presented based on this scenario. In
Figure 1, there are two traffic points
A and
B connected by
N candidate routes grouped as
. Travellers at one time can travel in one of the routes from group
using one of the
M possible modalities, e.g., foot, cycle, car, bus, etc., from group
. For a travel route
i, its traffic volume is composed of its incoming travellers and existing traffic
, and the maximal capacity of the route
i is denoted as
. In practice, through RIF, a traveller
t with specific travel route and modality preferences could compete for the available capacity resources within the candidate routes from
while using a specific modality from
between traffic points
A and
B.
According to the user scenario shown in
Figure 1, RIF should work as an intelligent traffic instructor for the travellers between traffic points
A and
B. In this paper, RIF shapes a series of incentives issued to travellers to suggest to them a less congested route and a better modality to save travel time, cost and pollution. In order to do so, RIF has to consider travellers’ preferences and real-time traffic situations, which will be summarized as incentive criteria later. Two important incentive design issues are first when and under what conditions will RIF shape and issue incentives to travellers and, second, what real-time traffic information is available to travellers. As discussed before, the incentive considered in this paper is of the proactive type. Therefore, RIF will shape incentives to those travellers when they start travelling from traffic point
A to
B (or from
B to
A) at a designated time point. Of course, there are other cases where incentives could be shaped and issued to travellers either proactively or reactively in a temporary or un-predictable manner, like a route change becomes more preferable than sticking with a problematic route. In order to illustrate the use of RIF, we consider the case given in
Figure 1.
Accordingly, we use to denote the incentives to be issued to travellers in RIF. An incentive suggests its recipient traveller to travel through route i by modality j between traffic points A and B at a specific time point. Obviously, for a specific route k, its involved incentives could be . The objective of RIF therefore is to guide any traveller t, grouped as , to find his/her best fit incentive between traffic points A and B following certain incentive criteria at a time. We use to denote whether an incentive is issued to traveller t () or not () in RIF. Therefore, is the output that RIF tries to reach. In this paper, we think about the personal incentive issuing at a specific, i.e., static time point, therefore, in rest part of this paper, we will not mention time any more.
3.2. Incentive Criteria and Weighting
When RIF generates incentives to travellers, the incentive criteria under consideration in this paper include:
A traveller’s preference for a route that indicates the traveller’s willingness to take up a suggestion for an alternative route.
A traveller’s preference for a modality that indicates the traveller’s willingness to take up a suggestion for an alternative modality.
Current route traffic volume that determines whether a traveller is in congested traffic or not.
Such incentive criteria, i.e., travel route, modality and traffic volume, were selected with the aim to balance between the need for guiding traffic travellers to select other travel options (route and modality) in the event of congestion while maintaining the overall traveller satisfaction. If a traveller has a high willingness to change travel route or modality, a suggestion to change the route will not adversely affect the traveller’s satisfaction. On the other hand, suggesting a change to less willing travellers will lower their satisfaction. In reality, travellers can be provided with a smartphone application or related device that tracks their personal mobility and travel patterns. For instance, the system can gain an insight into one traveller’s personal mobility profile by referring to the traveller’s frequent locations, habitual routes and modalities. In addition, the system can acquire personal preferences from a traveller, such as one’s modality and route preferences and his/her tolerance towards congestion, using an appropriate Human-Computer Interaction (HCI) design, such as a questionnaire. In practice, measurements of traffic volume can be taken in real time using appropriate sensors such as inductive loop detectors under the road surface at traffic lights. For traffic that exhibits daily, seasonal or other patterns, historical data can be analysed to provide short-term traffic predictions.
To solve the personal incentive problem, we can weight each criteria to represent the importance of a criterion to a traveller, which is according to the sensitivity of the traveller to the criterion using AHP. AHP is a methodology that has been widely used for complex decision support in different fields such as government, business, industry, healthcare and education [
36]. In this paper, the incentive-related decision making can be a very complex process, particularly when a number of dependent elements, i.e., criteria, including travel route, modality and route volume, are involved. These elements are dependent because they are co-related highly in transportation model. For example, there might not be a public transport route between the origin and destination, so no bus modality is associated with a specific route. Same goes for volume and modality, etc. In practice, it is complicated and difficult to make correct incentive decisions considering the dependent elements, even when these elements in ITS are tangible, carefully estimated, and well understood for the personal incentive problem at hand. As each element has a different influence on the final incentive decision, there is no absolute or correct personal incentive decision. Fortunately, AHP can help the personal incentive decision making to find one that best suits the ITS goal and the problem involved.
According to AHP, whether an incentive decision is suitable or not to the problem is reflected by the incentive utility with respect to the weight of each incentive element and the historical and real-time sensor readings for each elements from travellers’ previous and current trips. Here, we first apply AHP to weight the incentive element, i.e., criteria, then calculate the incentive utility based on those weights. Using the reasonable weights calculated by AHP, the contributions of each element to the final decision, made by utility comparisons, can be rationally modelled.
To apply AHP to weight the incentive criteria, the process is closely related to travellers’ sensitivity to a criterion. Specifically, if a traveller is highly sensitive to a criterion, e.g., traffic route, then this means that this traveller has a strong preference on his/her habitual choice with respect to this criterion, i.e., a strong preference for his/her habitual route. We measure the sensitivity of a traveller to a criterion by using a commonly-used scale between one and nine, where ‘1’ indicates the lowest sensitivity and ‘9’ is the highest [
17,
38].
Table 1 gives the sensitivity measurements based on whether any traveller from
is sensitive to each of the criteria or not as listed in
Table 2. According to
Table 1, the sensitivity of the emphasized criterion, such as “travel modality” for Travellers 1, 3,
T in
Table 2, is rated within the scale between six and nine and the sensitivity of a less-emphasized criterion, such as “travel volume”, for Travellers 1, 3 in
Table 2 is rated within the scale between one and five.
Based on the scaled sensitivity of a traveller to a criterion, pairwise comparisons are carried out between all pairs of criteria for each traveller to evaluate the relative importance of one criterion over another to the traveller. For example, considering traveller
t, the pairwise comparisons result in a square matrix
c of
, where
denotes the pair comparison between criteria
i and
j as shown in Equation (
2).
Based on Equation (
2), for the given traveller
t, the eigenvector for each criterion, say criterion
k, can be calculated using the geometric mean method as:
According to Equation (
3), the normalization of
will determine
as the weight of criterion
k for the traveller
t, which is formulated as:
3.3. Incentive Utility and Optimization Problem
On the historical and real-time data of travellers and the weights of incentive criteria, we define the utility of a traveller
t choosing an incentive
as:
where
is the historical times traveller
t travels through route
i and
is the maximal times of a traveller in group
travelling through one route in group
.
is the historical time of traveller
t taking modality
j, and
is the maximal times of a traveller in group
taking one modality in group
.
is the existing traffic volume of route
i, and
is the maximal traffic volume out of all the candidate routes. The utility formulated in Equation (
5) denotes that if a traveller receives an incentive suggesting a less congested route or their habitual route or modality, the utility of this incentive to the traveller will be relatively high.
The optimization problem in this paper therefore is to maximize the overall incentive utilities of all the travellers in
, which is formulated as:
where
is subject to the fact that the traffic volume of each route should not exceed its maximal allowed capacity, e.g.,
, as formulated in Equation (
7).
5. The PE Method
To solve problem
apart from the decision tree method, we designed another PE-based algorithm for personal incentive generation as Algorithm 1, which is based on a series of comparisons of incentive utilities, instead of data mining on a large amount of historical traffic data. Algorithm 1 carries out the incentive generation for all the travellers in group
. In the beginning of the algorithm, it employs a greedy approach such that each traveller selects an incentive out of all the candidate incentives that have the highest utility (Steps 3–8). The incentive utilities are calculated according to Equation (
5). Then, the travellers with the same incentive selected will form an incentive population (Step 6). For example,
is the incentive population formed by all of the travellers selecting incentive
. Afterwards, the incentive utility of each traveller and average utility of each incentive population, e.g.,
of
, will be updated and calculated using Steps 9 and 10. Obviously, the greedy incentive selection will lead to traffic congestion and cause the incentive utilities to drop down.
After the greedy incentive selection, Algorithm 1 runs a finite loop to obtain the optimized incentive results from Step 12 to 23. As shown in Step 12, if an incentive population has homogeneous utility smaller than its previous average utility, i.e., , then this means the population is not in its optimal situation and some travellers in are not satisfied with an incentive . Unsatisfied travellers could be those being affected by a congested traffic route i or the ones not taking route i or modality j as their preferred choices. The homogeneous utility of an incentive population represents the minimal utility, i.e., payoff of the entire incentive population. According to PE, if an incentivized population has a higher homogeneous utility, the incentivized population is more attractive to travellers. On the contrary, in the case of , Algorithm 1 has to adjust population and search the travellers in population to find specific travellers to be moved to another population, which will potentially benefit population while not jeopardizing the incentive utilities of the receiving population. The travellers to be moved are the ones that have a higher incentive utility in the receiving population than in the current population. For example, if traveller t in incentive population has (Step 15), then traveller t will be moved to the alternative incentive population: (Steps 16 and 17). After the travellers are relocated, the incentive utilities of all travellers and the homogeneous utilities of all incentive populations will be updated (Steps 20 and 21). Algorithm 1 will finish, i.e., the PE game converges to an equilibrium, when the homogeneous utilities are higher than the average utilities of all the incentive populations or the algorithm runs out of its allowed steps . Then, , as the result of the algorithm is obtained, which allocates incentives to the right travellers.
Algorithm 1: The PE-based Algorithm for personal incentive generation. |
|
6. Analysis of the Decision Tree and PE Methods
This paper provides two different solutions to the personal incentive problem. In this section, we discuss their pros and cons with respect to usability, scalability and computation complexity.
In theory, the decision tree method can effectively provide a solution to the personal incentive problem. In practice, the ID3 algorithm as a classical decision tree algorithm takes travellers’ historical travel data into account to explore travellers’ common sense behaviours. Through ID3, the incentive distribution for each traveller is not based on a heuristic searching algorithm such as the PE method, but is based on data mining of a large amount of travellers’ historical data. As long as the historical data used in the training are at the big data level, the ID3 algorithm will help each traveller to select the best fit incentive, instead of by random attempts. This enables the ID3 algorithm to closely reach the global optimization and not fall into a backwards or forwards dilemma. However, the ID3 algorithm relies heavily on travellers’ historical data and behaviour. According to [
26], the ID3 algorithm applied to the incentive problem still faces the problem of overfitting, false data, etc. Considering scalability and computational complexity, the time complexity of ID3 algorithm is
, where
n is the number of sample data used in the training. If the sample data reach a big data level, the algorithm execution time will be enormous and require distributed computing to guarantee finishing the computation in a short time. This will limit the scalability for the use of the ID3 algorithm. This paper only considers a simple scenario with two traffic points, where the ID3 algorithm is affordable for RIF. However, when applying the ID3 algorithm at the smart city level with a large number of pairs of traffic points to be considered, the requirement of heavy computing will limit the application of the decision tree method.
According to Algorithm 1, the main activities of the PE method are to carry out heuristic searching to make sure each incentive population has member travellers with relatively high utilities. The PE method does not depend on a large amount of historical data, but works out the personal incentive problem using an evolutionary game with a series of utility comparisons. It involves less computation and training time compared to the decision tree method, but gives a better result, which is validated by simulations in the next section. However, there are two challenges concerning the PE method. First, there is a possibility that the heuristic search cannot find the optimal solution, but only find a local optimization. This is because the heuristic searching is for each traveller, out of the travellers’ incentive population, without looking at the global problem . Even through the PE game itself tries to reach a global equilibrium to benefit all of the travellers, this issue still exists. Second, the heuristic search may cause the travellers’ incentive population regrouping to fall into a backwards versus forwards dilemma, where a traveller may be regrouped back to its previous incentive population without reaching an optimization. Those two challenges will escalate if the traveller group , route group and modality group are large, leading to Algorithm 1 not being able to converge after a long time of running. The time complexity of Algorithm 1 is , where n is the size of the travellers under consideration. Compared to the decision tree method, the PE method requires much less computation, as the number of travellers between two traffic points is relatively low. This makes the PE method more suitable to work as part of a smart city initiative.
7. Simulations
To validate the decision tree and PE methods in RIF, we apply a series of MATLAB-based simulations with the configurations listed in
Table 4. Based on the configurations, we run the ID3 algorithm and Algorithm 1 a finite number of times and list the overall incentive utilities of all the travellers as the outputs in
Table 5. We implement the greedy method mentioned in
Section 5 as a baseline solution to validate our personal incentive solutions. For the greedy method, each traveller will always choose the incentive that suggests his/her habitual route and modality. The greedy method is prone to cause a travel bias without promoting its effectiveness and green transportation use.
According to the outputs in
Table 5, it is obvious that the PE method gives the best performance, while the decision tree method gives more of a mediocre performance. For the greedy method, because every traveller takes a selfish approach without cooperation, it causes route congestion and a modality bias that worsens the performance. The PE method outperforms the decision tree method because it utilizes a sophisticated evolutionary game theory and the use of recent historical data of travellers to more effectively, in real time, work out a traveller’s strategy. In contrast, the decision tree method depends heavily on a large amount of historical data and lacks timeliness and cooperation.
To better demonstrate the outputs,
Figure 4 demonstrates the average utility of each incentive population provided by different methods, considering different traveller scenarios respectively.
Figure 4 demonstrates the same numerical results as summarized in
Table 5. According to the outputs, it is clear that the average utility of all the members in an incentive population is improved mostly by the PE method, then medially by the decision tree method, compared to the greedy method.
8. Conclusions and Future Work
This paper provides a Rule-based Incentive Framework (RIF) implemented using the decision tree and PE methods. By solving the personal incentive problem, RIF can enable each traveller to pursue more effective and green transportation choices as part of a smart city initiative. The decision tree and PE methods both have their pros and cons. Generally, the PE method gives a better performance through the use of evolutionary game theory and real-time travel information. However, the PE method may not reach the best optimization. The decision tree method depends heavily on the historical data, but it will give effective results if a large amount of sample data is provided.
For future work, we aim to combine the merits of the machine learning methodology, such as decision tree, with game theory, like PE, to come up with more effective intelligent personal incentive solutions, under more complicated system scenarios. In the meantime, we will consider the timeliness of RIF and the acceptance of personalized incentives by travellers, which will make RIF more applicable.