Next Article in Journal
On the Study of Fixed Points for a New Class of α-Admissible Mappings
Next Article in Special Issue
A Guaranteed Deterministic Approach to Superhedging—The Case of Convex Payoff Functions on Options
Previous Article in Journal
A Closed Form for Slant Submanifolds of Generalized Sasakian Space Forms
Previous Article in Special Issue
Robust Portfolio Optimization in an Illiquid Market in Discrete-Time
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Moving Information Horizon Approach for Dynamic Game Models

1
School of Automation, Qingdao University, Qingdao 266071, China
2
Digital Workplace Product Technology Department, Huawei, Shenzhen 518000, China
3
Faculty of Applied Mathematics and Control Processes, St. Petersburg State University, Universitetskiy Prospekt, 35, Petergof, 198504 St. Petersburg, Russia
4
School of Mathematics and Statistics, Qingdao University, Qingdao 266071, China
*
Author to whom correspondence should be addressed.
Mathematics 2019, 7(12), 1239; https://doi.org/10.3390/math7121239
Submission received: 25 August 2019 / Revised: 12 November 2019 / Accepted: 26 November 2019 / Published: 14 December 2019
(This article belongs to the Special Issue Mathematical Game Theory 2019)

Abstract

:
In the paper, a new class of dynamic game models with a moving information horizon or dynamic updating is studied. In this class of games, players do not have full information about the game structure (motion equations, payoff functions) on the interval on which the game is defined. It is supposed that the players at each stage of the dynamic game have only truncated information about the game structure defined by the information horizon. Cooperative and noncooperative settings are considered in the paper. Results are illustrated using the oligopoly advertising game model, and comparison between the solution in the initial game model and in the game model with moving information horizon is presented. Simulation results are presented.

1. Introduction

Noncooperative game theory deals with strategic interactions among multiple decision makers with the objective functions depending on the choices of all the players. It is supposed that the players do not cooperate, and it is important to model their behavior using some optimality principal. There are many results that have been obtained in the field of noncooperative game theory; for a general description. see the fundamental book [1]. In the papers [2,3], the solution concept called the Nash equilibrium was proposed; in the current paper, this solution concept is used, but for a dynamic setting. The following book [4] can be used to study the main recent results in the field of dynamic cooperative game theory, which is a focus of the current paper.
The cooperative dynamic game theory offers socially convenient and group efficient solutions to different decision problems involving strategic actions. One of the fundamental questions in the theory of cooperative dynamic games is to define the allocation rule for cooperative payoff or set of imputations based on the optimal behavior of players, and the notation of characteristic function is used for that, which shows the strength of coalition S. In this paper, it is defined as in [5] as a total payoff of players from coalition S in Nash equilibrium in the game with the following set of players: coalition S (acting as one player) and players from the set N∖S. To compute the feedback Nash equilibrium, we analyze [4]. A set of imputations or solutions of the game is determined using the characteristic function at the beginning of each subinterval. In the paper [6], Haurie noted the problem of time inconsistency for the Nash bargaining solution in one differential game model. The notion of time consistency was formalized mathematically by Petrosyan in the paper [7]. In the next paper on time consistency [8], L. Petrosyan defined the notion of the Imputation Distribution Procedure (IDP), which is used to compose a time consistent cooperative solution or single imputation. Later on, L. Petrosyan defined the notion of strong time consistency in the paper [9]. See the recent publications on this topic in [10,11,12]. In order to determine a solution for the game model with dynamic updating, it is necessary to combine partial solutions and their IDP on subintervals. The theories of time consistency and strong time consistency introduced by L. Petrosyan in [7,9] are also studied for the offered solution.
Classical dynamic game models assume that the game structure does not change in the interval on which the game is defined or that the players have all information about the change of the game structure. However, if we consider a long term process, then these assumptions cannot cope with the reality. In the class of games with continuous updating, it is supposed that the players have or use only information about motion equations and payoff functions defined on the interval with a length equal to the value of the information horizon. This information is updated as the current time evolves, see Figure 1. In order to define the best possible behavior for players in this type of dynamic game, it is necessary to use a special approach, which is called the Looking Forward Approach (LFA).
The concept of the looking forward approach was introduced in [13] for differential games, where a new class of games was presented and the foundation for further study of games with dynamic updating or a moving information horizon was laid. To get more information about the approach, one may consider the following papers [13,14,15,16,17,18,19,20]. In the paper [13], the looking forward approach was applied to the cooperative differential game with a finite horizon. The paper [17] was focused on studying the looking forward approach with stochastic forecast and dynamic adaptation. In the paper [15], the looking forward approach was applied to a cooperative differential game of pollution control. In the paper [16], the looking forward approach was applied to the cooperative differential game with an infinite horizon. The paper [14] was devoted to studying the cooperative differential games with an infinite horizon and different types of forecasts and information structures. The last papers on the looking forward approach [18,21] were devoted to either studying the looking forward approach for cooperative differential games with nontransferable utility and real-life application of the looking forward approach to economic simulation. The class of differential games with continuous updating was considered in the papers [19,20]; here, it was supposed that the updating process evolves continuously in time. In the paper [19], the system of Hamilton–Jacobi–Bellman equations was derived for the Nash equilibrium in the game with continuous updating. In the paper [20], the class of linear-quadratic differential games with continuous updating was considered, and the explicit form of the Nash equilibrium was derived. In this paper, we apply the looking forward approach to the class of games with dynamic updating. The importance of the results of this paper is supported by the applicability of dynamic models to real-life processes, the more simple structure of dynamic models, and the dynamic nature of numerical model predictive control. In the papers described above, we studied differential games. Another important thing is that in this paper, we study the dynamic oligopoly marketing model of advertising, where we apply the looking forward approach. This result is essentially new.
In order to demonstrate the looking forward approach, we present the dynamic marketing model of advertising with a finite horizon. We refrain from reviewing the literature on advertising models, given that there are recent surveys on this topic. Huang, Leng, and Liang [22] gave an extensive and comprehensive survey of dynamic advertising competition since 1994, their coverage starting where the previous survey in Feichtinger, Hartl, and Sethi [23] stopped. Jørgensen and Zaccour [24] covered advertising models in oligopolies (horizontal strategic interactions) and in marketing channels (vertical strategic interactions), whereas He, Prasad, Gutierrez, and Sethi [25] focused on Stackelberg differential game models in supply chains and marketing channels that include advertising. Finally, Aust and Buscher [26] and Jørgensen and Zaccour [27] concentrated on cooperative advertising in marketing channels. This model was based on the model presented by Naik, Prasad, and Sethi in [28]. The model in [28] is the extension of the Sethi model for awareness of auto brands with the churn term, which is the extension of the decay market share term in monopoly models capturing forgetting and noise. The closed-loop Nash equilibrium concept is used to obtain the optimal advertising expenditure for the noncooperative game model. Considering the applications of game theory in different fields of research, many subjects from society modeling have been covered, for example see [29]. In this paper, the looking forward approach is applied to the dynamic case of the model presented in [28]. Both noncooperative and cooperative models are considered.
The paper is structured as follows. In Section 2 , a description of the looking forward approach is presented for a general dynamic game model, and a solution for a game with dynamic updating is presented for both fixed and random information horizons. In Section 3, the looking forward approach is applied to a dynamic oligopoly marketing model of advertising for noncooperative game and cooperative game models, respectively. In Section 4, simulation results for a marketing game model are presented.

2. Looking Forward Approach

Information plays one of the main roles in game theory. Generally, classical differential game models consider only the complete information, which means the information about the game structure is known and will not change during the whole game. This paper focuses on, however, the case that when information updates dynamically, players use only the truncated information and update it at each stage. In order to model the behavior of players when information updates dynamically, a novel approach called the looking forward approach in [13] is introduced.
Consider a general N-stage n-person nonzero-sum discrete-time dynamic game Γ ( x 0 , N ) [30] with initial state x 0 R m . The state dynamics of the game is defined by the following difference equation:
x k + 1 = f k ( x k , u 1 k , u 2 k , , u n k ) , x 1 = x 0 ,
where u i k U i R m i is the control vector of player i at stage k, x k X is the state of the game, and k 1 , 2 , N T .
The payoff function of player i is:
K i ( x 0 , N , U ) = j = 1 N g i j [ x j , u 1 j , u 2 j , , u n j , x j + 1 ] 1 1 + r j 1 , f o r i 1 , 2 , n I ,
where r is the discount rate.
This game is a complete information game; however, in real-life-like games, players do not have full information about the game structure on the interval on which the game is defined. We apply the looking forward approach in order to model the behavior of players when information about the game updates dynamically.

2.1. Truncated Subgame

Assume that at each stage k, players have full information about the motion equations and payoff functions within k + T ¯ stages, where T ¯ is a fixed value, namely the information horizon. At the stage k, information about the game is updated. At the stage k + 1, players have full information about the game structure on the interval until the stage k + 1 + T ¯ , see Figure 1.
Definition 1.
Let k = 1 , , N T ¯ . Truncated subgame Γ k ( x k , 0 , k , k + T ¯ ) is defined on the interval k , k + T ¯ . The motion equation and payoff function of truncated subgame on the stage interval k , k + T ¯ coincide with that of the initial game on the same stage interval. The motion equation and initial condition of truncated subgame have the following form:
x k + 1 = f k ( x k , u 1 k , u 2 k , , u n k ) , x 1 , 0 = x 1 = x 0 ,
(denote x k , l = x k + l , u i k , l = u i k + l , l = 0 , T ¯ ¯ for the truncated subgame Γ k ( x k , 0 , k , k + T ¯ ) ).
The payoff function of firm i in truncated subgame Γ k ( x k , 0 , k , k + T ¯ ) has the form:
K i ( x k , 0 , k , k + T ¯ , U ) = j = 0 T ¯ g i k + j [ x k , j , u 1 k , j , u 2 k , j , , u n k , j , x k , j + 1 ] 1 1 + r k + j 1 ,
for i 1 , 2 , n I .

2.2. Noncooperative Outcomes in the Game Model with a Moving Information Horizon

Considering each truncated subgame, let u k N E , l ( x ) = ( u 1 N E , k , l ( x ) , , u n N E , k , l ( x ) ) , l [ 0 , T ¯ ] be a vector of strategies that provides a feedback Nash equilibrium solution for the truncated subgame Γ k ( x k , 0 , k , k + T ¯ ) , and by V ¯ i k ( l , x ) , denote the payoff of player i I in feedback Nash equilibrium in the subgame defined on the interval [ l , k + T ¯ ] , l = k , , k + T ¯ :
V ¯ i k ( l , x ) = j = l k + T ¯ g i k + j [ x k , j , u 1 N E , k , j , u 2 N E , k , j , , u n N E , k , j , x k , j + 1 ] 1 1 + r k + j 1 .
Theorem 1.
([31]) Suppose that there exist functions V ¯ i k ( l , x ) satisfying (5), then strategies that maximize the right hand side of (5) constitute the feedback Nash equilibrium:
V ¯ i k ( l , x ) = m a x u i k , l { g i k [ x , u 1 N E , k , l ( x ) , , u i 1 N E , k , l , u i k , l , u i + 1 N E , k , l , , u n N E , k , l ( x ) , f ¯ k ( x , u i k , l ) ] 1 1 + r k + l 1 + V ¯ i k ( l + 1 , f ¯ k ( x , u i k , l ) ) } = g i k [ x , u 1 N E , k , l ( x ) , , u n N E , k , l ( x ) , f k ( x k , u 1 N E , k , l ( x ) , , u n N E , k , l ( x ) ) ] × 1 1 + r k + l 1 + V ¯ i k ( l + 1 , f k ( x k , u 1 N E , k , l ( x ) , , u n N E , k , l ( x ) ) ) ,
where V ¯ i k ( T ¯ + 1 , x ) = 0 , i I .
Denote the strategies obtained by maximization the right hand side of (5) by u k N E , l ( x ) , l [ 0 , T ¯ ] . By x k N E = ( x k , 0 N E , , x k , T ¯ N E ) , denote the trajectory obtained with strategies u k N E , l ( x ) involved. The presented solution is valid only for a truncated subgame, not for the whole game defined on the interval [ 0 , N ] . It is also impossible to construct the Nash equilibrium using the classical approaches, since information about the game updates dynamically, and it is different on different intervals. In order to construct the resulting equilibrium strategies and corresponding trajectory for a class of games with dynamic updating, we introduce the notion of the resulting strategies and corresponding resulting trajectory:
Definition 2.
The resulting feedback Nash equilibrium strategies { u ^ i N E , j } j = 1 N , i I are constructed in the following way:
u ^ i N E , j = u i N E , j , 0 , j 1 , , N T ¯ , u i N E , N T ¯ , j N T ¯ , j N T ¯ + 1 , , N .
Definition 3.
The resulting noncooperative trajectory x ^ j N E j = 1 N is a combination of x k N E for each truncated subgame Γ k ( x k , 0 N E , k , k + T ¯ ) :
x ^ j N E j = 0 N = x j , 0 N E , j = 1 , N T ¯ ¯ , x N T ¯ , j ( N T ¯ ) N E , j = N T ¯ + 1 , N ¯ .
Along the resulting noncooperative trajectory, players receive a payoff according to the following formula:
V ^ ( x ^ k N E , N k ) = j = k N 1 T ¯ [ V ¯ i j ( 0 , x j , 0 N E ) V ¯ i j ( 1 , x j , 1 N E ) ] + V ¯ i N T ¯ ( 0 , x N T ¯ , 0 N E ) , k = 1 , N T ¯ ¯ , V ¯ i N T ¯ ( k N + T ¯ , x N T ¯ , k N + T ¯ N E ) , k = N T ¯ + 1 , N ¯ .

2.3. Cooperative Game Model with a Moving Information Horizon

The theory of cooperative dynamic games offers socially convenient and group efficient solutions to different decision problems involving strategic actions. One of the fundamental problems in the theory of cooperative dynamic games is the formulation of optimal or cooperative behavior for players and the definition of the allocation rule or cooperative solution, which is the set of imputations chosen by the players in the game. We suppose that the players decide to cooperate in each truncated subgame. By Γ k c ( x k , 0 , k , k + T ¯ ) , we denote a truncated cooperative subgame defined on the interval [ k , k + T ¯ ] with the initial condition x k , 0 . The maximal joint payoff of all players in this game can be defined by maximizing functional:
i I K i ( x k , 0 , k , k + T ¯ , U ) = i I j = 0 T ¯ g i k + j [ x k , j , u 1 k , j , u 2 k , j , , u n k , j , x k , j + 1 ] 1 1 + r k + j 1 ,
subject to:
x k + 1 = f k ( x k , u 1 k , u 2 k , , u n k ) ,
x 1 , 0 = x 1 = x 0 .
Assume that the Bellman function for each truncated subgame Γ k c ( x k , 0 , k , k + T ¯ ) has the following form [32]:
W k ( 0 , x k , 0 ) = m a x u i ( k ) , i I i I j = 0 T ¯ g i k + j [ x k , j , u 1 k , j , , u n k , j , f k ( x k , j , u 1 k , j , , u n k , j ) ] 1 1 + r k + j 1 ,
where u i ( k ) = u i k , 0 , , u i k , T ¯ .
Theorem 2.
Suppose that there exists function W k ( l , x ) in truncated subgame Γ k c ( x k , 0 , k , k + T ¯ ) , l [ 0 , T ¯ ] such that the following recursive relations are satisfied:
W k ( l , x ) = m a x u i ( k ) , i I { i I g i k [ x , u 1 k , l , , u n k , l , f k ( x , u 1 k , l , , u n k , l ) ] 1 1 + r k + l 1 + W k ( l + 1 , f k ( x , u 1 k ( x ) , , u n k ( x ) ) ) } ,
where W k ( T ¯ + 1 , x ) = 0 . Assume that the maximum in (12) is achieved under control u i * ( k ) , then u i * ( k ) is optimal in the control problem (8), (9).
Definition 4.
The resulting cooperative strategies { u ^ i * , j } j = 1 N , i I are constructed in the following way:
u ^ i * , j = u i * , j , 0 , j 1 , , N T ¯ , u i * , N T ¯ , j N T ¯ , j N T ¯ + 1 , , N .
In a similar way, we define the resulting cooperative trajectory. By u * , denote the optimal strategies obtained by maximization of the right hand side of (12). The corresponding trajectory is defined by x k * = ( x k , 0 * , , x k , T ¯ * ) .
Definition 5.
The resulting cooperative trajectory x ˜ l * j = 1 N is a combination of cooperative trajectories x k * for each truncated cooperative subgame Γ k c ( x k * , k , k + T ¯ ) defined on the successive intervals [ k , k + T ¯ ] :
x ˜ j * j = 1 N = x j , 0 * , j = 1 , N T ¯ ¯ , x N T ¯ , j ( N T ¯ ) * , j = N T ¯ + 1 , N ¯ .
On the interval [ k , k + 1 ] , the resulting cooperative trajectory coincides with the cooperative trajectory x k * in the truncated cooperative subgame Γ k c ( x k * , k , k + T ¯ ) . At the stage k + 1, information about the game structure updates. On the interval [ k + 1 , k + 2 ] , trajectory x ˜ k * coincides with cooperative trajectory x k + 1 * in the truncated cooperative subgame Γ k + 1 c ( x k + 1 * , k + 1 , k + 1 + T ¯ ) , etc.
The characteristic function of the coalition is an essential concept in the theory of cooperative games. In this paper, it is defined as in [5] as a total payoff of players from coalition S I in Nash equilibrium in the game with the following set of players: coalition S (acting as one player) and players from the set I∖S(acting as individuals). For each coalition S I , define the values of the characteristic function for each truncated subgame as was done in [16]:
V k ( S ; x k , 0 * ) = i I K i k ( x k , 0 * ; u ) , S = I , V ˜ k ( S , x k , 0 * ) , S I , 0 , S = ,
where V ˜ k ( S , x k , 0 * ) is defined as the total payoff of players from coalition S \ I in feedback Nash equilibrium u N E = ( u ¯ 1 N E , , u ¯ n N E ) in the game with the following set of players: coalition S (acting as one player) and players from the set | I \ S | , i.e., in the game with | I \ S | + 1 players. For | S | = 1 , the cooperative case turns into the noncooperative case. In other cases, players from coalition S act as one player maximizing their total payoff.

2.4. Resulting Cooperative Solution and Theoremerties

An imputation ξ k ( x k * , k , k + T ¯ ) for each truncated cooperative subgame Γ k + 1 c ( x k * , k , k + T ¯ ) is defined as an arbitrary vector satisfying the conditions:
individual rationality : ξ i k ( x k * , k , k + T ¯ ) V ( { i } ; x k * , k , k + T ¯ ) ,
group rationality : i I ξ i k ( x k * , k , k + T ¯ ) = V ( I ; x k * , k , k + T ¯ ) .
Denote the set of all possible imputations for each truncated subgame by E k ( x * , k , k + T ¯ ) . Suppose that for each truncated subgame, a non-empty cooperative solution is defined by:
M k ( x * , k , k + T ¯ ) E k ( x * , k , k + T ¯ ) ,
it can be a core, nucleus, or Shapley value.
In order to guarantee time consistency Theoremerty [30] in each cooperative truncated subgame, it is necessary to use the notion of the imputation distribution procedure. The Theoremerty of time consistency was introduced in 1977 by L. Petrosyan [33].
Following the continuous time analysis of Yeung and Petrosyan [34] for cooperative differential games, we formulate a discrete time version of the imputation distribution procedure [8] so that the agreed upon imputations will be time consistent. By B i k ( x i * , k ) , denote the payment that firm i receives in stage k under the cooperative agreement along the cooperative trajectory x i * , k k N . The payment scheme involving B i k ( x i * , k ) constitutes an IDP in the sense that the imputation of firm i over the stages from k to N can be expressed as:
ξ i k ( x i * , k ) = B i k ( x i * , k ) + j = k + 1 N B i j ( x i * , j ) , i I .
Definition 6.
Solution M ( x k * , k ) is called time consistent, if for any imputation ξ k ( x * , k ) M ( x k * , k ) , there exists IDP B k ( x i * , k ) = ( B 1 k ( x i * , k ) , , B n k ( x i * , k ) ) such that for k = 1 , N ¯ , this satisfies:
k N B k ( x i * , k ) M ( x k * , k ) k N B k ( x i * , k ) = ξ k ( x i * , k ) .
The agreed-upon optimality principle or cooperative solution is time consistent, if the condition (6) is maintained at any time instant throughout the game along the cooperative trajectory.
In order to construct the resulting cooperative solution or cooperative solution in the game with dynamic updating Γ ( x 0 , N ) , it is necessary to use a special approach, because the standard approach can only be applied to each truncated subgame with dynamic updating defined on the interval [ k , k + T ¯ ] . IDP also provides the time consistency Theoremerty of the new solution concept and the ability to determine solutions within the interval [ k , k + T ¯ ] .
Suppose that in each truncated subgame Γ k c ( x k * , k , k + T ¯ ) , cooperative solution M k ( x * , k , k + T ¯ ) along the cooperative trajectory x k * is selected. Suppose that the imputation:
ξ k ( x k * , k , k + T ¯ ) M k ( x * , k , k + T ¯ )
and corresponding IDP for k 1 , N T ¯ ¯ are selected in a way so as to guarantee the time consistency Theoremerty:
β k , l ( x k , l * ) = [ β 1 k , l ( x k , l * ) , , β n k , l ( x k , l * ) ] , for l 0 , T ¯ ¯ .
The corresponding IDP β k , l ( x k , l * ) can be obtained in the following form [30]:
β k , l ( x k , l * ) = ξ k ( x k , l * , k + l , k + T ¯ ) ξ k ( x k , l + 1 * , k + l + 1 , k + T ¯ ) , l 0 , T ¯ ¯ ,
where ξ k ( x k , T ¯ + 1 * , k + T ¯ + 1 , k + T ¯ ) = 0 .
However, the time consistency only holds on the interval [ k , k + T ¯ ] for each truncated subgame Γ k c ( x k * , k , k + T ¯ ) , k 1 , , N . In order to guarantee time consistency for the whole interval [ 1 , N ] , we introduce the resulting IDP.
Definition 7.
The resulting IDP β ^ k ( x ˜ * ) is defined for a set of chosen imputations in each truncated subgame ξ k ( x k , 0 * , k , k + T ¯ ) M k ( x * , k , k + T ¯ ) using the corresponding β k ( x k * ) as follows:
β ^ k ( x * ) k = 1 N = β k , 0 ( x k , 0 * ) , k = 1 , N T ¯ ¯ , β N T ¯ , k N + T ¯ ( x N T ¯ , k N + T ¯ * ) , k = N T ¯ + 1 , N ¯ .
Using the resulting IDP β ^ k ( x ˜ * ) , we define the following vector:
Definition 8.
The resulting imputation ξ ^ ( x ˜ * , N k ) is a vector defined using the resulting IDP β ^ k ( x * ) in the following way, for stage k = 1 , N ¯ :
ξ ^ ( x ˜ * , N k ) = l = k N β ^ l ( x ˜ * ) ,
in particular ξ ^ ( x ˜ * , N ) = l = 1 N β ^ l ( x ˜ * ) .
Using the notion of resulting imputation, it is possible to define the allocation rule for the joint payoff:
Definition 9.
Resulting solution M ^ k ( x ˜ * , N k ) is the set of resulting imputations ξ ^ ( x ˜ * , N k ) for all possible resulting IDPs, β ^ k ( x ˜ * ) (defined by different imputations in each truncated subgame).
Any resulting imputation ξ ^ ( x ˜ * , N ) M ^ ( x ˜ * , N ) and corresponding resulting IDP β ^ k ( x ˜ * ) allocate the total payoff of players along the resulting cooperative trajectory x ˜ l * in the initial game with prescribed duration or the following condition holds:
i I j = 1 k β ^ i j ( x ˜ j * ) = i I j = 1 k g i j [ x ˜ j * , u ˜ * , x ˜ l + 1 * ] 1 1 + r j 1 .
The resulting solution M ^ k ( x ˜ * , N k ) is time consistent by construction. However, it can also be proven that an arbitrary resulting cooperative solution is strong time consistent in the game model with dynamic updating. By an arbitrary solution, we understand a solution that for example is composed of some vectors from the core in the first stage, the Shapley value in the second stage, etc.
Definition 10.
Cooperative solution M ( x 0 , N ) is called strong time consistent, if for k = 1 , N ¯ and every ξ ( x 0 , N ) M ( x 0 , N ) , the corresponding IDP β k ( x k * ) satisfies the condition:
j = 1 k β j ( x j * ) M ( x k , 0 * , N k ) M ( x 0 , N ) ,
where a A = a + a : a A .
Theorem 3.
The arbitrary resulting cooperative solution M ^ ( x ˜ * , N ) is strong time consistent in the game model with dynamic updating.

2.5. Construction of the Characteristic Function in the Game Model with Dynamic Updating

In the previous section, we defined the notion of the resulting cooperative solution or we defined a way for allocating cooperative payoff along the chosen trajectory in the game with dynamic updating. However, we did not show that the constructed set was a set of imputations or was a set of vectors satisfying individual and group rationality Theoremerties. In order to justify the resulting cooperative solution, we need to define the characteristic function for the game model with dynamic updating.
Definition 11.
The resulting characteristic function V ¯ ( S ; x ˜ k * , N k ) , S I in the game Γ ( x ˜ k * , N k ) with dynamic updating was the function calculated using the values of characteristic functions V ( S ; x k * , k , k + T ¯ ) in every truncated subgame Γ k c ( x k , 0 , k , k + T ¯ ) along the resulting cooperative trajectory x ˜ k * for k 1 , , N :
V ¯ ( S ; x ˜ k * , N k ) = l = k N T ¯ 1 [ V ( S ; x l * , l , l + T ¯ ) V ( S ; x l + 1 * , l + 1 , l + T ¯ ) ] + V ( S ; x N T ¯ * , N T ¯ , N ) , k = 1 , N T ¯ ¯ , V ( S ; x k * , k , N ) , k = N T ¯ + 1 , N ¯ .
Using the resulting characteristic function, we can show that the resulting imputation ξ ^ ( x ˜ * , N ) was the imputation in the game with dynamic updating:
Theorem 4.
The resulting imputation ξ ^ ( x ˜ * , N ) is the imputation in the game Γ ( x 0 , N ) with dynamic updating, if for k 1 , , N T ¯ , the following condition is satisfied:
ξ i k ( x k , l * , k + l , k + T ¯ ) ξ i k ( x k , l + 1 * , k + l + 1 , k + T ¯ ) V ( { i } ; x k , l * , k + l , k + T ¯ ) V ( { i } ; x k , l + 1 * , k + l + 1 , k + T ¯ ) , i I ,
where l = 0 , T ¯ ¯ .
The proof of Theorems 4, 5, 6, 8, 9, 10, 11 in the paper can be found in the Appendix A.

2.6. Relationship of the Solutions in Truncated Subgames and Resulting Solutions

In this section, it is shown that if the players choose imputations ξ k ( x k * , k , k + T ¯ ) E k ( x * , k , k + T ¯ ) based on V ( S ; x * , k , k + T ¯ ) for k 1 , , N T ¯ , S I in every truncated subgame by the same rule, then the resulting imputation ξ ^ ( x ˜ * , N k ) corresponds to the imputation chosen by the same rule using the resulting characteristic function V ¯ ( S ; x * , N k ) , S I . Further, we prove it for the number of classical optimality principals.
We assume that if in every truncated subgame Γ k c ( x k * , k , k + T ¯ ) , players choose Shapley value S h k ( x k * , k , k + T ¯ ) as the imputation, then the corresponding resulting imputation ξ ^ ( x ˜ * , N k ) coincides with the Shapley value S h ^ ( x ˜ k * , N k ) calculated using the resulting characteristic function V ¯ ( S ; x ˜ * , N k ) , S I .
Theorem 5.
Suppose that in every truncated subgame Γ k c ( x k * , k , k + T ¯ ) :
ξ k ( x k , 0 * , k , k + T ¯ ) = S h k ( x k * , k , k + T ¯ ) ,
where k 1 , , N T ¯ , then corresponding resulting imputation ξ ^ ( x ˜ k * , N k ) coincides with S h ^ ( x ˜ k * , N k ) :
ξ ^ ( x ˜ k * , N k ) = S h ^ ( x ˜ k * , N k ) , k [ 1 , N ] ,
where S h ^ ( x ˜ k * , N k ) is the Shapley value, calculated using the resulting characteristic function V ¯ ( S ; x ˜ k * , N k ) , S I .
Suppose that in every truncated subgame Γ k c ( x k * , k , k + T ¯ ) , players choose the core C ( x k * , k , k + T ¯ ) as the cooperative solution, then each resulting imputation ξ ^ ( x * , N k ) belongs to the core, defined using the resulting characteristic function V ¯ ( S , x ˜ k * , N k ) .
Theorem 6.
Suppose that in every truncated subgame Γ k c ( x k * , k , k + T ¯ ) :
M k ( x * , k , k + T ¯ ) = C ( x k * , k , k + T ¯ ) ,
where k 1 , , N T ¯ , then for every ξ k ( x k * , k , k + T ¯ ) C ( x k * , k , k + T ¯ ) , the following is satisfied:
ξ ^ ( x ˜ k * , N k ) C ^ ( x ˜ k * , N k ) , k [ k , N ] ,
where C ^ ( x ˜ k * , N k ) is the core, calculated using the resulting characteristic function V ¯ ( S , x ˜ k * , N k ) , S I , if this condition holds:
i S ξ ¯ i k ( x k , 0 * , k , k + T ¯ ) V ( S ; x * , k , k + T ¯ ) i S ξ ¯ i k ( x k , 1 * , k + 1 , k + T ¯ ) V ( S ; x * , k + 1 , k + T ¯ ) 0 , S I .

2.7. Random Information Horizon

In this section, we consider a case when the length of the information horizon is not fixed; this means that the players may have certain information on truncated intervals [ k , T ¯ k ] , but the length of intervals is a random variable (see Figure 2). By T ¯ k , we denote the length of the information horizon in the stage k. T ¯ k can take values from the interval [ T ^ k 1 + 1 , N ] , where T ^ k 1 is the realization of random horizon T ¯ k 1 in the previous truncated subgame, in particular T ^ 0 = 0 . It is worth mentioning that once T ¯ k = N , then the updating procedure stops. T ¯ k is a discrete random variable. At the initial position x 0 , the following probabilities are defined for realization on the interval [ 1 , N ] : { γ 1 , , γ N } . In the position x 1 (second position), the probabilities are recalculated in the following way:
γ T ¯ k 1 + 1 ζ = T ¯ k 1 + 1 N γ ζ , , γ N ζ = T ¯ k 1 + 1 N γ ζ .
The motion equation on the interval [ k , T ¯ k ] is:
x k + 1 = f k ( x k , u 1 k , u 2 k , , u n k ) , x 1 , 0 = x 1 = x 0 .
The payoff function of player i { 1 , , n } for the k th truncated subgame is:
E ζ = k T ¯ k g i ζ ( x ζ , u ζ , x ζ + 1 ) = T ¯ k = T ¯ k 1 + 1 N γ T ¯ k j = T ¯ k 1 + 1 N γ j ζ = k T ¯ k g i ζ ( x ζ , u ζ , x ζ + 1 ) .
Consider the noncooperative game model. In order to define the feedback Nash equilibrium in this game, denote the Bellman function V i k ( x , k ) for the k th truncated subgame as follows:
V i k ( x , k ) = m a x u i k , , u i T ¯ k T ¯ k = T ¯ k 1 + 1 N γ T ¯ k j = T ¯ k 1 + 1 N γ j ζ = k T ¯ k g i ζ ( x ζ , u ζ , x ζ + 1 ) | x k = x
with the set of strategies { u i j = ψ i j , j = k , T ¯ k ¯ , i I } , then:
V i k ( x , k ) = m a x u i k , , u i T ¯ k T ¯ k = T ¯ k 1 + 1 N γ T ¯ k j = T ¯ k 1 + 1 N γ j ζ = k T ¯ k g i ζ ( x ζ , u ζ , x ζ + 1 ) = T ¯ k = T ¯ k 1 + 1 N γ T ¯ k j = T ¯ k 1 + 1 N γ j ζ = k T ¯ k g i ζ ( x ζ * , ψ ζ , x ζ + 1 * ) = l = k T ¯ k 1 g i l ( x l * , ψ l , x l + 1 * ) + T ¯ k = T ¯ k 1 + 1 N γ T ¯ k j = T ¯ k 1 + 1 N γ j ζ = T ¯ k 1 + 1 T ¯ k g i ζ ( x ζ * , ψ ζ , x ζ + 1 * ) ,
where x k + 1 * = f k ( x k * , ψ k ( x k * ) ) .
As in the theorem in [35], a feedback Nash equilibrium in the noncooperative case of each truncated subgame with random horizon can be characterized as follows.
Theorem 7.
A set of strategies { ψ i j , j = k , N ¯ , i I } provides a feedback Nash equilibrium solution to the game (25), (26), if there exist functions V i k ( x , j ) , for j { k , , N } and i I , such that the following recursive relations are satisfied:
V ^ i k ( x , N ) = m a x u i N g i N ( x , ψ 1 N , , ψ i 1 N , u i N , ψ i + 1 N , , ψ n N , f ˜ N ( x , u i N ) ) , V ^ i k ( x , τ ) = m a x u i τ { g i τ ( x , ψ 1 τ , , ψ i 1 τ , u i τ , ψ i + 1 τ , , ψ n τ , f ˜ τ ( x , u i τ ) ) + j = τ + 1 N γ j j = τ N γ j V ^ i k ( f ˜ τ ( x , u i τ ) , τ + 1 ) } , for τ { T ¯ k 1 + 1 , , N 1 } , V i k ( x , ν ) = m a x u i ν { g i ν ( x , ψ 1 ν , , ψ i 1 ν , u i ν , ψ i + 1 ν , , ψ n ν , f ˜ ν ( x , u i ν ) ) + V i k ( f ˜ ν ( x , u i ν ) , ν + 1 ) } , for ν k , T ¯ k 1 ¯ , V i k ( x , T ¯ k 1 + 1 ) = V ^ i k ( x , T ¯ k 1 + 1 ) ,
where f ˜ k ( x , u i k ) = f k ( x , ψ 1 k , , ψ i 1 k , u i k , ψ i + 1 k , , ψ n k ) .
Consider the case when all players agree to cooperate in each truncated subgame. To maximize their expected joint payoff for the k th truncated subgame, we need to solve the following discrete time dynamic programming problem:
E i I ζ = k T ¯ k g i ζ ( x ζ , u ζ , x ζ + 1 ) = i I T ¯ k = T ¯ k 1 + 1 N γ T ¯ k j = T ¯ k 1 + 1 N γ j ζ = k T ¯ k g i ζ ( x ζ , u ζ , x ζ + 1 )
subject to (25).
As a noncooperative case, we can characterize an optimal solution for (25) and (28) as follows:
Corollary 1.
A set of strategies { ϕ i j , j = k , N ¯ , i I } provides an optimal cooperative strategies to the game (25) and (28), if there exists function W i k ( x , j ) , for j { k , , N } and i I , such that the following recursive relations are satisfied:
W ^ i k ( x , N ) = m a x u 1 N , , u n N i I g i N ( x , u 1 N , , u n N , f N ( x , u 1 N , , u n N ) ) , W ^ i k ( x , τ ) = m a x u 1 τ , , u n τ { i I g i τ ( x , u 1 τ , , u n τ , f τ ( x , u 1 τ , , u n τ ) ) + j = τ + 1 N γ j j = τ N γ j W ^ i k ( f τ ( x , u 1 τ , , u n τ ) , τ + 1 ) } , for τ { T ¯ k 1 + 1 , , N 1 } , W i k ( x , ν ) = m a x u 1 ν , , u n ν { i I g i ν ( x , u 1 ν , , u n ν , f ν ( x , u 1 ν , , u n ν ) ) + W i k ( f ν ( x , u 1 ν , , u n ν ) , ν + 1 ) } , for ν k , T ¯ k 1 ¯ ,
where W i k ( x , T ¯ k 1 + 1 ) = W ^ i k ( x , T ¯ k 1 + 1 ) .
Using the set of optimal strategies, { ϕ i j , j = k , N ¯ , i I } x k * can be generated from (25), and we obtain the corresponding cooperative trajectory.

3. Dynamic Oligopoly Marketing Model of Advertising

3.1. Initial Game Model

Consider an n-firm dynamic oligopoly marketing model of advertising defined on the closed time interval [ t 0 , T ] . The solution for the noncooperative case of this model in differential form was presented in [28]. In this paper, the model is transformed to the dynamic setting and studied for not only the noncooperative case, but mainly for the cooperative case [36]. Advertising efforts as strategies are used by firms to compete on the oligopoly market, and each firm tries to increase its market share while competitors try to minimize it using their advertising efforts. Denote by x i k the market share of firm i I 1 , , n at stage k. The full list of notations is presented in the Table 1. The dynamic oligopoly marketing model of advertising is defined as follows. Motion equation of firm i I or its market share:
x i k + 1 = h ( n n 1 ρ i u i k 1 x i k 1 n 1 j I ρ j u j k 1 x j k ) ( h δ 1 ) x i k + h δ 1 n , x i 0 = z i 0 ,
where z i 0 is a positive constant.
The payoff function of firm i I in the dynamic game model or the profit of firm i I :
K i 0 ( x 0 , u 0 ) = l = 0 N m i x i l ( u i l ) 2 .
Unlike the motion Equation (2) in the model example, here we use the discount rate r = 0 . The main reason is that we do not want to complicate the result of the model and show the difference between the looking forward approach and the initial differential game first. The next thing is that we may model the advertising process on the short time interval, i.e., with a small value of T.
In the table, the list of parameters of the model is presented.
The initial marketing game model is defined on the interval [0,N]. As in Section 2, suppose that players have full information about the motion equations and payoff functions on the interval [ k , k + T ¯ ] . In order to construct a solution in the game model with dynamic updating, we use the notion of truncated subgame Γ k ( x k , 0 , k , k + T ¯ ) .
The payoff function of firm i I in truncated subgame Γ k ( x k 0 , k , k + T ¯ ) has the form:
K ¯ i k ( x i k , 0 , u ) = j = k k + T ¯ m i x i j ( u i j ) 2 .
In Expression (30), there is a logical consistency requirement that the market shares should satisfy:
i I x i γ = 1 , x i γ 0 , i I , γ = 0 , N ¯ .

3.2. Noncooperative Outcomes in a Truncated Subgame

Firstly, we consider the noncooperative case for each truncated subgame Γ k ( x k 0 , k , k + T ¯ ) , k { 0 , , N T ¯ } with the initial condition x k 0 . According to Theorem 1, the set of strategies { u ¯ i k , l , for i 1 , , n , l = 0 , T ¯ ¯ and k 0 , , N T ¯ }, provides a feedback Nash equilibrium solution to the game (30), (32) if there exist functions V i k ( l , x i k , l ) , for i 1 , , n , l = 0 , T ¯ ¯ and k 0 , , N T ¯ in each truncated subgame Γ k ( x k 0 , k , k + T ¯ ) , such that the following recursive relations are satisfied, for each i I :
V i k ( l , x i k , l ) = m a x u i k , l m i x i k , l ( u i k , l ) 2 + V i k ( l + 1 , x i k , l + 1 ) = m i x i k , l ( u ¯ i k , l ) 2 + V i k ( l + 1 , x i k , l + 1 ) .
Theorem 8.
The payoff of players for each truncated subgame Γ k ( x k 0 , k , k + T ¯ ) in feedback Nash equilibrium on the interval [ k , k + T ¯ ] has the following form:
V i k ( l , x i k , l ) = i I A ˜ i k , l x i k , l + B ˜ k , l ( i ) , i 1 , , n , k 0 , , N T ¯ , l = 0 , T ¯ ¯ ,
where A ˜ i k , l , B ˜ k , l ( i ) are determined from the relations:
A ˜ i k , l = m i ( G ¯ ^ i k , l + 1 Z i ) 2 A ˜ i k , l + 1 ( δ h 1 ) , A ˜ j k , l = 2 ( G ¯ ^ j k , l + 1 Z j ) 2 A ˜ j k , l + 1 ( δ h 1 ) , j i , j I \ i , B ˜ k , l ( i ) = i I [ 2 ( G ¯ ^ i k , l + 1 Z i ) 2 + A ˜ i k , l + 1 h δ n ] + B ˜ k , l + 1 ( i ) ( G ¯ ^ i k , l + 1 Z i ) 2 ,
with initial conditions A ˜ i k , T ¯ + 1 = B ˜ k , T ¯ + 1 ( i ) = 0 , i 1 , , n , where G ^ i k , l + 1 = n A ˜ i k , l + 1 j I A ˜ j k , l + 1 , Z i = h ρ i 2 ( n 1 ) , G ¯ ^ i k , l + 1 = max 0 , G ^ i k , l + 1 .
The corresponding feedback Nash equilibrium strategies are:
u ¯ i k , l = G ¯ ^ i k , l + 1 Z i 1 x i k , l , k 0 , , N T ¯ , i 1 , , n .
Consider now the case of a random horizon for a truncated subgame. Suppose that the players know certain information about the game structure, but the duration of this information being right (or not changed) is unknown. Suppose that the probabilities are symmetric on the interval [ 1 , N ] , γ 1 = = γ N = 1 N . In our model, the first truncated subgame starts at Stage 0. Invoking Theorem 7 for the k th truncated subgame, if there exist continuously differentiable functions V i k ( x , f ) , i { 1 , , n } , f { k , , T ¯ k 1 } , and V ^ i k ( x , τ ) , for τ { T ¯ k 1 + 1 , , N } , such that the following recursive relations are satisfied:
V ^ i k ( x , N ) = m a x u i N 0 m i x i N ( u i N ) 2 , V ^ i k ( x , τ ) = m a x u i τ 0 m i x i τ ( u i τ ) 2 + N τ N τ + 1 V ^ i k ( x i τ + 1 , τ + 1 ) , for τ { T ¯ k 1 + 1 , , N 1 } , V i k ( x , f ) = m a x u i ν 0 m i x i f ( u i f ) 2 + V i k ( x i f + 1 , f + 1 ) , for f k , T ¯ k 1 ¯ . V i k ( x , T ¯ k 1 ) = V ^ i k ( x , T ¯ k 1 ) ,
then strategies u i , i I that maximize the right hand side of System (35) constitute the feedback Nash equilibrium. The corresponding Bellman functions can be obtained as follows:
Theorem 9.
The Bellman functions for the feedback Nash equilibrium of the k th truncated subgame in (35) are:
1.
V ^ i k ( x , τ ) = i I A i k , τ x i τ + B k , τ ( i ) , τ { T ¯ k 1 + 1 , , N 1 } ,
where A i k , τ = m i ( N τ N τ + 1 Y ¯ i k , τ + 1 Z i ) 2 N τ N τ + 1 A i k , τ + 1 ( δ h 1 ) , A j k , τ = 2 ( N τ N τ + 1 Y ¯ j k , τ + 1 Z j ) 2 N τ N τ + 1 A j k , τ + 1 ( δ h 1 ) , j I \ i , B k , τ ( i ) = i I [ 2 ( N τ N τ + 1 Y ¯ i k , τ + 1 Z i ) 2 + N τ N τ + 1 A i k , τ + 1 h δ n ] + N τ N τ + 1 B k , τ + 1 ( i ) ( N τ N τ + 1 Y ¯ i k , τ + 1 Z i ) 2 .
2.
V i k ( x , f ) = i I A i k , f x i f + B k , f ( i ) , f { k , , T ¯ k 1 } ,
where A i k , f = m i ( Y ¯ i k + 1 Z i ) 2 A i k , f + 1 ( δ h 1 ) , A j k , f = 2 ( Y ¯ j k , f + 1 Z j ) 2 A j k , f + 1 ( δ h 1 ) , j I \ i , B k , f ( i ) = i I [ 2 ( Y ¯ i k , f + 1 Z i ) 2 + A i k , f + 1 h δ n ] + B k , f + 1 ( i ) ( Y ¯ i k , f + 1 Z i ) 2
with initial conditions A i k , N = m i , B k , N + 1 ( i ) = 0 , i 1 , , n , A j k , N = 0 , for j I \ i and Y i k , t + 1 = n A i k , t + 1 j I A j k , t + 1 , Y ¯ i k , t + 1 = max 0 , Y i k , t + 1 t { k , , N 1 } .

3.3. Cooperative Outcomes in a Truncated Subgame

In this section, we focus on the cooperative setting, where all firms agree to cooperate in each truncated subgame. Denote by Γ k c ( x k 0 , k , k + T ¯ ) the truncated cooperative subgame on the interval [ k , k + T ¯ ] with the initial condition x k 0 . All firms aim to maximize their joint payoff:
i I K ¯ i k ( x i k , 0 , u ) = i I j = k k + T ¯ m i x i j ( u i j ) 2 .
Suppose that the set of strategies { u i * , k , l , for i 1 , , n , l = 0 , T ¯ ¯ and k 0 , , N T ¯ }, provides an optimal control to the game (30), (38). Invoking Theorem 2.2, if there exist functions { W k ( l , x k l ) , x k l = ( x 1 k , l . , x n k , l ) } , for i 1 , , n , l = 0 , T ¯ ¯ and k 0 , , N T ¯ in each truncated subgame Γ k c ( x k 0 , k , k + T ¯ ) , then the following recursive relations are satisfied:
W k ( l , x k l ) = m a x u 1 k , l , , u n k , l 0 i I m i x i k , l ( u i k , l ) 2 + W k ( l + 1 , x k l + 1 ) = i I m i x i k , l ( u i * , k , l ) 2 + W k ( l + 1 , x k l + 1 ) .
Theorem 10.
The maximum joint payoff of players for each truncated subgame Γ k c ( x k 0 , k , k + T ¯ ) has the following form:
W k ( l , x k l ) = i I C ^ i k , l x i k , l + D ^ i k , l ,
where C ^ i k , l , D ^ i k , l , i 1 , , n , k 0 , , N T ¯ , l = 0 , T ¯ ¯ satisfy the relations:
C ^ i k , l = m i ( G ¯ ¯ i k , l + 1 Z i ) 2 C ^ i k , l + 1 ( δ h 1 ) , D ^ i k , l = ( G ¯ ¯ i k , l + 1 Z i ) 2 + C ^ i k , l + 1 δ h n + D i k , l + 1
with initial conditions C ^ i k , T ¯ + 1 = 0 , D ^ i k , T ¯ + 1 = 0 , G ¯ i k , l + 1 = n C ^ i k , l + 1 j I C ^ j k , l + 1 and G ¯ ¯ i k , l + 1 = max 0 , G ¯ i k , l + 1 .
The optimal cooperative strategies for each truncated subgame have the form:
u i * , k , l = G ¯ ¯ i k , l + 1 Z i 1 x i k , l , i I .
Denote the solution of (39) with optimal strategies u * involved with the cooperative trajectory, x k * , l = ( x 1 * , k , l , , x n * , k , l ) , l = 0 , T ¯ ¯ . The cooperative trajectory in each truncated subgame can be calculated using (30) as follows, k 0 , , N T ¯ :
x i * , k , l + 1 = 2 n G ¯ ¯ i k , l + 1 Z i 2 ( 1 x i * , k , l ) j I 2 G ¯ ¯ j k , l + 1 Z j 2 ( 1 x j * , k , l ) ( h δ 1 ) x i * , k , l + h δ n .
So far, we have formalized the case when all firms cooperate in the game model with dynamic updating and fixed information horizon. Hence, we start to study the case when the information horizon is random, where T ¯ 1 has symmetric probabilities on the interval [ 1 , N ] , γ 1 = = γ N = 1 N . In the k th cooperative truncated subgame, if there exist functions W k ( x , f ) , for i { 1 , , n } , f { k , , T ¯ k 1 } , and W ^ k ( x , τ ) , for τ { T ¯ k 1 + 1 , , N } , such that the following recursive relations are satisfied:
W ^ k ( x , N ) = m a x u 1 N , , u n N i I m i x i N ( u i N ) 2 , W ^ k ( x , τ ) = m a x u 1 τ , , u n τ i I m i x i τ ( u i τ ) 2 + N τ N τ + 1 W ^ k ( x i τ + 1 , τ + 1 ) , for τ { T ¯ k 1 + 1 , , N 1 } , W k ( x , f ) = m a x u 1 f , , u n f i I m i x i f ( u i f ) 2 + W k ( x i f + 1 , f + 1 ) , for f k , T ¯ k 1 ¯ , W i k ( x , T ¯ k 1 ) = W ^ i k ( x , T ¯ k 1 ) ,
then the strategies { u i * , k , l , for i 1 , , n , l = 0 , T ¯ ¯ and k 0 , , N T ¯ } are optimal.
The value function can be obtained as follows:
Theorem 11.
The Bellman functions of the k th cooperative truncated subgame in (40) are:
1.
W ^ k ( x , τ ) = i I C ˜ i k , τ x i τ + D ˜ i k , τ , τ { T ¯ k 1 + 1 , , N 1 } ,
where C ˜ i k , τ = m i ( N τ N τ + 1 G ¯ ^ i k , τ + 1 Z i ) 2 N τ N τ + 1 C ˜ i k , τ + 1 ( δ h 1 ) , D ˜ i k , τ = ( N τ N τ + 1 G ¯ ^ i k , τ + 1 Z i ) 2 + N τ N τ + 1 C ˜ i k , τ + 1 h δ n + N τ N τ + 1 D ˜ i k , τ + 1 ;
2.
W i k ( x , f ) = i I C ˜ i k , f x i f + D ˜ i k , f , f { k , , T ¯ k 1 } ,
where C ˜ i k , l = m i ( G ¯ ^ i k , l + 1 Z i ) 2 C ˜ i k , l + 1 ( δ h 1 ) , D ˜ i k , l = ( G ¯ ^ i k , l + 1 Z i ) 2 + C ˜ i k , l + 1 δ h n + D i k , l + 1 ,
with initial conditions C ˜ i k , N = m i , D ˜ i k , N = 0 , i 1 , , n , G ^ i k , t + 1 = n C ˜ i k , t + 1 j I C ˜ j k , t + 1 , G ¯ ^ i k , t + 1 = max 0 , G ^ i k , t + 1 , t { k , , N 1 } .

3.4. Characteristic Function in a Truncated Subgame

Using the characteristic function formula (14), we define the characteristic function for each truncated subgame:
V k ( S ; x k * , 0 ) = i I K i k ( x k * , 0 ; u ) , S = I , V ˜ k ( S , x k * , 0 ) , S I , 0 , S = ,
where V ˜ k ( S , x k * , 0 ) is defined as the total payoff of players from coalition S in feedback Nash equilibrium u N E = ( u ¯ 1 , , u ¯ n ) in the game with the following set of players: coalition S (acting as one player) and players from the set I \ S , i.e., in the game with | I \ S | + 1 players.
Suppose that the Bellman function for coalition S I and players i I \ S in truncated subgame Γ k c ( x k 0 , k , k + T ¯ ) has the following form:
V k ( S , x k 0 ) = i I C ˜ i k , 0 x i k , 0 + E ¯ k , 0 , S I V k ( i , x k 0 ) = i I A ^ i k , 0 x i k , 0 + B ^ ( i ) k , 0 , i I \ S ,
then the corresponding system of Bellman equations will have the form:
V k ( S , x k 0 ) = m a x u i k , i S i S [ m i x i k , 0 ( u i k ) 2 ] + V k ( S , x k 1 ) , V k ( i , x k 0 ) = m a x u i k m i x i k , 0 ( u i k ) 2 + V k ( i , x k 1 ) , i I \ S , V k ( S , x k T ¯ + 1 ) = 0 , V k ( i , x k T ¯ + 1 ) = 0 .

4. Numerical Simulation for a Dynamic Advertising Game Model with Updating

We consider a specific three firm oligopoly case n the stage interval N = 8, with fixed information horizon T ¯ = 2 . In order to illustrate our model, let the parameters for each firm be ρ = [ 0.4 , 0.5 , 0.3 ] , h = 0.4 , δ = 0.09 , m = [ 0.6 , 1 , 1 . 2 ] and with the initial conditions z 0 = [ 0.3 , 0.5 , 0.2 ] .
Figure 3 and Figure 4 show the feedback Nash equilibrium strategies for each firm in the noncooperative case and cooperative strategies in the cooperative case respectively for both the initial game model and the game model with the looking forward approach.
Figure 5 and Figure 6 show the dynamics of players’ market shares corresponding to the feedback Nash equilibrium strategies for each firm in the noncooperative case and the dynamics of market shares corresponding to the optimal cooperative strategies in the cooperative case respectively for both the initial game and the game model with dynamic updating.
Figure 7 and Figure 8 show the optimal payoff of each firm in the noncooperative case and the characteristic function of the cooperative case for coalitions: { 1 , 2 } , { 1 , 3 } , and { 2 , 3 } , respectively, for both the initial game model and the game model with dynamic updating.
Figure 9 shows the difference between the resulting Shapley value and the Shapley value in the initial game model. As we can see from above, the resulting Shapley value of the game with moving information horizon changed more steadily over stages rather than the Shapley value of the initial game. It can be seen that the competition among the companies was more fierce in the initial game rather than in the game with a moving information horizon, considering market shares in Figure 6 and advertising expenditure in Figure 7. The Shapley value of each company barely changed in the last stage, which means that the profit of each company did not change much, as is shown in Figure 9.
Suppose that the realization of the game information horizon: T ¯ 1 = 2 , T ¯ 2 = 5 , T ¯ 3 = 8 . Figure 10 shows the difference of the resulting noncooperative trajectory with a fixed information horizon and a random information horizon, respectively.
Figure 11 shows the difference between the resulting noncooperative outcomes of each firm in the noncooperative case with the fixed information horizon and random information horizon approaches.
Figure 12 shows the difference between the resulting cooperative trajectory with fixed and random information horizons.

5. Conclusions

In this paper, a novel approach was presented for defining the solution in the dynamic game model, where information updated dynamically. Here, players did not have full information about the game structure through the whole time interval on which the game was defined, but they had certain information about the game structure on a fixed interval, namely the information horizon. As the time evolved, information about the game structure updated. Moreover, a random information horizon was considered for this approach. Therefore, the approach can be used to perform more realistic modeling of real-life conflict processes.
The dynamic oligopoly marketing model of advertising was considered, and the corresponding numerical example was used to illustrate the comparison between the initial dynamic marketing model and the corresponding model with dynamic updating. Besides, the comparison between the looking forward approach with a fixed information horizon and the looking forward approach with a random information horizon was presented.

Author Contributions

Methodology, O.P.; formal analysis, L.S.; investigation, Y.L.; supervision, H.G.

Funding

The work of first author is supported by Postdoctoral International Exchange Program of China and funded by Russian Foundation for Basic Research (RFBR) according to the research project No. 18-00-00727 (18-00-00725). The work of last author is supported by National Natural Science Foundation of China (71571108).

Acknowledgments

Great thanks to Lihong Shi who while working in Huawei found time to contribute to this paper.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Appendix A.1. Proof of Theorem 4

It is necessary to prove that the following conditions are satisfied:
i = 1 n ξ ^ i ( x ˜ k * , N k ) = V ¯ ( I ; x ˜ k * , N k ) ,
ξ ^ i ( x ˜ k * , N k ) V ¯ ( { i } ; x ˜ k * , N k ) , i I .
Denote function
y ( k ) = N T ¯ + 1 , k = 1 , N T ¯ ¯ , k , k = N T ¯ + 1 , N ¯ .
According to Definition 8, the right hand side of (A1) can be rewritten as:
i = 1 n ξ ^ i ( x ˜ k * , N k ) = i = 1 n j = k N β ^ i j ( x ˜ k * ) = i = 1 n [ j = k N T ¯ ( ξ ¯ i j ( x j , 0 * , j , j + T ¯ ) ξ ¯ i k ( x j , 1 * , j + 1 , j + T ¯ ) ) + j = y ( k ) N ( ξ ¯ i N T ¯ ( x N T ¯ , j N + T ¯ * , j , N ) ξ ¯ i N T ¯ ( x N T ¯ , j + 1 N + T ¯ * , j + 1 , N ) ) ] = j = k N T ¯ V ( I ; x j , 0 * , j , j + T ¯ ) V ( I ; x j , 1 * , j + 1 , j + T ¯ ) + j = y ( k ) N V ( I ; x N T ¯ , j N + T ¯ * , j , N ) V ( I ; x N T ¯ , j + 1 N + T ¯ * , j + 1 , N ) .
The right hand side of (A3) is equal to V ¯ ( I ; x ˜ k * , N k ) according to Definition 11; therefore, (A1) is correct. Now, consider (A2), where the left hand side can be rewritten as:
ξ ^ i ( x ˜ k * , N k ) = j = k N T ¯ ( ξ ¯ i j ( x j , 0 * , j , j + T ¯ ) ξ ¯ i k ( x j , 1 * , j + 1 , j + T ¯ ) ) + j = y ( k ) N ( ξ ¯ i N T ¯ ( x N T ¯ , j N + T ¯ * , j , N ) ξ ¯ i N T ¯ ( x N T ¯ , j + 1 N + T ¯ * , j + 1 , N ) ) .
(A2) is fulfilled for each stage, if for k = 1 , N ¯ , the following condition is satisfied:
ξ ¯ i j ( x j , 0 * , j , j + T ¯ ) ξ ¯ i k ( x j , 1 * , j + 1 , j + T ¯ ) V ( { i } ; x j , 0 * , j , j + T ¯ ) V ( { i } ; x j , 1 * , j + 1 , j + T ¯ ) , j = k , N T ¯ ¯ , ξ ¯ i N T ¯ ( x N T ¯ , j N + T ¯ * , j , N ) ξ ¯ i N T ¯ ( x N T ¯ , j + 1 N + T ¯ * , j + 1 , N ) V ( { i } ; x N T ¯ , j N + T ¯ * , j , N ) V ( { i } ; x N T ¯ , j + 1 N + T ¯ * , j + 1 , N ) , j = N T ¯ + 1 , N ¯ .
Fulfillment of Condition (A5) implies the fulfillment of the condition for each truncated subgame k = { 1 , , N T ¯ } :
ξ i k ( x k , l * , k + l , k + T ¯ ) ξ i k ( x k , l + 1 * , k + l + 1 , k + T ¯ ) V ( { i } ; x k , l * , k + l , k + T ¯ ) V ( { i } ; x k , l + 1 * , k + l + 1 , k + T ¯ ) , l = 0 , T ¯ ¯ , i I .
The theorem is proven.

Appendix A.2. Proof of Theorem 5

Due to Definition 8 and (17), the resulting imputation is calculated using the Shapley values S h k ( x k * , k , k + T ¯ ) in each truncated subgame:
ξ ^ ( x k * , N k ) = j = k N T ¯ [ S h j ( x j , 0 * , j , j + T ¯ ) S h j ( x j , 1 * , j + 1 , j + T ¯ ) ] + j = y ( k ) N [ S h N T ¯ ( x N T ¯ , j N + T ¯ * , j , N ) S h N T ¯ ( x N T ¯ , j + 1 N + T ¯ * , j + 1 , N ) ] = S I , i S ( I S ) ! ( S 1 ) ! I ! { j = k N T ¯ [ ( V ( S ; x j , 0 * , j , j + T ¯ ) V ( S \ i ; x j , 0 * , j , j + T ¯ ) ) ( V ( S ; x j , 1 * , j + 1 , j + T ¯ ) V ( S \ i ; x j , 1 * , j + 1 , j + T ¯ ) ) ] + j = y ( k ) N [ ( V ( S ; x N T ¯ , j N + T ¯ * , j , N ) V ( S \ i ; x N T ¯ , j N + T ¯ * , j , N ) ) ( V ( S ; x N T ¯ , j + 1 N + T ¯ * , j + 1 , N ) V ( S \ i ; x N T ¯ , j + 1 N + T ¯ * , j + 1 , N ) ) ] } = S I , i S ( I S ) ! ( S 1 ) ! I ! { j = k N T ¯ [ ( V ( S ; x j , 0 * , j , j + T ¯ ) V ( S ; x j , 1 * , j + 1 , j + T ¯ ) ) ( V ( S \ i ; x j , 0 * , j , j + T ¯ ) V ( S \ i ; x j , 1 * , j + 1 , j + T ¯ ) ) ] + j = y ( k ) N [ ( V ( S ; x N T ¯ , j N + T ¯ * , j , N ) V ( S ; x N T ¯ , j + 1 N + T ¯ * , j + 1 , N ) ) ( V ( S \ i ; x N T ¯ , j N + T ¯ * , j , N ) V ( S \ i ; x N T ¯ , j + 1 N + T ¯ * , j + 1 , N ) ) ] } ,
S h ^ ( x k * , N k ) is the Shapley value, calculated using the resulting characteristic function V ¯ ( S , x ˜ k * , N k ) :
S h ^ ( x k * , N k ) = S I , i S ( I S ) ! ( S 1 ) ! I ! ( V ¯ ( S ; x ˜ k * , N k ) V ¯ ( S \ i ; x ˜ k * , N k ) ) = S I , i S ( I S ) ! ( S 1 ) ! I ! { j = k N T ¯ [ ( V ( S ; x j , 0 * , j , j + T ¯ ) V ( S ; x j , 1 * , j + 1 , j + T ¯ ) ) ( V ( S \ i ; x j , 0 * , j , j + T ¯ ) V ( S \ i ; x j , 1 * , j + 1 , j + T ¯ ) ) ] + j = y ( k ) N [ ( V ( S ; x N T ¯ , j N + T ¯ * , j , N ) V ( S ; x N T ¯ , j + 1 N + T ¯ * , j + 1 , N ) ) ( V ( S \ i ; x N T ¯ , j N + T ¯ * , j , N ) V ( S \ i ; x N T ¯ , j + 1 N + T ¯ * , j + 1 , N ) ) ] } .
Comparing (A7) with (A8), the theorem is proven.

Appendix A.3. Proof of Theorem 6

The statement of the theorem is satisfied when the following two statements are true:
1.
If players in every truncated subgame choose the imputation ξ k ( x k * , k , k + T ¯ ) C ( x k * , k , k + T ¯ ) calculated using V ( S ; x * , k , k + T ¯ ) for k 1 , , N T ¯ , then the resulting imputation ξ ^ ( x ˜ k * , N k ) belongs to core C ^ ( x ˜ k * , N k ) , calculated using the resulting characteristic function V ¯ ( S , x ˜ k * , N k ) .
2.
Core C ^ ( x ˜ k * , N k ) should not contain imputation ξ ^ ( x ˜ k * , N k ) , for which it is impossible to find the set of imputations in truncated subgame ξ k ( x k * , k , k + T ¯ ) C ( x k * , k , k + T ¯ ) .
For the first statement, if the set of imputations ξ k ( x k * , k , k + T ¯ ) satisfies the system of inequalities:
i S ξ i k ( x k , 0 * , k , k + T ¯ ) V ( S ; x * , k , k + T ¯ ) , S I ,
for k 1 , , N T ¯ , i I , then the resulting imputation ξ ^ ( x ˜ k * , N k ) satisfies the system of inequalities:
i S ξ ^ i ( x ˜ k * , N k ) V ¯ ( S , x ˜ k * , N k ) .
Substitute the expression for ξ i k ( x k , 0 * , k , k + T ¯ ) into the left hand side of (A10). To the right hand side of (A10), substitute (22). It shows that for every S I , k 1 , , N T ¯ , the fulfillment of (24) leads to the fulfillment of (A9); hence, (A10) is satisfied. The first statement is proven.
For the second statement, any ξ ^ ( x ˜ k * , N k ) C ^ ( x ˜ k * , N k ) , k 1 , , N T ¯ , the following inequalities are satisfied in the set of imputations ξ i k ( x k , 0 * , k , k + T ¯ ) , i 1 , , n , k 1 , , N T ¯ :
i S [ j = k N T ¯ ( ξ i j ( x j , 0 * , j , j + T ¯ ) ξ i k ( x j , 1 * , j + 1 , j + T ¯ ) ) + j = y ( k ) N ( ξ i N T ¯ ( x N T ¯ , j N + T ¯ * , j , N ) ξ i N T ¯ ( x N T ¯ , j + 1 N + T ¯ * , j + 1 , N ) ) ] j = k N T ¯ ( V ( S ; x j , 0 * , j , j + T ¯ ) V ( S ; x j , 1 * , j + 1 , j + T ¯ ) ) + j = y ( k ) N ( V ( S ; x N T ¯ , j N + T ¯ * , j , N ) V ( S ; x N T ¯ , j + 1 N + T ¯ * , j + 1 , N ) ) .
If there exists one set of imputations ξ i k ( x k , 0 * , k , k + T ¯ ) , i 1 , , n , k 1 , , N T ¯ in (A11) for which the inequalities (A9) are not satisfied, this unsatisfied inequalities will lead to the following condition for k 1 , , N T ¯ being satisfied:
i S ξ i k ( x k , 0 * , k , k + T ¯ ) V ( S ; x * , k , k + T ¯ ) < 0 ,
which contradicts (24), so the theorem is proven.

Appendix A.4. Proof of Theorem 8

The Bellman equation for firm i I is given by:
V i k ( l , x i k , l ) = m a x u i k 0 m i x i k , l ( u i k , l ) 2 + V i k ( l + 1 , x i k , l + 1 ) ) .
Invoking (30) and V i k ( l , x i k , l ) = i I A ˜ i k , l x i k , l + B ˜ k , l ( i ) , for i I , l = 0 , T ¯ ¯ and k 0 , , N , the above Bellman equation becomes:
i I A ˜ i k , l x i k , l + B ˜ k , l ( i ) = m a x u i k 0 m i x i k , l ( u i , l k ) 2 + i I A ˜ i k , l + 1 x i k , l + 1 + B ˜ k , l + 1 ( i ) = m a x u i k 0 { i I A ˜ i k , l + 1 [ h ( n n 1 ρ i u i k , l 1 x i k , l 1 n 1 j I ρ j u j k , l 1 x j k , l ) ( h δ 1 ) x i k , l + h δ 1 n ] + B ˜ k , l + 1 ( i ) + m i x i k , l ( u i k , l ) 2 } .
Performing the indicated maximization in the above yields:
2 u i k , l + h ρ i n 1 n A ˜ i k , l + 1 1 x i k , l h ρ i n 1 1 x i k , l j I A ˜ j k , l + 1 = 0 ,
for i I , l = 0 , T ¯ ¯ and k 0 , , N T ¯ , then the feedback Nash equilibrium strategy of firm i I can be obtained in the form:
u ¯ i k , l = m a x 0 , [ n A ˜ i k , l + 1 j I A ˜ j k , l + 1 ] h ρ i 2 ( n 1 ) 1 x i k , l
= h ρ i 1 x i k , l 2 ( n 1 ) m a x 0 , n A ˜ i k , l + 1 j I A ˜ j k , l + 1 .
Anticipating that the controls will be shown to be positive, substituting it into (A13), collecting the terms together, then the parameters A ˜ i k , l and B ˜ k , l ( i ) can be expressed as:
A ˜ i k , l = m i ( G ¯ ^ i k , l + 1 Z i ) 2 A ˜ i k , l + 1 ( δ h 1 ) , A ˜ j k , l = 2 ( G ¯ ^ j k , l + 1 Z j ) 2 A ˜ j k , l + 1 ( δ h 1 ) , j i , j I \ i , B ˜ k , l ( i ) = i I [ 2 ( G ¯ ^ i k , l + 1 Z i ) 2 + A ˜ i k , l + 1 h δ n ] + B ˜ k , l + 1 ( i ) ( G ¯ ^ i k , l + 1 Z i ) 2 ,
where G ^ i k , l + 1 = n A ˜ i k , l + 1 j I A ˜ j k , l + 1 , G ¯ ^ i k , l + 1 = m a x 0 , G ^ i k , l + 1 , Z i = h ρ i 2 ( n 1 ) .

Appendix A.5. Proof of Theorem 9

Consider the Bellman function in the last stage, in the stage N. Performing the indicated maximization in it yields:
ψ i N = 0 , for i I .
Invoking (36) and V ¯ i k ( x i N , N ) = m i x i N , the second equation in (35) becomes:
i I A i k , τ x i τ + B k , τ ( i ) = m a x u i τ m i x i τ ( u i τ ) 2 + N τ N τ 1 ( i I A i k , τ + 1 x i τ + 1 + B k , τ + 1 ( i ) ) .
Performing the indicated maximization in the above yields:
2 ( u i τ ) + N τ N τ 1 ( h ρ i n 1 n A i k , τ + 1 1 x i k , τ h ρ i n 1 1 x i k , τ j I A j k , τ + 1 ) = 0
for i I and τ { T ¯ k 1 + 1 , , N 1 } , then the feedback Nash equilibrium strategy of firm i I can be obtained in the form:
ψ i τ = m a x 0 , N τ N τ 1 ( n A i k , τ + 1 j I A j k , τ + 1 ) h ρ i 2 ( n 1 ) 1 x i k , τ
= N τ N τ 1 h ρ i 1 x i k , τ 2 ( n 1 ) m a x 0 , n A i k , τ + 1 j I A j k , τ + 1 .
Anticipating that the controls will be shown to be positive, substituting it into the second equation in (35), collecting the terms together, the parameters A i k , τ and B k , τ ( i ) can be expressed as:
A i k , τ = m i ( N τ N τ + 1 Y ¯ i k , τ + 1 Z i ) 2 N τ N τ + 1 A i k , τ + 1 ( δ h 1 ) , A j k , τ = 2 ( N τ N τ + 1 Y ¯ j k , τ + 1 Z j ) 2 N τ N τ + 1 A j k , τ + 1 ( δ h 1 ) , j I \ i , B k , τ ( i ) = i I [ 2 ( N τ N τ + 1 Y ¯ i k , τ + 1 Z i ) 2 + N τ N τ + 1 A i k , τ + 1 h δ n ] + N τ N τ + 1 B k , τ + 1 ( i ) ( N τ N τ + 1 Y ¯ i k , τ + 1 Z i ) 2 .
In the stage f { k , , T ¯ k 1 } , performing the indicated maximization in the third equation in (35) yields:
ψ i f = m a x 0 , ( n A i k , f + 1 j I A j k , f + 1 ) h ρ i 2 ( n 1 ) 1 x i k , f
= h ρ i 1 x i k , f 2 ( n 1 ) m a x 0 , n A i k , f + 1 j I A j k , f + 1 .
Anticipating that the controls will be shown to be positive, substituting it into the third equation in (35), collecting the terms together, then the parameters A i k , f and B k , f ( i ) can be expressed as:
A i k , f = m i ( Y ¯ i k + 1 Z i ) 2 A i k , f + 1 ( δ h 1 ) , A j k , f = 2 ( Y ¯ j k , f + 1 Z j ) 2 A j k , f + 1 ( δ h 1 ) , j I \ i , B k , f ( i ) = i I [ 2 ( Y ¯ i k , f + 1 Z i ) 2 + A i k , f + 1 h δ n ] + B k , f + 1 ( i ) ( Y ¯ i k , f + 1 Z i ) 2 ,
where Y i k , t + 1 = n A i k , t + 1 j I A j k , t + 1 , Y ¯ i k , t + 1 = m a x 0 , n A i k , t + 1 j I A j k , t + 1 , t { k , , N 1 } .

Appendix A.6. Proof of Theorem 10

The Bellman equation for the cooperative setting is given by:
W k ( l , x i k , l ) = m a x u 1 k , l , , u n k , l i I m i x i k , l ( u i * , k , l ) 2 + W k ( l + 1 , x i k , l + 1 ) .
Substituting (39) into (A14):
W k ( l , x k l ) = m a x u 1 k , l , , u n k , l i I m i x i k , l ( u i * , k , l ) 2 + W k ( l + 1 , x k l + 1 ) i I C ^ i k , l x i k , l + D ^ i k , l = m a x u 1 k , , u n k i I [ m i x i k , l ( u i k , l ) 2 ] + i I C ^ i k , l + 1 x i k , l + 1 + D ^ i k , l + 1 ,
W k ( T ¯ + 1 , x k T ¯ ) = 0 .
Substitute the right hand side of (30) to (A15). Performing the indicated maximization in (A15) yields:
2 u i k + ( n C ^ i k , l + 1 j I C ^ j k , l + 1 ) h ρ i 1 x i k , l n 1 = 0 .
Optimal cooperative strategies can be obtained in the form:
u i * , k , l = m a x 0 , ( n C ^ i k , l + 1 j I C ^ j k , l + 1 ) h ρ i 1 x i k , l 2 ( n 1 )
= h ρ i 1 x i k , l 2 ( n 1 ) m a x 0 , n C ^ i k , l + 1 j I C ^ j k , l + 1
for i I , l = 0 , T ¯ ¯ and k 0 , , N .
Anticipating the cooperative strategies will be shown to be positive, substituting it into (A15), hence C ^ i k , l and D ^ i k , l are defined as:
C ^ i k , l = m i ( G ¯ ¯ i k , l + 1 Z i ) 2 C ^ i k , l + 1 ( δ h 1 ) , D ^ i k , l = ( G ¯ ¯ i k , l + 1 Z i ) 2 + C ^ i k , l + 1 δ h n + D i k , l + 1 .

Appendix A.7. Proof of Theorem 11

Consider the Bellman function in the last stage, that is in the stage N. Performing the indicated maximization in it yields:
ϕ i N = 0 , f o r i I .
Invoking (41) and W ¯ i k ( x i N , N ) = i I m i x i N , then the second equation in (40) becomes:
i I C ˜ i k , τ x i τ + D ˜ i k , τ = m a x u 1 τ , , u n τ i I m i x i τ ( u i τ ) 2 + N τ N τ + 1 ( C ˜ i k , τ + 1 x i τ + 1 + D ˜ i k , τ + 1 ) .
Performing the indicated maximization in the above yields:
2 ( u i τ ) + N τ N τ 1 ( h ρ i n 1 n C ˜ k , τ + 1 1 x i k , τ h ρ i n 1 1 x i k , τ j I C ˜ j k , τ + 1 ) = 0
for i I and τ { T ¯ k 1 + 1 , , N 1 } , then the feedback Nash equilibrium strategy of firm i I can be obtained in the form:
ϕ i τ = m a x { 0 , N τ N τ 1 ( n C ˜ i k , τ + 1 j I C ˜ j k , τ + 1 ) h ρ i 2 ( n 1 ) 1 x i k , τ }
= N τ N τ 1 h ρ i 1 x i k , τ 2 ( n 1 ) m a x { 0 , n C ˜ i k , τ + 1 j I C ˜ j k , τ + 1 } .
Anticipating that the controls will be shown to be positive, substituting this into the second equation in (40), collecting the terms together, then the parameters C ˜ i k , τ and D ˜ i k , τ can be expressed as:
C ˜ i k , τ = m i ( N τ N τ + 1 G ¯ ^ i k , τ + 1 Z i ) 2 N τ N τ + 1 C ˜ i k , τ + 1 ( δ h 1 ) , D ˜ i k , τ = ( N τ N τ + 1 G ¯ ^ i k , τ + 1 Z i ) 2 + N τ N τ + 1 C ˜ i k , τ + 1 h δ n + N τ N τ + 1 D ˜ i k , τ + 1 .
For stage f { k , , T ¯ k 1 } , performing the indicated maximization in third equation of (40) yields:
ϕ i f = m a x 0 , ( n C ˜ i k , τ + 1 j I C ˜ j k , τ + 1 ) h ρ i 2 ( n 1 ) 1 x i k , f
= h ρ i 1 x i k , f 2 ( n 1 ) m a x 0 , n C ˜ i k , τ + 1 j I C ˜ j k , τ + 1 .
Anticipating that the controls will be shown to be positive, substituting this into the third equation in (40), collecting the terms together, then the parameters C ˜ i k , f and D ˜ i k , f can be expressed as:
C ˜ i k , l = m i ( G ¯ ^ i k , l + 1 Z i ) 2 C ˜ i k , l + 1 ( δ h 1 ) , D ˜ i k , l = ( G ¯ ^ i k , l + 1 Z i ) 2 + C ˜ i k , l + 1 δ h n + D i k , l + 1 ,
where G ^ i k , t + 1 = n C ˜ i k , t + 1 j I C ˜ j k , t + 1 , G ¯ ^ i k , t + 1 = m a x 0 , G ^ i k , t + 1 , t { k , , N 1 } .

References

  1. Owen, G. Game Theory; Emerald Group Publishing Limited: Bingley, UK, 2013. [Google Scholar]
  2. Nash, J.F. Equilibrium points in n-person games. Proc. Natl. Acad. Sci. USA 1950, 36, 48–49. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Nash, J.F. Non-cooperative games. Ann. Math. 1951, 54, 286–295. [Google Scholar] [CrossRef]
  4. Basar, T.; Olsder, G.J. Dynamic Noncooperative Game Theory; SIAM: Philadelphia, PA, USA, 1999; Volume 23. [Google Scholar]
  5. Chander, P.; Tulkens, H. A core-theoretic solution for the design of cooperative agreements on transfrontier pollution. In Public Goods, Environmental Externalities and Fiscal Competition; Springer: Berlin/Heidelberg, Germany, 2006; pp. 176–193. [Google Scholar]
  6. Haurie, A. A note on nonzero-sum differential games with bargaining solution. J. Optim. Theory Appl. 1976, 18, 31–39. [Google Scholar] [CrossRef]
  7. Petrosyan, L. Time-consistency of solutions in multi-player differential games. Vestn. Leningr. State Univ. 1977, 4, 46–52. [Google Scholar]
  8. Petrosyan, L.; Danilov, N. Stability of solutions in non-zero sum differential games with transferable payoffs. Vestn. Leningr. Univ. 1979, 1, 52–59. [Google Scholar]
  9. Petrosyan, L. Strongly time-consistent differential optimality principles. Vestn. St. Petersburg Univ. Math. 1993, 26, 40–46. [Google Scholar]
  10. Petrosyan, L.; Yeung, D.W.K. Dynamically stable cooperative solutions in randomly furcating differential games. Proc. Steklov Inst. Math. 2006, 253, S208–S220. [Google Scholar] [CrossRef]
  11. Jørgensen, S.; Yeung, D.W. Inter-and intragenerational renewableresource extraction. Ann. Oper. Res. 1999, 88, 275–289. [Google Scholar] [CrossRef]
  12. Jørgensen, S.; Martin-Herran, G.; Zaccour, G. Agreeability and time consistency in linear-state differential games. J. Optim. Theory Appl. 2003, 119, 49–63. [Google Scholar] [CrossRef]
  13. Petrosian, O. Looking Forward Approach in Cooperative Differential Games. Int. Game Theory Rev. 2016, 18, 1640007. [Google Scholar]
  14. Yeung, D.; Petrosian, O. Infinite Horizon Dynamic Games: A New Approach via Information Updating. Int. Game Theory Rev. 2017, 19, 1–16. [Google Scholar] [CrossRef]
  15. Gromova, E.; Petrosian, O. Control of information horizon for cooperative differential game of pollution control. In Proceedings of the International Conference on Stability and Oscillations of Nonlinear Control Systems (Pyatnitskiy’s Conference), Moscow, Russia, 1–3 June 2016; pp. 1–4. [Google Scholar]
  16. Petrosian, O.L. Looking Forward Approach in Cooperative Differential Games with Infinite-Horizon Vestn. St. Petersburg Univ. Math. 2016, 10, 18–30. [Google Scholar]
  17. Petrosian, O.; Barabanov, A. Looking Forward Approach in Cooperative Differential Games with Uncertain Stochastic Dynamics. J. Optim. Theory Appl. 2017, 172, 328–347. [Google Scholar] [CrossRef]
  18. Petrosian, O.; Nastych, M.; Volf, D. Differential game of oil market with moving informational horizon and non-transferable utility. In Proceedings of the Constructive Nonsmooth Analysis and Related Topics (Dedicated to the Memory of VF Demyanov) (CNSA), St. Petersburg, Russia, 22–27 May 2017; pp. 1–4. [Google Scholar]
  19. Petrosian, O.; Tur, A. Hamilton-Jacobi-Bellman Equations for Non-cooperative Differential Games with Continuous Updating. Commun. Comput. Inf. Sci. 2019. to be published. [Google Scholar]
  20. Kuchkarov, I.; Petrosian, O. On Class of Linear Quadratic Non-cooperative Differential Games with Continuous Updating. Lect. Notes Comput. Sci. 2019. to be published. [Google Scholar]
  21. Petrosian, O.; Nastych, M.; Volf, D. Non-cooperative Differential Game Model of Oil Market with Looking Forward Approach. Frontiers of Dynamic Games, Game Theory and Management, St. Petersburg, 2017; Petrosyan, L.A., Mazalov, V.V., Zenkevich, N., Eds.; Birkhäuser: Basel, Switzerland, 2018. [Google Scholar]
  22. Huang, J.; Leng, M.; Liang, L. Recent developments in dynamic advertising research. Eur. J. Oper. Res. 2012, 220, 591–609. [Google Scholar] [CrossRef]
  23. Feichtinger, G.; Hartl, R.F.; Sethi, S.P. Dynamic optimal control models in advertising: Recent developments. Manag. Sci. 1994, 40, 195–226. [Google Scholar] [CrossRef]
  24. Jorgensen, S.; Zaccour, G. Differential Games in Marketing. International Series in Quantitative Marketing; Springer: New York, NY, USA, 2004. [Google Scholar]
  25. He, X.; Prasad, A.; Sethi, S.P.; Gutierrez, G.J. A survey of Stackelberg differential game models in supply and marketing channels. J. Syst. Sci. Syst. Eng. 2007, 16, 385–413. [Google Scholar] [CrossRef]
  26. Aust, G.; Buscher, U. Vertical cooperative advertising in a retailer duopoly. Comput. Ind. Eng. 2014, 72, 247–254. [Google Scholar] [CrossRef]
  27. Jørgensen, S.; Zaccour, G. A survey of game-theoretic models of cooperative advertising. Eur. J. Oper. Res. 2014, 237, 1–14. [Google Scholar] [CrossRef]
  28. Prasad, A.; Sethi, S.P.; Naik, P.A. Optimal Control of an Oligopoly Model of Advertising. IFAC Proc. Vol. 2009, 42, 91–96. [Google Scholar] [CrossRef]
  29. Helbing, D.; Brockmann, D.; Chadefaux, T.; Donnay, K.; Blanke, U.; Woolley-Meza, O.; Moussaid, M.; Johansson, A.; Krause, J.; Schutte, S.; et al. Saving Human Lives: What Complexity Science and Information Systems can Contribute. Phys. Soc. 2015, 158, 735–781. [Google Scholar] [CrossRef] [PubMed]
  30. Yeung, D.W.; Petrosyan, L.A. Subgame Consistent Economic Optimization: An Advanced Cooperative Dynamic Game Analysis; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2011. [Google Scholar]
  31. Yeung, D. On differential games with a feedback Nash equilibrium. J. Optim. Theory Appl. 1994, 82, 181–188. [Google Scholar] [CrossRef]
  32. Bellman, R. Dynamic Programming; Dover Publications: New York, NY, USA, 2003. [Google Scholar]
  33. Petrosjan, L.A. Differential Pursuit Games; World Scientific Publishing Co Pte Ltd.: Singapore, 1993. [Google Scholar]
  34. Yeung, D.W.; Petrosjan, L.A. Cooperative Stochastic Differential Games; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2006. [Google Scholar]
  35. Yeung, D.W.; Petrosyan, L.A. Subgame consistent cooperative solution of dynamic games with random horizon. J. Optim. Theory Appl. 2011, 150, 78–97. [Google Scholar] [CrossRef]
  36. Shi, L.; Petrosian, O. A Dynamic Oligopoly Marketing Model of Advertising. In Proceedings of the 2017 International Conference on Eleventh International Conference “Game Theory and Management”, St. Petersburg, Russia, 28–30 June 2017; pp. 207–223. [Google Scholar]
Figure 1. The behavior of players in the game with truncated information can be modeled using a series of truncated subgames Γ k ( x k , 0 , k , k + T ¯ ) , k = 1 , , N T ¯ .
Figure 1. The behavior of players in the game with truncated information can be modeled using a series of truncated subgames Γ k ( x k , 0 , k , k + T ¯ ) , k = 1 , , N T ¯ .
Mathematics 07 01239 g001
Figure 2. The length of each blue oval is the realization of the infinite horizon.
Figure 2. The length of each blue oval is the realization of the infinite horizon.
Mathematics 07 01239 g002
Figure 3. The feedback Nash equilibrium of the noncooperative case in the initial game (solid line) and the feedback Nash equilibrium of noncooperative case in the game model with dynamic updating (dashed line).
Figure 3. The feedback Nash equilibrium of the noncooperative case in the initial game (solid line) and the feedback Nash equilibrium of noncooperative case in the game model with dynamic updating (dashed line).
Mathematics 07 01239 g003
Figure 4. Optimal cooperative strategies in the cooperative case of the initial game (solid line) and optimal cooperative strategies in the cooperative game model with dynamic updating (dashed line).
Figure 4. Optimal cooperative strategies in the cooperative case of the initial game (solid line) and optimal cooperative strategies in the cooperative game model with dynamic updating (dashed line).
Mathematics 07 01239 g004
Figure 5. Noncooperative trajectory in the initial game (solid line) and resulting noncooperative trajectory (dashed line) in the game model with dynamic updating.
Figure 5. Noncooperative trajectory in the initial game (solid line) and resulting noncooperative trajectory (dashed line) in the game model with dynamic updating.
Mathematics 07 01239 g005
Figure 6. Cooperative trajectory in the initial game (solid line) and resulting cooperative trajectory (dashed line) in the game model with dynamic updating.
Figure 6. Cooperative trajectory in the initial game (solid line) and resulting cooperative trajectory (dashed line) in the game model with dynamic updating.
Mathematics 07 01239 g006
Figure 7. Noncooperative outcomes in the initial game (solid line) and resulting noncooperative outcomes (dashed line) in the game model with dynamic updating.
Figure 7. Noncooperative outcomes in the initial game (solid line) and resulting noncooperative outcomes (dashed line) in the game model with dynamic updating.
Mathematics 07 01239 g007
Figure 8. Characteristic function in the initial game model (solid line) and resulting characteristic function (dashed line) in the game model with dynamic updating.
Figure 8. Characteristic function in the initial game model (solid line) and resulting characteristic function (dashed line) in the game model with dynamic updating.
Mathematics 07 01239 g008
Figure 9. Shapley value in the initial game model (solid line) and the resulting Shapley value (dashed line) in the game model with dynamic updating.
Figure 9. Shapley value in the initial game model (solid line) and the resulting Shapley value (dashed line) in the game model with dynamic updating.
Mathematics 07 01239 g009
Figure 10. The resulting noncooperative trajectory with a fixed information horizon (dashed line) and the resulting noncooperative trajectory with a random information horizon (solid line).
Figure 10. The resulting noncooperative trajectory with a fixed information horizon (dashed line) and the resulting noncooperative trajectory with a random information horizon (solid line).
Mathematics 07 01239 g010
Figure 11. The resulting noncooperative outcomes with a fixed information horizon (dashed line) and the resulting noncooperative outcomes with a random information horizon (solid line).
Figure 11. The resulting noncooperative outcomes with a fixed information horizon (dashed line) and the resulting noncooperative outcomes with a random information horizon (solid line).
Mathematics 07 01239 g011
Figure 12. The resulting cooperative trajectory with a fixed information horizon (dashed line) and the resulting cooperative trajectory with a random information horizon (solid line).
Figure 12. The resulting cooperative trajectory with a fixed information horizon (dashed line) and the resulting cooperative trajectory with a random information horizon (solid line).
Mathematics 07 01239 g012
Table 1. List of variables and parameters.
Table 1. List of variables and parameters.
NotationExplanation
x i k 1 Market share of firm i I 1 , , n at stage k.
u i k 0 Advertising effort rate of firm i I at stage k.
ρ i > 0 Advertising effectiveness parameter of firm i I .
δ > 0 Churn parameter.
m i > 0 Industry sales multiplied by the per unit profit margin of firm i I .
C ( u i k ) Cost of advertising of firm i I at stage k, parameterized by ( u i k ) 2 .

Share and Cite

MDPI and ACS Style

Petrosian, O.; Shi, L.; Li, Y.; Gao, H. Moving Information Horizon Approach for Dynamic Game Models. Mathematics 2019, 7, 1239. https://doi.org/10.3390/math7121239

AMA Style

Petrosian O, Shi L, Li Y, Gao H. Moving Information Horizon Approach for Dynamic Game Models. Mathematics. 2019; 7(12):1239. https://doi.org/10.3390/math7121239

Chicago/Turabian Style

Petrosian, Ovanes, Lihong Shi, Yin Li, and Hongwei Gao. 2019. "Moving Information Horizon Approach for Dynamic Game Models" Mathematics 7, no. 12: 1239. https://doi.org/10.3390/math7121239

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop