Next Article in Journal
Cheap Talk with Transparent and Monotone Motives from a Seller to an Informed Buyer
Previous Article in Journal
Information Transmission and Countervailing Biases in Organizations
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Communication

A Controlled Discrete-Time Queueing System as a Model for the Orders of Two Competing Companies

Department of Mathematics and Industrial Engineering, Polytechnique Montréal, C.P. 6079, Succursale Centre-ville, Montréal, QC H3C 3A7, Canada
Games 2024, 15(3), 19; https://doi.org/10.3390/g15030019
Submission received: 30 April 2024 / Revised: 26 May 2024 / Accepted: 27 May 2024 / Published: 29 May 2024
(This article belongs to the Section Applied Game Theory)

Abstract

:
We consider two companies that are competing for orders. Let X 1 ( n ) denote the number of orders processed by the first company at time n, and let τ ( k ) be the first time that X 1 ( n ) < j or X 1 ( n ) = r , given that X 1 ( 0 ) = k . We assume that { X 1 ( n ) , n = 0 , 1 , } is a controlled discrete-time queueing system. Each company is using some control to increase its share of orders. The aim of the first company is to maximize the expected value of τ ( k ) , while its competitor tries to minimize this expected value. The optimal solution is obtained by making use of dynamic programming. Particular problems are solved explicitly.
AMS Subject Classification:
Primary 93E20; Secondary 60K25

1. Introduction

Many papers have been published on the optimal control of queueing systems. Often the authors assume that it is possible to control the service rate; see, for example, Laxmi et al. [1], Chen et al. [2] and Tian et al. [3]. In other papers, the aim is to control the arrival of customers into the system; see Wu et al. [4]. Sometimes, it is assumed that there are different types of customers; see Wen et al. [5] and Dudin et al. [6]. Another problem that is often considered is to determine the optimal number of servers in the system; see Asadzadeh et al. [7].
In this paper, we consider a single-server queueing model in discrete time. Orders arrive one at a time. We assume that the time needed for an order to arrive is a random variable A that follows a geometric distribution with parameter 2 p A , so that
P [ A = n ] = ( 1 2 p A ) n 1 2 p A for n = 1 , 2 ,
Moreover, the service time is a random variable S having a geometric distribution with parameter p S .
There are two competing companies. In equilibrium, orders arrive at each company according to a random variable A i having a geometric distribution with parameter p A , for i = 1 , 2 .
Instead of solving an optimal control problem, our aim in this paper is to present a stochastic dynamic game that can serve as a model for the behavior of companies competing for orders. We must, of course, make simplifying assumptions in order to obtain a mathematically tractable problem. However, we believe that the model is realistic enough to be useful.
Let X 1 ( n ) be the number of orders being processed by the first company at time n, and let
τ ( k ) [ = τ ( k , j , r ) ] : = inf { n 0 : X 1 ( n ) < j or X 1 ( n ) = r X 1 ( 0 ) = k } ,
where j k r and j , k , r N .
Now, suppose that each company can try to use some form of control in order to increase its share of the orders. More precisely, we assume that the first company uses the control u n { 0 , p A } at time n to increase the arrival rate of its orders. Similarly, the second company uses v n { p A , 0 } at time n to decrease the order arrival rate of its competitor.
Remark 1.
Writing that v n = p A means that Company 2 is making some efforts (for instance, by lowering its prices) in order to reduce to zero the rate at which Company 1 receives orders, unless this company also tries to increase or at least keep its share of the orders.
Let
p A ( n ) : = p A + u n + v n for n = 1 , 2 ,
We define the cost function
J ( k ) = n = 0 τ ( k ) 1 q 1 u n 2 + q 2 v n 2 + λ ( v n u n ) ,
where q 1 , q 2 and λ are positive constants. We look for the values u n * and v n * of the control variables that are such that the expected value of J ( k ) is minimized.
The idea behind the cost function is as follows: since the parameter λ is positive, if r is large, then Company 1 wants to maximize the time it stays in business. To do this, it would like to use the control u n = p A ; however, this leads to quadratic control costs. Company 2, for its part, would like to use v n = p A in order to bankrupt Company 1 as quickly as possible, but this also entails quadratic control costs.
Problems in which the optimizers try to minimize or maximize the expected value of a certain cost function until a given random event occurs are known as homing problems. Whittle [8] considered the case when the optimizer controls an n-dimensional diffusion process until it leaves a given subset of R n . Rishel [9] also treated the homing problems for n-dimensional diffusion processes; these processes were more realistic models for the wear of devices than those proposed by various authors.
The author has recently extended homing problems to the optimal control of queueing systems in continuous time; see [10,11,12]. In these three papers, the aim was to determine the optimal number of servers working at time t. He also published a paper ([13]) on a homing problem for a controlled random walk with two optimizers.
Next, we define the value function
F ( k ) = min ( u n , v n ) 0 n τ ( k ) 1 E [ J ( k ) ] .
The function F ( k ) is the expected cost incurred (which can sometimes be a reward) if both optimizers choose the optimal value of u n and v n between the initial time n = 0 and time τ ( k ) 1 .
In Section 2, the dynamic programming equation satisfied by the value function F ( k ) will be derived, and a particular problem will be solved explicitly. In Section 3, the problem formulation will be modified. We will then assume that the value of u n is known, and we will look for the value of v n that maximizes the expected value of a certain cost function. Concluding remarks will be made in Section 4.

2. Dynamic Programmic Equation

We will derive the dynamic programming equation satisfied by F ( k ) . We have
F ( k ) = min ( u n , v n ) 0 n τ ( k ) 1 { q 1 u 0 2 + q 2 v 0 2 + λ ( v 0 u 0 )   + E n = 1 τ ( k ) 1 q 1 u n 2 + q 2 v n 2 + λ ( v n u n ) } .
Then, making use of Bellman’s principle of optimality (see [14]), we can write that
F ( k ) = min ( u 0 , v 0 ) q 1 u 0 2 + q 2 v 0 2 + λ ( v 0 u 0 ) + E F ( X 1 ( 1 ) ) .
Indeed, whatever the two optimizers decide to do at time n = 0 , the decisions they make from time n = 1 to time τ ( k ) 1 must be optimal.
Remark 2.
Equation (7) is valid because of our assumptions that A i , for i = 1 , 2 , and S have a geometric distribution. Indeed, as is well known, this distribution possesses the memoryless property. If we assume instead that these random variables have any other (discrete) distribution, then we will have to take the past into account, rendering the optimization problem almost intractable.
Furthermore, we have
E F ( X 1 ( 1 ) ) = F ( k + 1 ) p A ( 0 ) ( 1 p S ) + F ( k 1 ) [ 1 p A ( 0 ) ] p S     + F ( k ) p A ( 0 ) p S + [ 1 p A ( 0 ) ] ( 1 p S ) .
Hence, we can state the following proposition.
Proposition 1.
The value function F ( k ) satisfies the dynamic programming equation (DPE)
F ( k ) = min ( u 0 , v 0 ) { q 1 u 0 2 + q 2 v 0 2 + λ ( v 0 u 0 )     + F ( k + 1 ) p A ( 0 ) ( 1 p S ) + F ( k 1 ) [ 1 p A ( 0 ) ] p S     + F ( k ) p A ( 0 ) p S + [ 1 p A ( 0 ) ] ( 1 p S ) } .
Moreover, we have the boundary condition F ( k ) = 0 if k < j or k = r .
Remark 3.
(i) 
We take for granted that each optimizer does not know what the other has decided to do. In Section 3, we will assume that Company 2 knows the decision made by Company 1.
(ii) 
There are four possibilities for ( u 0 , v 0 ) : ( 0 , 0 ) , ( 0 , p A ) , ( p A , 0 ) and ( p A , p A ) . If we solve the difference equation corresponding to each possible value of ( u 0 , v 0 ) , we actually obtain the value of the function F ( k ) if the optimizers choose the same value of ( u 0 , v 0 ) for any k. Hence, we cannot obtain the value function and/or the optimal controls for any value of k by comparing the four expressions for F ( k ) obtained by solving the four difference equations.
(iii) 
We can write that ( u n , v n ) = ( u n ( k ) , v n ( k ) ) . The number of possible pairs ( u 0 ( k ) , v 0 ( k ) ) for j k < r is equal to 4 r j . If we have the values of ( u 0 ( k ) , v 0 ( k ) ) for k = j , , r 1 , we can solve a system of r j linear equations to obtain the corresponding values of F ( k ) for any k. If r j is small, it is a simple matter to consider all the possible values of ( u 0 ( k ) , v 0 ( k ) ) and compute the function F ( k ) for k = j , , r 1 . We can then determine the optimal controls and the associated value function.
In the following subsection, a particular problem will be solved explicitly.

An Example

Suppose that q 1 = q 2 = λ = 1 , j = 1 and r = 3 . Moreover, we take p A = 1 / 4 and p S = 1 / 2 . The values of F ( 1 ) and F ( 2 ) for the 4 2 = 16 possible choices for ( u 0 ( 1 ) , v 0 ( 1 ) ) and ( u 0 ( 2 ) , v 0 ( 2 ) ) are presented in Table 1.
For example, if ( u 0 ( 1 ) , v 0 ( 1 ) ) = ( 0 , 0 ) and ( u 0 ( 2 ) , v 0 ( 2 ) ) = ( 0 , p A ) (Case no. 5), then we have p A ( 0 ) = p A if k = 1 and p A ( 0 ) = 0 if k = 2 . We must solve the system of linear equations
(10) F ( 1 ) = 1 8 F ( 2 ) + 1 2 F ( 1 ) (11) F ( 2 ) = 3 16 + 1 2 F ( 2 ) + 1 2 F ( 1 ) ,
whose solution is F ( 1 ) = 1 / 8 and F ( 2 ) = 1 / 2 .
We see that the optimal strategy is to choose Case no. 14; that is, we take ( u 0 ( 1 ) , v 0 ( 1 ) ) = ( p A , 0 ) and ( u 0 ( 2 ) , v 0 ( 2 ) ) = ( p A , p A ) .
Remark 4.
The four difference equations that must be solved, subject to the boundary conditions F ( 0 ) = F ( 3 ) = 0 , are given below. We denote their solutions by F i ( k ) , for i = 1 , 2 , 3 , 4 .
(1)
( u 0 , v 0 ) = ( 0 , 0 ) :
F ( k ) = 1 8 F ( k + 1 ) + 3 8 F ( k 1 ) + 1 2 F ( k ) .
We easily find that the solution is F 1 ( k ) 0 .
(2)
( u 0 , v 0 ) = ( p A , 0 ) :
F ( k ) = 3 8 + 1 4 F ( k + 1 ) + 1 4 F ( k 1 ) + 1 2 F ( k ) .
We find that
F 2 ( k ) = 3 8 k ( 3 + k ) .
(3)
( u 0 , v 0 ) = ( 0 , p A ) :
F ( k ) = 3 8 + 1 2 F ( k 1 ) + 1 2 F ( k ) .
Because p A ( 0 ) = 0 , X 1 ( n ) cannot reach the value 3 if k = 1 or 2. The solution that satisfies the boundary condition F ( 0 ) = 0 is
F 3 ( k ) = 3 8 k .
(4)
( u 0 , v 0 ) = ( p A , p A ) :
F ( k ) = 3 8 + 1 8 F ( k + 1 ) + 3 8 F ( k 1 ) + 1 2 F ( k ) .
We have
F 4 ( k ) = 3 52 ( 3 + 3 k + 1 26 k ) .
The functions F i ( k ) , for i = 2 , 3 , 4 , are shown in Figure 1. Moreover, the values of F i ( k ) for k = 1 , 2 and i = 1 , 2 , 3 , 4 are presented in Table 2.
Notice that F 1 ( k ) corresponds to Case no. 1, while F 4 ( k ) corresponds to Case no. 16. We observe that none of the functions F 1 ( k ) , F 4 ( k ) is the value function. However, F 4 ( k ) yields values which are quite close to those obtained with the value function. Therefore, if r j is large, so that the number of equations to consider is also large, then a suboptimal solution can be obtained by assuming that ( u 0 ( k ) , v 0 ( k ) ) will be the same for any k.

3. Optimal Control When u n Is Known

In this section, we assume that Company 2 knows the strategy of Company 1, and tries to maximize the expected value of the following cost function:
C ( k ) = n = 0 τ ( k ) 1 q 1 u n + q 2 v n + K [ X 1 ( τ ( k ) ) ] .
The terminal cost function K ( · ) is defined by
K [ X 1 ( τ ( k ) ) ] = K 1 if X 1 ( τ ( k ) ) = r , K 2 if X 1 ( τ ( k ) ) = j 1 ,
where K 1 < 0 and K 2 > 0 . The constant r N could be the maximum number of orders that Company 1 can process at the same time.
Suppose that Company 1 uses the control u n ( k ) = p A for any n and any k. Company 2 must decide whether to choose v n ( k ) = 0 or p A . We define the value function
V ( k ) = max v n , 0 n τ ( k ) 1 E [ C ( k ) ] .
Proceeding as in the previous section, we can prove the following proposition.
Proposition 2.
The value function V ( k ) satisfies the dynamic programming equation
V ( k ) = max v 0 { p A , 0 } { q 1 p A + q 2 v 0 + V ( k + 1 ) p A ( 0 ) ( 1 p S )     + V ( k 1 ) [ 1 p A ( 0 ) ] p S     + V ( k ) p A ( 0 ) p S + [ 1 p A ( 0 ) ] ( 1 p S ) } ,
where p A ( 0 ) = 2 p A + v 0 . Furthermore, the function V ( k ) is such that V ( r ) = K 1 and V ( j 1 ) = K 2 .
There are now 2 r j possible strategies for Company 2. If r j is small, we can proceed as in the previous section and calculate the expected value of C ( k ) for each possible strategy.
Assume that q 1 = q 2 = 1 , p A = 1 / 4 , p S = 1 / 2 , K 1 = 10 , K 2 = 10 , r = 5 and j = 1 . We present in Table 3 the value of V ( 1 ) , , V ( 4 ) for each of the 16 possible strategies that Company 2 can choose. These values are obtained by solving the system of four linear equations
V ( k ) = 1 4 + v 0 ( k ) + V ( k + 1 ) 1 2 + v 0 ( k ) 1 2     + V ( k 1 ) 1 2 v 0 ( k ) 1 2 + V ( k ) 1 2 ,
for k = 1 , 2 , 3 , 4 , together with the boundary conditions V ( 0 ) = 10 and V ( 5 ) = 10 . We conclude that the optimal strategy is the one that corresponds to Case no. 15; that is, v 0 ( 1 ) = 0 and v 0 ( k ) = p A for k = 2 , 3 , 4 .
Remark 5.
As in the previous section, we can obtain at least a suboptimal solution which is close the optimal one by solving the difference equations obtained by assuming first that v 0 ( k ) 0 , namely
V ( k ) = q 1 p A + V ( k + 1 ) 2 p A ( 1 p S ) + V ( k 1 ) ( 1 2 p A ) p S   + V ( k ) 2 p A p S + ( 1 2 p A ) ( 1 p S ) ,
and then that v 0 ( k ) p A :
V ( k ) = q 1 p A q 2 p A + V ( k + 1 ) p A ( 1 p S ) + V ( k 1 ) ( 1 p A ) p S   + V ( k ) p A p S + ( 1 p A ) ( 1 p S ) .
The first equation corresponds to Case no. 1, and the second one to Case no. 16. We find that the solutions to these equations in the particular case considered above are, respectively,
V 1 ( k ) = 10 3 2 k 1 2 k 2
and
V 16 ( k ) = 1220 121 10 121 3 k .
Notice that the solution obtained when v 0 ( k ) 0 (Case no. 1) actually gives the minimum of the expected value of the cost function C ( k ) . Moreover, we see in Table 3 that the choice that corresponds to Case no. 11 almost yields the optimal solution that we are looking for.

4. Conclusions

In this paper, a homing problem for a queueing model in discrete time has been considered. The problem can be seen as a dynamic game because there are two optimizers with opposing objectives.
In Section 2, dynamic programming was used to derive the equation satisfied by the value function. From this equation, one can deduce the optimal values of the control variables. However, we have seen that, in order to do this, one has to solve a possibly large number of systems of linear equations, subject to the appropriate boundary conditions. Although solving each system is straightforward, repeating this procedure a large number of times can become tedious. We have also seen that it is possible to obtain a good suboptimal solution to our problem fairly quickly.
In Section 3, the problem formulation was modified. We assumed that the strategy of Company 1 was known, and we looked for the strategy that Company 2 should adopt to maximize the expected value of a certain cost function. We treated the case when the control variable u n is always equal to p A . However, the same type of analysis could be carried out for any choice of u n . In particular, we could find the optimal strategy of Company 2 if u n 0 .
In theory, we could easily extend the problems considered to the case when each optimizer can choose between more than two possible values for the variable it controls. The calculations would, however, become quite complex. One could possibly use numerical simulations to determine the optimal solutions. Indeed, instead of solving a large number of difference equations, simulating the proposed model can enable us to determine the optimal solution by computing the value function for each simulation. Simulating geometric random variables is not a difficult task.
Finally, it is also possible to consider optimal control problems for queueing models in continuous time with two optimizers.

Funding

This research was supported by the Natural Sciences and Engineering Research Council of Canada.

Data Availability Statement

No data was used for this research.

Acknowledgments

The author wishes to thank the anonymous reviewers of this paper for their constructive comments.

Conflicts of Interest

The author reports that there are no competing interests to declare.

References

  1. Laxmi, P.V.; Bhavani, E.G.; Jyothsna, K. Analysis of Markovian queueing system with second optional service operating under the triadic policy. OPSEARCH 2023, 60, 256–275. [Google Scholar] [CrossRef]
  2. Chen, G.; Liu, Z.; Xia, L. Event-based optimization of service rate control in retrial queues. J. Oper. Res. Soc. 2023, 74, 979–991. [Google Scholar] [CrossRef]
  3. Tian, R.; Su, S.; Zhang, Z.G. Equilibrium and social optimality in queues with service rate and customers’ joining decisions. Qual. Technol. Quant. Manag. 2024, 21, 1–34. [Google Scholar] [CrossRef]
  4. Wu, C.-H.; Yang, D.-Y.; Yong, C.-R. Performance evaluation and bi-objective optimization for F-policy queue with alternating service rates. J. Ind. Manag. Optim. 2023, 19, 3819–3839. [Google Scholar] [CrossRef]
  5. Wen, J.; Geng, N.; Xie, X. Optimal insertion of customers with waiting time targets. Comput. Oper. Res. 2020, 122, 105001. [Google Scholar] [CrossRef]
  6. Dudin, A.; Dudin, S.; Dudina, O. Analysis of a queueing system with mixed service discipline. Methodol. Comput. Appl. Prob. 2023, 25, 57. [Google Scholar] [CrossRef]
  7. Asadzadeh, S.; Akhavan, B.; Akhavan, B. Multi-objective optimization of Gas Station performance using response surface methodology. Int. J. Qual. Reliab. Manag. 2021, 38, 465–483. [Google Scholar] [CrossRef]
  8. Whittle, P. Optimization over Time; Wiley: Chichester, UK, 1982; Volume I. [Google Scholar]
  9. Rishel, R. Controlled wear process: Modeling optimal control. IEEE Trans. Automat. Control 1991, 36, 1100–1102. [Google Scholar] [CrossRef]
  10. Lefebvre, M. An optimal control problem for a modified M/G/k queueing system. In Proceedings of the Workshop on Intelligent Information Systems, Chişinǎu, Republic of Moldova, 19–21 October 2023; pp. 141–148. [Google Scholar]
  11. Lefebvre, M. Reducing the size of a waiting line optimally. WSEAS Trans. Syst. Control 2023, 18, 342–345. [Google Scholar] [CrossRef]
  12. Lefebvre, M.; Yaghoubi, R. Optimal control of a queueing system. 2024; Submitted for publication. [Google Scholar]
  13. Lefebvre, M. A discrete-time homing problem with two optimizers. Games 2023, 14, 68. [Google Scholar] [CrossRef]
  14. Bellman, R. Dynamic Programming; Princeton University Press: Princeton, NJ, USA, 1957. [Google Scholar]
Figure 1. Functions F 2 ( k ) (solid line), F 3 ( k ) (dotted line) and F 4 ( k ) (dashed line) for k [ 1 , 2 ] .
Figure 1. Functions F 2 ( k ) (solid line), F 3 ( k ) (dotted line) and F 4 ( k ) (dashed line) for k [ 1 , 2 ] .
Games 15 00019 g001
Figure 2. Functions V 1 ( k ) (solid line) and V 16 ( k ) (dotted line) for k [ 1 , 4 ] .
Figure 2. Functions V 1 ( k ) (solid line) and V 16 ( k ) (dotted line) for k [ 1 , 4 ] .
Games 15 00019 g002
Table 1. Values of F ( 1 ) and F ( 2 ) for the 16 possible choices for the control variables.
Table 1. Values of F ( 1 ) and F ( 2 ) for the 16 possible choices for the control variables.
No. u 0 ( 1 ) v 0 ( 1 ) u 0 ( 2 ) v 0 ( 2 ) F ( 1 ) F ( 2 )
1000000
2 p A 000 0.6 0.45
30 p A 00 0.375 0.28125
400 p A 0 0.107 0.4286
5000 p A 0.125 0.5
6 p A p A 00 0.923 0.692
7 p A 0 p A 0 0.75 0.75
8 p A 00 p A 1.125 1.5
90 p A p A 0 0.375 0.5625
100 p A 0 p A 0.375 0.75
1100 p A p A 0.231 0.923
12 p A p A p A 0 0.964 0.857
13 p A p A 0 p A 1.125 1.5
14 p A 0 p A p A 1.2 1.65
150 p A p A p A 0.375 1.03125
16 p A p A p A p A 1.154 1.615
Table 2. Values of F i ( k ) for k = 1 , 2 and i = 1 , 2 , 3 , 4 .
Table 2. Values of F i ( k ) for k = 1 , 2 and i = 1 , 2 , 3 , 4 .
i F i ( 1 ) F i ( 2 )
100
2 0.375 0.75
3 0.75 0.75
4 1.154 1.615
Table 3. Values of V ( 1 ) , V ( 4 ) for the 16 possible choices for the control variables v 0 ( 1 ) , , v 0 ( 4 ) .
Table 3. Values of V ( 1 ) , V ( 4 ) for the 16 possible choices for the control variables v 0 ( 1 ) , , v 0 ( 4 ) .
No. v 0 ( 1 ) v 0 ( 2 ) v 0 ( 3 ) v 0 ( 4 ) V ( 1 ) V ( 2 ) V ( 3 ) V ( 4 )
10000851 4
2 p A 000 8.92 5.69 1.46 3.77
30 p A 00 9.36 7.73 2.82 3.09
400 p A 0 9.56 8.11 5.67 1.67
5000 p A 9.29 7.57 4.86 1.14
6 p A p A 00 9.45 7.81 2.81 3.06
7 p A 0 p A 0 9.52 8.08 5.64 1.68
8 p A 00 p A 9.42 7.68 4.95 1.21
90 p A p A 0 10.13 9.26 6.65 1.17
100 p A 0 p A 10962
1100 p A p A 10.47 9.93 8.40 3.80
12 p A p A p A 0 9.72 8.87 6.31 1.34
13 p A p A 0 p A 9.67 8.69 5.76 1.82
14 p A 0 p A p A 9.84 9.35 7.86 3.40
150 p A p A p A 10.49 9.98 8.44 3.83
16 p A p A p A p A 9.83 9.34 7.85 3.39
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lefebvre, M. A Controlled Discrete-Time Queueing System as a Model for the Orders of Two Competing Companies. Games 2024, 15, 19. https://doi.org/10.3390/g15030019

AMA Style

Lefebvre M. A Controlled Discrete-Time Queueing System as a Model for the Orders of Two Competing Companies. Games. 2024; 15(3):19. https://doi.org/10.3390/g15030019

Chicago/Turabian Style

Lefebvre, Mario. 2024. "A Controlled Discrete-Time Queueing System as a Model for the Orders of Two Competing Companies" Games 15, no. 3: 19. https://doi.org/10.3390/g15030019

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop