Next Article in Journal
Hour-Ahead Photovoltaic Output Forecasting Using Wavelet-ANFIS
Next Article in Special Issue
Monte Carlo Algorithms for the Extracting of Electrical Capacitance
Previous Article in Journal
Learning Impulsive Pinning Control of Complex Networks
Previous Article in Special Issue
Bounds on the Rate of Convergence for MtX/MtX/1 Queueing Models
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Optimal Prefetching in Random Trees

1
Deloitte India (Offices of the US), Hyderabad 500032, Telangana, India
2
Inria, University of Montpellier, 34095 Montpellier, France
3
Inria, Université Côte d’Azur, 06902 Sophia Antipolis, France
*
Author to whom correspondence should be addressed.
The work of this author was performed partly as a Master’s student at IISER Mohali, India, and as an intern at Inria.
Mathematics 2021, 9(19), 2437; https://doi.org/10.3390/math9192437
Submission received: 31 August 2021 / Revised: 21 September 2021 / Accepted: 22 September 2021 / Published: 1 October 2021
(This article belongs to the Special Issue Stability Problems for Stochastic Models: Theory and Applications II)

Abstract

:
We propose and analyze a model for optimizing the prefetching of documents, in the situation where the connection between documents is discovered progressively. A random surfer moves along the edges of a random tree representing possible sequences of documents, which is known to a controller only up to depth d. A quantity k of documents can be prefetched between two movements. The question is to determine which nodes of the known tree should be prefetched so as to minimize the probability of the surfer moving to a node not prefetched. We analyzed the model with the tools of Markov decision process theory. We formally identified the optimal policy in several situations, and we identified it numerically in others.

1. Introduction

Prefetching is a basic technique underlying many computer science applications. Its main purpose is to reduce the time needed to access some information by loading it in advance and concurrently with the process that needs this information. From prefetching of data and code in CPUs and memory architectures, to prefetching of web pages and video segments in Internet-based applications, this technique is ubiquitous. Yet, the technique fundamentally involves a tradeoff between access latency and the consumption of resources (memory, network), and the optimization of this tradeoff is not completely understood.
Clearly, the issue here is randomness: the entity in charge of prefetching, let us call it the “controller”, does not know in advance what is the precise data access sequence of the process needing the data. It must therefore make decisions based on the current state of said process and its knowledge of the possible evolution. The adequate formalism for modeling optimal decisions in such a context is that of Markov Decision Processes (MDPs). The principle of using Markov decision processes to optimize prefetching in the context of video applications was first demonstrated in [1,2]. The model was extended in [3,4] and further extended in [5].
The basic principle of these models is that the set of (video) documents to be viewed is represented by a directed graph. The nodes represent the documents, and the edges represent the possible transitions: which documents can be viewed after the viewing of the current document is completed. The edges can be labeled with probabilities or frequencies. A random “surfer” alternates viewing periods and moves to another node/document according to the probabilities of the edges. The controller knows where the surfer stands and knows all about the graph, but does not know which way the surfer will go: only the odds. Its decision is to choose which nodes to download during the time the surfer views the current document. The amount of nodes that can be downloaded is constrained by network resources and is called the “prefetching budget”. The amount of storage memory available to the controller is assumed to be sufficient: no memory management is involved in the decision. The criterion to be minimized is typically the average number of times the surfer moves to a document that has not been prefetched: this is a measure of the user’s dissatisfaction. The criterion might also involve some measure of the waste of network and memory resources. An optimal policy can, in principle, be computed using dynamic programming.
In practical situations, the probabilities that the surfer moves to some new document after viewing the current document are not known a priori. However, these probabilities can be learned from data using Markov models, as in [6,7,8,9,10,11]. Moreover, the optimal control of a prefetching agent can be approximated using machine learning techniques such as reinforcement learning, as in [2]. A way to evaluate the efficiency of a machine learning algorithm is to test it on a problem for which the exact solution is known. The purpose of this paper is to provide such a benchmark situation, by determining the optimal policy and the minimal possible cost it induces, to which heuristics and learning algorithms can be compared.
While these previous modeling attempts demonstrated that the MDP formalism is flexible enough to take into account many features of a real system, they also illustrate that finding an optimal policy is a complex problem. Indeed, computing an optimal prefetching policy is very hard in general, that is when the graph of documents does not have a particular property. In [12], the authors studied the feasibility variant of the problem. There, it was assumed that the controller has a prefetching budget k, representing the number of documents that can be prefetched while some document is viewed. The question is to decide whether k is large enough so that there exists a policy that prefetches all nodes of a graph before the random surfer tries to access them. This is a subproblem of the Markov optimization model: if such a policy exists, the MDP model should find it as a policy that realizes a zero cost. The results of [12,13] concluded that finding the minimum possible k is difficult when the graph is general. Computing the optimal policy in the corresponding MDP must be even more difficult.
However, if the underlying graph is a tree, it was proven in [12] that the minimal budget k that ensures the existence of a costless policy can be computed in polynomial time. The corresponding prefetching strategy is also easy to compute and has the feature of being “connected”. This property, which we also call “greedy”, means that the controller can choose the documents to download in the set of neighbors (in the document graph) of the documents already downloaded.
The models and results reviewed thus far assumed that the complete space of documents is known to the controller. This ideal situation may be either unrealistic or undesirable. For instance, if the documents are pages on the global web, storing all the knowledge about this graph is probably impossible and also useless since the web surfer will not visit all the graph during a surfing session. Furthermore, since the complexity of the decision grows exponentially with the size of the graph, it may help the controller to limit, on purpose, the size of the known graph to a neighborhood of the current document.
The current literature lacks a model for the optimal prefetching problem, which features a dynamic graph of documents. It also lacks situations where an optimal control can be formally (and not just numerically) identified, even when the underlying graph of documents is static. We fill these gaps in two ways. First, we propose a new optimal prefetching model in which the graph of documents is dynamic. We focused on trees, since those are the simplest graphs, with the potential for having computable solutions as the literature review suggests. Second, we compute exactly the optimal control for some instances of this model. We proceed with an informal description of the model, then we highlight our contribution.

1.1. The Model

We propose to use the modeling described above, but replace the graph of known documents with a tree of depth d. The root of this tree is the current position of the surfer. The tree represents all the possible sequences of d moves of the surfer. After the surfer has moved to one of the neighbors of its current position, a discovery phase adds a new generation of documents at depth d. The rest of the tree is then ignored: the possibility that the surfer moves back to a node already viewed is neglected, as well as the possibility that several paths exist from one node to another. If any of these possibilities happens in practice, the task will become easier to the controller.
In the discovery phase, we assume that a random number of new documents is attached to every leaf of the current tree, with a uniform distribution between 1 and some integer p, which we refer to as the “fanout”. In practical graphs of documents, this assumption is not very realistic. The advantages of making such an assumption are that the space of possible configurations will remain finite and that probabilities relative to objects in this space will be easier to write.
As in previous models, the controller is assumed to have a fixed prefetching budget: some integer number k. Given a tree of documents with some nodes already downloaded, the problem is to decide which k nodes to download so as to minimize the cost. We chose as criterion the stationary probability that the surfer moves to a node that is not prefetched. All these elements are converted in the specification of a Markov decision process, with criterion the infinite-horizon average cost. The model has only three parameters: the depth d and fanout p of trees and the prefetching budget k. The question is whether there is a simple rule based on these three parameters that leads to an optimal decision.

1.2. Contribution

The first contribution of this paper is the precise specification of this MDP, in Section 3. This specification is based on sets of trees, presented in Section 2, together with their basic properties.
We then turn to the identification of optimal prefetching policies, in Section 5, Section 6 and Section 7. The results we obtained include: (a) A bound on the optimal cost in general trees; (b) The characterization of optimal policies in trees with depth 1, arbitrary fanout, and arbitrary budget; (c) The characterization of optimal policies in trees with depth 2, arbitrary fanout, and budget 1; (d) An exploration of the optimal policy in trees with depth 2, budget 2 and fanout less than 5. In the process of obtaining these results, we show, in Section 4, the properties of underlying Markov chains on the “shape of trees”, which do not depend on the specific policy used and are of independent interest. We discuss the results and the modeling assumptions in Section 8 and conclude in Section 9.
The main notation that is used throughout the paper is summarized in Table A1 in Appendix A.

2. Preliminaries: Sets of Trees

This section is devoted to the presentation of mathematical objects that are used in the definition of our problem and in its solution.
The state space of the MDP we are about to construct is a set of marked trees. We introduce it now, together with other sets of trees that will be useful in the analysis. We shall use “with fanout p” as a shorthand for “with nodes having between 1 and p sons”.
Definition 1 (Trees and marked trees).
Define:
(a) 
T p , d the set of rooted trees of depth d with fanout p;
(b) 
M p , d the set of rooted trees of depth d with fanout p and a mark in the set { 0 , 1 } ;
(c) 
M p , d + the set of rooted trees of depth d with fanout p and a mark in the set { 0 , 1 } except for leaves that have the mark 0.
These sets are represented mathematically with the following recursive formulas:
T p , 0 = { 0 }
T p , d = { 0 } × S E Q 1 . . p ( T p , d 1 ) d 1
M p , 0 = { 0 , 1 }
M p , d = { 0 , 1 } × S E Q 1 . . p ( M p , d 1 ) d 1
M p , 1 + = M p , 0 × S E Q 1 . . p ( T p , 0 )
M p , d + = M p , 0 × S E Q 1 . . p ( M p , d 1 + ) d 2 .
In these expressions, S E Q 1 . . p ( A ) denotes, in the notation of [14], a sequence of objects in the set A, with the length between 1 and p. In (1), we associate the constant mark “0” with nodes in unmarked trees. At the risk of being confusing, we shall say that a node in marked trees of M p , d or M p , d + is “unmarked” if it has the mark 0. With this convention, we can say that T p , d M p , d + M p , d .
A tree t in M p , d is represented as follows. If the depth is d = 0 , t = ( μ ) where μ is the mark. If d > 0 , t = ( μ , s ) where μ = 0 or μ { 0 , 1 } depending on the set, and s = ( s 1 , , s m ) is a list of length m [ 1 . . p ] . The elements of s are called “subtrees”. The root nodes of these subtrees are called “sons” of t. The following notation will be useful to designate the components of a tree. Figure 1 illustrates this terminology.
Definition 2 (Mark, subtrees, internal nodes, leaves).
For a tree represented as t = ( μ , s ) , let μ ( t ) : M p , d { 0 , 1 } denote the mark of the root and s ( t ) : M p , d S E Q 1 . . k ( M p , d 1 ) denote the list of subtrees of t. Let also inode ( t ) denote the number of internal nodes in t and leaves ( t ) denote the number of leaves in t.
The cardinal of the sets defined in Definition 1 is important to know, in case we want to turn to numerical experiments. From the recursive definition of the different sets of trees, the following result is easily established.
Lemma 1.
Let T p , d , M p , d , and M p , d + denote respectively the cardinals of sets T p , d , M p , d and M p , d + . Then:
  • T p , 0 = 1 , T p , 1 = p and for d 2 ,
    T p , d = m = 1 p ( T p , d 1 ) m = ( T p , d 1 ) p + 1 T p , d 1 T p , d 1 1 ;
  • M p , 0 = 2 and for d 1 ,
    M p , d = 2 m = 1 p ( M p , d 1 ) m = 2 ( M p , d 1 ) p + 1 M p , d 1 M p , d 1 1 ;
  • M p , 1 + = 2 p and for d 2 ,
    M p , d + = 2 m = 1 p ( M p , d 1 + ) m = 2 ( M p , d 1 + ) p + 1 M p , d 1 + M p , d 1 + 1 .
Similar formulas can be established for generating functions of the size of trees in each set. We shall not develop this analysis further. Table 1 shows the values of T p , d , M p , d , and M p , d + for small values of p and d. Clearly, these numbers grow extremely fast with d. The sets remain manageable for small values of p and d. For instance, Figure 2 lists the 6 trees of T 2 , 2 .

3. The MDP Model

In this section, we formally describe the five elements of the prefetching MDP. An MDP model is formally defined by a state space, an action space, transitions, costs, and an evaluation criterion. The state is usually made of numerical variables or discrete structures that summarize the information needed for the following definitions. Actions and transitions specify what controls are allowed to the controller, how they modify the states, and with what probabilities. The cost function, which depends on the states and actions, quantifies the impact of actions. The costs incurred at different time steps are aggregated into a numerical criterion: the objective is to minimize this criterion.

3.1. Prefetching Process Flow

The prefetching process flow is summarized as follows. The current state of the prefetching program is the currently known graph of depth d, together with the knowledge of nodes that have been already prefetched. This is represented as a marked tree of depth d. The surfer is assumed to stand at the root. The controller then prefetches up to k documents, which is represented by marking the corresponding nodes in the tree. Then, the surfer moves randomly to one of the sons, with uniform probabilities among them. If the document corresponding to this node is not already prefetched, then some cost is incurred. Finally, the controller discovers a new generation of nodes. We assume that every possible exploration/discovery of the current subtree is equally likely. After discovery, the controller is back at the beginning of this decision loop.

3.2. State Space and Action Space

According to the prefetching process flow described previously, the state space S of the MDP is M p , d + as defined in Definition 1, since the leaves of a state were just discovered and are not prefetched: their mark is 0. For a given tree t S , the action space A t is the set of all subsets of the vertices of t with cardinal at most k. An action a A t will have the effect of marking the nodes in a. Some actions will not make much sense: those that mark already marked nodes. We choose to include them nevertheless as possible actions, which greatly simplifies the description of the set A t . The parameter k 1 is called the prefetching budget.

3.3. Transitions

We first formalize the transition between the different trees using random variables and set mappings. Then, we quantify the probabilities of these transitions and the costs.
Definition 3 (Prefetching process random variables).
Define:
(a) 
t the random variable denoting the state (tree) of the prefetching MDP; t takes values in M p , d + ;
(b) 
t a the random variable denoting the tree after some marking action has been performed; t a takes values in M p , d ;
(c) 
t b the random variable denoting the tree after the movement of the surfer and before the exploration of the tree; t b takes values in M p , d 1 .
In the evolution of the controlled process, these three random variables will depend on the time step n. When necessary, we use the notation t ( 0 ) , t a ( 0 ) , t b ( 0 ) , t ( 1 ) , t a ( 1 ) , t b ( 1 ) , , t ( n ) , t a ( n ) , t b ( n ) , to denote this (random) succession of trees.
Definition 4
(Discovery).Let D : M p , d 1 P ( M p , d + ) denote the mapping such that D ( t ) is the set of trees that can be discovered from tree t M p , d 1 . Elements of D ( t ) are in M p , d + and are obtained from the tree t by updating its leaves according to the following rule: for a leaf l = ( μ ) (i.e., depth-0 tree) in t, update it as:
( μ ) ( μ , l new ) w h e r e   l new { ( 0 ) , ( 0 , 0 ) , , ( 0 ) p } .
Definition 5 (Successors after discovery).
Let SD : M p , d + P ( M p , d + ) be defined by:
SD ( ( μ , s ) ) = t s s D ( t s )
where ⊔ refers to the disjoint union of sets.
The set SD ( t ) contains all trees that are the possible results from a combination of surfer movement and the discovery of new leaves.
For a tree t = ( μ , s ) M p , d + , let a ( t ) M p , d denote the tree after marking according to action a A t . Then, s ( a ( t ) ) is the set of subtrees of t after marking action a. The transition probability of moving from a tree t to a tree t under action a is:
P ( t , a , t ) = 1 | s ( t ) | | D ( t b ) | if   t D ( t b )   and   t b s ( a ( t ) ) 0 otherwise .

3.4. Immediate Cost

The immediate cost of moving to tree t by choosing action a while in tree t is:
c ( t , a , t ) = 0 if   μ ( t ) = 1 1 if   μ ( t ) = 0 .
Accordingly, the expected cost incurred when applying action a to tree t is:
c ( t , a ) = t SD ( t ) P ( t , a , t ) c ( t , a , t )
but a simpler expression is available by substituting the explicit values for the probability term and the cost term, as stated in the next lemma.
Lemma 2.
The expected cost can be written as:
c ( t , a ) = 1 1 | s ( t ) | t s ( a ( t ) ) μ ( t ) .
Given a budget or a specific family of policies, there are states in S that will never feature in the MDP. Thus, we shall focus only on the “usable states”.
Definition 6 (Usable states).
The states in S that are attained through transitions given a specific value of budget k or a family of policies are called usable states. Denote this set of states as U .
For example, budget dependent states for k = 1 , d = 2 , and p 2 will not include states where both nodes at depth 1 (if two exist) are marked. We will come back to this in Section 6.

3.5. Policies and Criterion

We choose as evaluation criterion the expected average cost over an infinite horizon. The class of policies in which we optimize is, in the terminology of [15] (Section 2.1.4), that of History-dependent, Randomized strategies (HR). However, the classical theory allows focusing on stationary strategies only. Some of our definitions are valid in the general class HR. In this context, γ n ( t ) denotes the action prescribed by the policy at time step n for tree t.
Definition 7 (Sensible policies, greedy policies).
A policy γ H R is called:
Sensible 
if, for every state t and every time instant n, γ n ( t ) marks k unmarked nodes, or the number of unmarked nodes in t;
Greedy 
if it is sensible and if for every t, γ n ( t ) marks: either k of the unmarked sons of t if these are more than k or else all unmarked sons of t plus possibly nodes at a depth of at least 2.
Sensible policies do not waste the marking budget on marked nodes, unless they have no other possibility. Among them, greedy policies mark sons as priority.
The following observation will be useful. Its proof is immediate by unrolling the marking/surfing/discovering cycle.
Lemma 3.
Consider a stationary policy γ such that for any t, γ ( t ) marks nodes only up to some depth d m . Let M γ be the Markov chain generated by this policy. Then, the usable states U , and in particular the recurrent classes of M γ , contain only trees with nodes marked up to depth d m 1 .
Remark 1.
Given the rules of surfing (uniform choice) and of discovery (independence of different subtrees), it seems possible to further reduce the size of the state space by exploiting symmetries. For instance, in Figure 2, the fourth and fifth trees will have exactly the same properties. We chose not to exploit these symmetries, because this would lead to an extra complexity in the formalism. Furthermore, it would render the enumeration of state spaces more complex and, as a consequence, complicate the description of the process on tree shapes; see the following section.

4. The Markov Chain of Tree Shapes

In this section, we temporarily forget the control part of the MDP and focus on the process of trees generated by the surfing/discovering mechanism. It turns out to contain two Markov chains, which we identify and analyze.

4.1. Definition and Basic Properties

An important feature of the MDP constructed in Section 3 is that the shape of the successive trees does not depend on the marking strategy. In order to formalize this, we first define the shape of trees.
Definition 8 (Shape of trees).
Consider the mapping σ p , d : M p , d T p , d , defined for all p and d recursively with:
σ p , 0 ( t ) = 0 , σ p , d ( μ , ( s 1 , , s m ) ) = ( 0 , ( σ p , d 1 ( s 1 ) , , σ p , d 1 ( s m ) ) ) , i f d 1 .
The tree σ p , d ( t ) is called the shape of tree t.
We now state the aforementioned property of the shape of trees in the MDP. Observe that, by the definition of the succession of trees t, t a , t b , we have σ p , d ( t ( n ) ) = σ p , d ( t a ( n ) ) for all n.
Proposition 1.
In the MDP defined in Section 3:
(i) 
The distribution of the sequence { σ p , d ( t ( 0 ) ) , σ p , d 1 ( t b ( 0 ) ) , σ p , d ( t ( 1 ) ) , } does not depend on the strategy γ H R ;
(ii) 
The processes { σ p , d ( t ( n ) ) ; n N } , and { σ p , d 1 ( t b ( n ) ) ; n N } are homogeneous Markov chains on the state spaces T p , d and T p , d 1 , respectively.
Proof. 
The proof of ( i ) follows from the fact that transition probabilities in (9) do not depend on the action a. Then, the Markov nature of embedded sequences σ p , d ( t ( n ) ) = σ p , d ( t a ( n ) ) and σ p , d 1 ( t b ( n ) ) is clear since random moves of the surfer and discoveries depend only on the shape of the current tree. □
We proceed with the identification of the stationary distributions of the Markov chains featured in Proposition 1. We first introduce the family of candidate distributions and state their basic properties. We then prove the result about Markov chains.
Definition 9.
Let π p , d : T p , d [ 0 , 1 ] be the sequence of functions defined recursively by:
π p , 0 ( t ) = 1
π p , d ( 0 , ( s 1 , , s m ) ) = 1 p k = 1 m π p , d 1 ( s k ) d 1 .
Lemma 4.
The functions π p , d introduced in Definition 9 have the following properties:
(i) 
For each fixed d, π p , d is a probability distribution on T p , d ;
(ii) 
The probabilities π p , d can be expressed as:
π p , d ( t ) = 1 p inode ( t ) .
The interpretation of Definition 9 and Lemma 4 ( i i ) is that the probability of a tree t is the probability that this tree is generated by a Galton–Watson process with branching probabilities uniform in { 1 , , p } , stopped at generation d.
Proof. 
The proof of ( i ) proceeds by recurrence. For d = 0 , the property is trivial. Assume then the property proven up to d. Then, selecting the number of sons of a tree and using (12), we have:
t T p , d + 1 π p , d + 1 ( t ) = m = 1 p s 1 , , s m T p , d π p , d + 1 ( 0 , ( s 1 , , s m ) ) = m = 1 p s 1 , , s m T p , d 1 p j = 1 m π p , d ( s j ) = m = 1 p 1 p s T p , d π p , d ( s ) m = m = 1 p 1 p = 1 .
We continue with the proof of ( i i ) . Recursively, inode ( t ) = 1 + t s ( t ) inode ( t ) . The result then follows from the definition (12). □
The following properties of the distribution π p , d will be useful. The proofs of the first two of them are straightforward and omitted.
Lemma 5.
Let t be a random tree in T p , d distributed according to π p , d . Then, | s ( t ) | is uniformly distributed in { 1 , , p } .
Lemma 6.
Let t be a random tree in T p , d distributed according to π p , d . Then, conditioned on the fact that | s ( t ) | = m , the subtrees s 1 , , s m are independent and uniformly distributed according to π p , d 1 .
Proposition 2.
Consider the Markov chains M = { σ p , d ( t ( n ) ) ;   n N } and M b = { σ p , d 1 ( t b ( n ) ) ; n N } :
( i )
Both chains are ergodic;
( i i )
The stationary distribution of M is π p , d ;
( i i i )
The stationary distribution of M b is π p , d 1 .
Proof. 
The property ( i ) is proven if we can show that both chains are irreducible and aperiodic. Irreducibility follows from the fact that there is a sequence of transitions with a nonzero probability leading to, say, the tree that is a chain (all its internal nodes have only one son, let us name it c p , d ): if the discovery phase adds just one leaf to every leaf of t b (this happens with positive probability), after d steps, the tree is c p , d , whatever the random surfing moves. The tree t b itself is c p , d 1 . Aperiodicity also follows from this construction since the transition from c p , d to c p , d has a nonzero probability.
In order to prove ( i i ) , we check that the distribution π p , d satisfies the equation π = π P . Since M is ergodic, this will be the unique solution. We first identify the set of trees that have a positive probability of transition to a given tree t T p , d . To that end, we have to reverse the process of the transformation of one tree into another. Reversing the discovery phase, we are led to define top ( t ) T p , d 1 as the tree deduced from t by removing the leaves. Then, reversing the surfer movement, we conclude that t can be transformed into t if and only if t has top ( t ) as one of its subtrees. Let Prec ( t ) T p , d be this set. For any t Prec ( t ) , we have:
P ( t , t ) = card { τ s ( t ) | τ = top ( t ) } | s ( t ) | 1 p leaves ( top ( t ) ) .
Accordingly, it is convenient to partition the set Prec ( t ) into “blocks” of states as follows:
Prec ( t ) = m = 1 p n = 1 m P ( m , n ) P ( m , n ) : = { t T p , d , | s ( t ) | = m , s ( t ) contains top ( t ) exactly n times } .
Trees that have the same number of sons and the same number of occurrences of top ( t ) among their sons are grouped together. By construction, we have:
P ( t , t ) = n m 1 p leaves ( top ( t ) ) t P ( m , n ) .
This transition probability is therefore constant in the block P ( m , n ) . Then, we can write:
t T p , d π p , d ( t ) P ( t , t ) = t Prec ( t ) π p , d ( t ) P ( t , t ) = m = 1 p n = 1 m t P ( m , n ) π p , d ( t ) n m 1 p leaves ( top ( t ) ) = 1 p leaves ( top ( t ) ) m = 1 p n = 1 m n m t P ( m , n ) π p , d ( t ) .
We evaluate the inner sum, which is the total probability of the block P ( m , n ) under the distribution π p , d . According to Lemma 6, the distribution of subtrees of t P ( m , n ) is that of m independent trees in T p , d 1 . Therefore, the probability, conditioned on m, that exactly n subtrees of t are top ( t ) is the following binomial distribution (resulting from picking the n locations for trees top ( t ) among the m possibilities):
m n π p , d 1 ( top ( t ) ) n ( 1 π p , d 1 ( top ( t ) ) ) m n .
We conclude that:
t P ( m , n ) π p , d ( t ) = 1 p m n π p , d 1 ( top ( t ) ) n ( 1 π p , d 1 ( top ( t ) ) ) m n .
Using this result, we can evaluate the product of the distribution π p , d and the matrix P through the following computation:
t T p , d π p , d ( t ) P ( t , t )
= 1 p leaves ( top ( t ) ) m = 1 p n = 1 m n m 1 p m n π p , d 1 ( top ( t ) ) n ( 1 π p , d 1 ( top ( t ) ) ) m n (16a)
= 1 p 1 + leaves ( top ( t ) ) m = 1 p n = 1 m m 1 n 1 ( π p , d 1 ( top ( t ) ) ) n 1 π p , d 1 ( top ( t ) ) m n
= π p , d 1 ( top ( t ) ) p 1 + leaves ( top ( t ) ) m = 1 p n = 0 m 1 m 1 n ( π p , d 1 ( top ( t ) ) ) n 1 π p , d 1 ( top ( t ) ) m 1 n
= π p , d 1 ( top ( t ) ) p 1 + leaves ( top ( t ) ) m = 1 p 1 (16b)
= π p , d 1 ( top ( t ) ) p 1 + leaves ( top ( t ) ) · p = 1 p inode ( top ( t ) ) + leaves ( top ( t ) ) = 1 p inode ( t ) (16c)
= π p , d ( t ) .
We used (13) and (14) to obtain (16a). The binomial expansion theorem was used in (16b). Finally, in (16c), we note that inode ( t ) is the sum of inode ( top ( t ) ) and leaves ( top ( t ) ) , which gives us the desired result, π p , d ( t ) .
Finally, we prove ( i i i ) . We know that σ ( t b ) results from σ ( t ) through the random choice of a son of t. Invoking again Lemma 6, we have: conditioned on the event { | s ( t ) | = m } , each subtree is τ with probability π p , d 1 ( τ ) , and the probability that a uniform random choice picks τ has probability π p , d 1 as well. Since this does not depend on m, the result is true also when removing the conditioning. □

4.2. Application to Greedy Policies

As an application of Lemma 5, we obtain an upper bound on the optimal cost that can be realized by any policy. This was based on the following result.
Lemma 7.
Consider a random variable t T p , d distributed as π p , d . Define the random variable:
C = [ | s ( t ) | k ] + | s ( t ) | .
Then, C = 0 with probability 1 (in particular, E C = 0 ) if k p and:
E C = 1 p H p k , k p ,
where:
H p k : = m = k + 1 p m k m = p k k ( H p H k ) .
Proof. 
According to Lemma 5, | s ( t ) | is uniformly distributed. Then:
E C = m = 1 p 1 p [ m k ] + m = 1 p m = k + 1 p m k m = 1 p H p k = 1 p m = k + 1 p 1 k m = 1 p p k k m = k + 1 p 1 m = 1 p p k k ( H p H k ) .
We now state the bound announced.
Proposition 3.
The optimal expected cost g * of the MDP described in Section 3 satisfies:
g * 1 k p 1 + H p H k .
The cost of any greedy policy satisfies the same bound.
Proof. 
Consider the policy that marks k sons of the current tree or all the sons if k is larger than this. This is a greedy policy, in the sense of Definition 7. The average cost of this policy is precisely given by E C as in Lemma 7, which is equal to the right-hand side in (19). Any greedy policy will mark at least the same nodes: it will necessarily have a smaller average cost. The optimal cost of the MDP is necessarily smaller than the cost of this specific policy. □

4.3. Metrics of Tree Shapes

We applied the results of Section 4.1 to evaluate several simple metrics in tree shapes generated by the MDP. This may be useful in particular for estimating the quantity of memory needed in a simulation of this process.

4.3.1. Average Number of Nodes

Let N p , d = E ( | t | ) be the average size of a tree t distributed according to π p , d . According to Lemma 6, we can write, conditional on the event { | s ( t ) | = m } : | t | = 1 + | s 1 | + + | s m | so that:
E ( | t | | | s ( t ) | = m ) = 1 + j = 1 m E ( | s j | ) = 1 + m N p , d 1 N p , d = E ( | t | ) = 1 + E ( | s ( t ) | ) N p , d 1 = 1 + p + 1 2 N p , d 1 .
Since the initial condition is N p , 0 = 1 , the recurrence in (20) has the solution:
N p , d = 2 p 1 p + 1 2 d + 1 1 , p 2 ,
with N 1 , d = d + 1 . We have proven the following result.
Lemma 8.
The average number of nodes in a tree t T p , d distributed according to π p , d is given by (21).

4.3.2. Average Number of Leaves

Let L p , d be the average number of leaves, in trees t distributed according to π p , d . The reasoning of Section 4.3.1 can be reproduced: according to Lemma 6, we can write, conditional on the event { | s ( t ) | = m } : | t | = | s 1 | + + | s m | . If follows that:
L p , d = E ( | s ( t ) | ) L p , d 1 = p + 1 2 L p , d 1 .
Since L p , 0 = 1 , we have the following result.
Lemma 9.
The average number of leaves in a tree t T p , d distributed according to π p , d is:
L p , d = p + 1 2 d .
Note that the maximal number of leaves of a tree in T p , d is p 2 .

4.3.3. Average Number of Nodes Created

Let C p , d be the average number of nodes created at some time step in the Markov chain { σ p , d ( t ( n ) ) ; n N } in its stationary regime. It is equal to the expected number of nodes created from some tree t T p , d distributed according to π p , d . This number itself is equal to the expected number of leaves in t, since t results from the addition of leaves to some t , also distributed according to π p , d . We therefore have, with Lemma 9:
C p , d = p + 1 2 d .

4.3.4. Average Number of Nodes Deleted

Let D p , d be the average number of nodes deleted at some time step in the Markov chain { σ p , d ( t ( n ) ) ; n N } in its stationary regime. By stationarity, it is expected that D p , d = C p , d . We verify this with a direct computation.
Consider a tree t T p , d . Conditioned on the events { | s ( t ) | = m } and the surfer moving to the j th son , the number of nodes deleted is: 1 + j | s | (the root and the other subtrees). Using Lemma 5 and the fact that each s is distributed according to π p , d 1 (Lemma 6), we have then:
D p , d = m = 1 p 1 p j = 1 m 1 m 1 + ( m 1 ) N p , d 1 = 1 p m = 1 p 1 + ( m 1 ) N p , d 1 = 1 + p ( p 1 ) 2 p N p , d 1 = 1 + p 1 2 N p , d 1 = p + 1 2 d
where the last equality results from (21). This is equal to the average number of nodes created C p , d , as expected.

5. Trees of Depth d = 1 with an Arbitrary Marking Budget

In this section, we consider trees of depth 1, and we prove the following result.
Theorem 1.
When d = 1 , any greedy policy is optimal.
It is quite clear intuitively that, indeed, no reasonable alternative exists. The exercise here is to check that the theory does provide a way to prove the result formally. In the process, we identify arguments that will be useful for the proofs of stronger results.
For the purpose of the forthcoming proof, we rename the elements of M p , 1 + as: M p , 1 + = { t μ , j : μ { 0 , 1 } , 1 j p } . In the notation of Section 2, we have t μ , j = ( μ ; ( 0 , , 0 ) j times ) . Furthermore, observe that trees t b belong to M p , 0 = { 0 , 1 } : these are trees reduced to a root with a mark.
Proof. 
We shall prove the result using Theorem A1. Define the constant g and the function f : T p , 1 R as:
g = H p k p
f ( t μ , j ) = ( j k ) + j , μ { 0 , 1 } .
The symbol H p k was defined in (18). We shall check that this g , f satisfies the optimality Equation (A1). For every state s = t μ , j and every action a, we write the quantity to be minimized in the right-hand side of this equation:
Q ( s , a ) : = c ( s , a ) + s M p , 1 + P ( s , a , s ) f ( s ) = c ( t μ , j , a ) + s M p , 1 + P ( t μ , j , a , s ) f ( s ) = c ( t μ , j , a ) + μ { 0 , 1 } P ¯ ( t μ , j , a , μ ) j = 1 p 1 p f ( t μ , j ) .
We obtained (24) by conditioning the transition s s on the value of the tree t b . The new notation P ¯ ( t , a , t b ) stands for the probability of moving from t to t b when action a is applied. Given the definition of f ( · ) in (23), we have further:
Q ( s , a ) = c ( t μ , j , a ) + μ { 0 , 1 } P ¯ ( t μ , j , a , μ ) j = 1 p 1 p ( j k ) + j = c ( t μ , j , a ) + μ { 0 , 1 } P ¯ ( t μ , j , a , μ ) H p k p = c ( t μ , j , a ) + g .
The actions a can be grouped according to the number of sons they mark in the tree t a : this number ranges from 0 to min { j , k } . When t a has sons marked, this determines the cost as:
c ( t μ , j , a ) = j j .
Finally, the minimization with respect to a amounts to the following minimization with respect to :
min a c ( s , a ) + s M p , 1 + P ( s , a , s ) f ( s ) = min 0 j k j j + g = ( j k ) + j + g = f ( s ) + g .
The constant g and the function f therefore solve Equation (A1). This function is bounded since the state space is finite. Therefore, there exists an optimal policy γ * with cost g. Clearly, this policy consists of marking up to k sons of any tree: a greedy policy in the sense of Definition 7. □
From the proof of Theorem 1 and also from Lemma 7, we have the corollary:
Corollary 1.
The average value of any tree of depth 1 is H p k / p .
Remark 2.
The fact that, in the present case, f ( s ) = min a c ( s , a ) is a consequence of the fact that the cost of the future tree s resulting from the transition is actually independent of the action a.
Remark 3.
It was proven in [16] that the finite-horizon, total-cost optimal-value function is given by:
W N * ( t μ , j ) = ( j k ) + j + ( N 1 ) p H p k ,
and is realized by any greedy policy. We obtained from this result the value of g = lim N + W N * ( t μ , j ) / N and the form of function f that must satisfy f ( s ) f ( s ) = lim N + ( W N * ( s ) W N * ( s ) ) .

6. Trees of Depth d = 2 with Marking Budget k = 1

In this section, we consider trees of depth 2, and we prove the following result.
Proposition 4.
When k = 1 and d = 2 , any greedy policy is optimal.
We begin with some notation and preliminary results. Then, we provide the proof.

6.1. Preliminaries

As a preliminary, observe that as a consequence of the marking/moving/discovery cycle, the subtrees of T 1 , p that appear have at most one leaf marked. Indeed, at the beginning of the cycle, trees t M 2 , p + have all leaves unmarked. The marking with budget k = 1 marks at most one of these leaves. Then, the surfer moves to one subtree t b M 1 , p , which inherits this property. The discovery phase merely adds unmarked leaves at depth 2. With this observation, we can restrict our attention to the usable set U of trees with at most one leaf marked, since only those can appear recurrently when some stationary policy is applied.
A second preparation is to calculate the average cost under some greedy policy γ , which therefore marks one node at depth one in any tree t. The choice of this node does not matter. According to Lemma 3, the Markov chain M γ generated by this policy has recurrent states with marks only at depth 0, that is at the root. Therefore, the cost (10) is always given by c ( t , γ ( t ) ) = 1 1 / | s ( t ) | . It is then of the form assumed in Lemma 7, and the application of (17) yields the expected cost for policy γ :
J γ = 1 p p 1 ( H p H 1 ) = 1 H p p .

6.2. Notation and Terminology

When d = 2 , the trees of interest are simpler than in the general case, and it is convenient to devise an appropriate notation. For trees in M p , 2 + , all subtrees have depth one and unmarked leaves. We shall adopt the simplified notation for such trees: ( μ ; m ) denotes a depth one tree with root marked with μ and m unmarked leaves. A typical tree of M p , 2 + is then denoted by t = ( μ ; ( μ 1 , j 1 ) , , ( μ m , j m ) ) for some m [ 1 . . p ] .
After marking, the subtrees of depth one will have at most one leaf marked: then ( μ ; m + ) will denote this tree with one marked leaf. Which leaf exactly is marked does not make a difference in the following reasoning.
In the analysis, the number of unmarked sons of a tree is a key criterion. Accordingly, we introduce the following typology for t = ( μ ; ( μ 1 , j 1 ) , , ( μ m , j m ) ) M p , 2 + :
Type 1 : r = 1 m μ r m 1 , Type 2 : r = 1 m μ r = m .

6.3. Proof

The proof uses the optimality equations and Theorem A1, as in Section 5. The “g” value needed for this was computed as J γ in (27). The next step is to evaluate the “f” function in the optimality Equation (A1) in Theorem A1. It is sufficient to provide a value to states that are usable in the sense of Definition 6. For other states, the value of f is defined by (A1), since the right-hand side only contains values of reachable states.
The function f that is proposed is the following:
f ( μ ; ( μ 1 , j 1 ) , , ( μ m , j m ) ) = 1 m 1 + r = 1 m μ r + r = 1 m 1 j r r = 1 m μ r m 1 ( type 1 ) 1 1 m 2 r = 1 m 1 j r + S L 1 ( t ) H p p p 1 r = 1 m μ r = m ( type 2 ) .
In this last line, S L 1 ( t ) = | { r | j r = 1 } | is the number of subtrees of t that have exactly one leaf.
Proof of Proposition 4. 
We apply Theorem A1 by checking that the function f, in (28), and the constant g = 1 H p / p satisfy the optimality equations. To that end, we first evaluate the expected value of trees t b in M p , 1 . Denote by P dis : M p , 1 M p , 2 + the transition probability from a tree t b to a “discovered” tree t . We can write:
f ˜ ( t b ) = t M p , 2 + P dis ( t b , t ) f ( t ) .
We have, for any μ { 0 , 1 } :
f ˜ ( μ ; j ) = 1 = 1 p j = 1 p 1 p j f ( μ ; ( 0 , 1 ) , , ( 0 , j ) ) ) = 1 j 1 = 1 p j = 1 p 1 p j 1 + r = 1 j 1 r = 1 j 1 + r = 1 j 1 = 1 p j = 1 p 1 p j 1 r = 1 j 1 + r = 1 j p j 1 p j r = 1 p 1 r = 1 j 1 + r = 1 j H p p = 1 j + H p p
[ j 2 ] f ˜ ( μ ; j + ) = 1 = 1 p j = 1 p 1 p j f ( μ ; ( 0 , 1 ) , , ( 1 , q ) , , ( 0 , j ) ) ) = 1 j 1 = 1 p j = 1 p 1 p j 2 + r = 1 j 1 r = 2 j + H p p
f ˜ ( μ ; 1 + ) = = 1 p 1 p f ( μ ; ( 1 , ) ) = 1 p = 1 p 1 2 1 = 1 H p p p 1 = 1 p p + 2 H p + H p p p 1 = 1 + 2 H p p + H p p p ( p 1 ) .
From these formulas, the following identities are obtained: for any μ and μ ,
f ˜ ( μ ; j ) f ˜ ( μ ; j + ) = 1 j j 2 f ˜ ( μ ; 1 ) f ˜ ( μ ; 1 + ) = H p p + H p p p ( p 1 ) = H p 1 p 1 .
The following result then immediately follows:
Lemma 10.
For all j { 1 , , p } , μ , μ { 0 , 1 } and all p 2 , 0 f ˜ ( μ ; j ) f ˜ ( μ ; j + ) 1 / 2 .
We now proceed with checking that f and g solve the optimality equations. We start with Type 1 trees. Let t = ( μ ; ( μ 1 , j 1 ) , , ( μ m , j m ) ) with r = 1 m μ m < m . The alternative actions are: ( a ) mark the root or any son already marked; ( b k ) mark an unmarked son k; ( c k ) mark a leaf of subtree ( μ k , j k ) . For actions ( b k ) , we denote by μ i the marks of the sons after marking: μ i = μ i for ( i k and μ k = μ k + 1 . Clearly, r = 1 m μ r = 1 + r = 1 m μ r . For actions ( c k ) , which leaf is marked does not matter, so we ignore this information.
The right-hand side of the optimality equation in the cases ( a ) , ( b k ) , and ( c k ) are respectively:
Q ( t , a ) = m r = 1 m μ r m + r = 1 m 1 m f ˜ ( μ r ; j r ) Q ( t , b k ) = m r = 1 m μ r 1 m + r = 1 m 1 m f ˜ ( μ r ; j r ) = m r = 1 m μ r 1 m r = 1 m 1 m 1 j r + H p p = 1 H p p r = 1 m μ r + 1 m 1 m r = 1 m 1 j r = g + f ( t ) Q ( t , c k ) = m r = 1 m μ r m + 1 m r = 1 | r k m f ˜ ( μ r ; j r ) + 1 m f ˜ ( μ k ; j k + ) .
Then:
Q ( t , c k ) Q ( t , a ) = 1 m f ˜ ( μ k ; j k + ) f ˜ ( μ k ; j k ) Q ( t , b ) Q ( t , c k ) = 1 m + 1 m f ˜ ( μ k ; j k ) f ˜ ( μ k ; j k + ) .
Both differences are negative according to Lemma 10. This implies that action ( b k ) dominates all actions ( c k ) , which in turn dominate action ( a ) . It therefore realizes the minimum in the right-hand side of the optimality equation. With (32), the right-hand side and the left-hand side coincide.
Next, we consider Type 2 trees. According to the preliminary remark, we can focus our attention on trees with at most one son marked: if this tree is of Type 2 (all sons marked), then it has only one son. Let then t = ( μ ; ( 1 , j ) ) , j [ 1 . . p ] be such a tree. The alternative actions are: ( a ) mark the root or the son; ( b ) mark leaf of the subtree ( μ k , j k ) . The right-hand side of the optimality equation is respectively: Q ( t , a ) = f ˜ ( 1 ; j ) and Q ( t , b ) = f ˜ ( 1 ; j + ) . From Lemma 10, we know that action ( b ) dominates action ( a ) . Further, from Definitions (28) and (31),
f ( μ ; ( 1 , j ) ) + g Q ( t , a ) = f ( μ ; ( 1 , j ) ) + 1 H p p f ˜ ( 1 ; j + ) = 1 2 + H p p p 1 + 1 H p p + 1 + 2 H p p + H p p p ( p 1 ) = 1 + H p p + H p p p 1 1 p 1 = 1 + H p p H p p p = 0 .
Remark 4.
It was proven in [16] that, for any tree t = ( μ ; ( s 1 , , s m ) ) , in the usable set, namely the set of trees with no sons marked, the finite-horizon total-cost optimal-value function is given by:
W n * ( t ) = n 1 m n 2 p H p 1 m r = 1 m 1 | s ( s r ) | f o r 2 n N ,
and that this cost is realized by any greedy policy. From this result, the average cost of this policy in the infinite horizon is then:
lim N W N * ( t ) N = lim N 1 N N 1 m N 2 p H p 1 m r = 1 m 1 | s ( s r ) | = 1 H p p .
This matches with (27). Furthermore, it is compatible with the form of the function f in (28). The interpretation of f ( t ) f ( t ) is the difference in the total expected cost when starting from trees t or t . According to (33), the tree-dependent cost for trees with unmarked sons would be:
f ( t ) = 1 | s ( t ) | 1 + r = 1 | s ( t ) | 1 | s ( s r ) | .
This is indeed the value in (28) since r μ r = 0 .

7. Trees of Depth d = 2 with Marking Budget k = 2

This section is devoted to the case where the marking budget is k = 2 . In this case, we do not present general results, but we focus on the case of trees with depth 2. For small values of p, we describe the optimal policy, and we conjecture that this policy is optimal for general values of p.
We begin with some additional notation, then we introduce the definitions of the policies of interest. This allows us to formulate Conjecture 1. We then present numerical experiments made with small values of p supporting this conjecture.

7.1. Preliminary

Similarly as in Section 6, we argue that usable states are necessarily such that the marking of sons r = 1 m i r takes values in { 0 , 1 , 2 } . Indeed, the sons of a tree are the leaves of a tree at the previous time step, and at most two leaves can be marked at any step.

7.2. Notation and Terminology

We first recall the representation of depth two trees of M 2 , p + from Section 6.2. Such a tree can be represented as ( μ , ( μ 1 , j 1 ) , , ( μ m , j m ) ) where m [ 1 , p ] , μ , μ r { 0 , 1 } and 1 j r p for all r [ 1 , m ] . μ , μ r are the markings of the root and the sons, m is the number of sons, and j r are the number of leaves of the depth one subtree ( μ r , j r ) . Based on the number of marked sons of the tree, we classify them into two types:
Type 1 : r = 1 m μ r m 2 , Type 2 : r = 1 m μ r m 1 .
Type 2 trees are further classified into three subtypes (remember that r μ r cannot exceed 2):
Type 2 a : r = 1 m μ r = m 1 , Type 2 b : m = 2 , μ 1 = μ 2 = 1 , Type 2 c : m = 1 , μ 1 = 1 .
We also introduce some shorthand notation for the different possible actions. Let a ( d 1 , d 1 ) represent the action of marking two sons ( d 1 as in “depth 1”), a ( d 1 , l j c ) represent the action of marking a son and a leaf of the tree ( μ c , j c ) ( l j as in “leaf j”), and a ( l j c 1 , l j c 2 ) represent the action of marking a leaf in each of the subtrees ( μ c 1 , j c 1 ) and ( μ c 2 , j c 2 ) . If c 1 = c 2 , two leaves are marked in this subtree.

7.3. Policies

All policies of interest in our study are greedy in the sense of Definition 7. These policies do not specify what happens when some marking budget is left after marking all unmarked sons of a tree. We therefore specify the following variants of the greedy policy by their precise behavior in this situation. We begin with four simple rules; three of them rely on an order defined on the subtrees. The terms “first” and “second” used in the specification are relative to this order:
  • Greedy depth 1: Ensure that only the sons of the tree are marked;
  • Greedy smallest: Ensure that the sons of the tree are marked. If budget remains, then mark the first leaf of the smallest subtree. If budget still remains, mark the second leaf of the smallest subtree, if any. Otherwise, mark the first leaf of the second smallest subtree;
  • Greedy largest: Ensure that the sons of the tree are marked. If budget remains, then mark the first leaf of the largest subtree. If budget still remains, mark the second leaf of the largest subtree, if any. Otherwise, mark the first leaf of the second largest subtree;
  • Greedy leftmost: Ensure that the sons of the tree are marked. If budget remains, then mark the first leaf of the leftmost subtree. If budget still remains, mark the second leaf of the leftmost subtree, if any. Otherwise, mark the first leaf of the second leftmost subtree.
The cost of policy “greedy depth 1” is known by Lemma 7:
J Greedy Depth 1 = H p 2 p = 1 + 1 2 H p p .
Finally, we introduce the “greedy finite optimal” policy, a name that we use as a shorthand for “the policy that seems to emerge as optimal with the finite horizon criterion”. Its behavior is specified in Table 2. It is explained in Appendix C how the features of this policy are extrapolated from the results obtained with the finite-horizon version of the MDP.
The behavior of the “greedy finite optimal” policy is obvious on Type 1 (more than two unmarked sons) and Type 2c (one son that is marked) trees. On Type 2a (one unmarked son) and Type 2b (two sons, both marked), it introduces a threshold of 3 or 4 on the size of the subtrees. When the tree is of Type 2a, one mark is left for marking subtrees. If all of them have a size less than 2, the largest one is marked. On the other hand, if some of them has a size larger than 3, the smallest of these is marked. When the tree is of Type 2b, two marks are left for marking subtrees. If both of them have a size larger than 4 ( j 2 j 1 > 3 ), the smallest subtree is marked. In the other case, both subtrees are marked.
A rationale for such a rule is as follows. When a tree has less than two unmarked sons, it can be made costless with the two marks of the budget. Therefore, when considering trees of Type 2a, a subtree that has less than two leaves has less priority than a subtree with more than three leaves, which is more “vulnerable”. Among the vulnerable trees, it is better to mark the smallest ones, in order to reduce future expected costs. Trees of Type 2b have, so to speak, one round in advance since all sons are already marked. Subtrees of a size less than 3 can be “protected” by devoting one mark to them: if the surfer moves to them, the budget of the next round will be used to complete the protection. If both trees are too large to be fully protected, the budget is devoted again to the smallest one.
We can now state the conjecture that is the focus of this section.
Conjecture 1.
When k = 2 and d = 2 , the “greedy finite optimal” policy is optimal.

7.4. Numerical Experiments

We provide support to Conjecture 1 with results for small p. We implemented the policy improvement algorithm [15] (Chapter 8.6), starting with a particular greedy policy. Prior to the implementation of the algorithm, we evaluated the average cost of the four variants of the greedy policy introduced in Section 7.3, for p = 3 , 4 , 5 . Of course, the greedy policies realize a zero cost for p 2 .
The average costs of the five greedy policies for small values of p are summarized in Table 3. The row concerning the “greedy depth 1” policy was evaluated using the exact formula (34), which gave respectively 1 / 9 , 5 / 24 , and 43 / 150 for p = 3 , 4 , and 5. Several observations can be made from the data of this table. First, there is a substantial gain of marking subtrees after having marked all the sons: there is a larger gap between the performance of greedy depth one and the group of the other ones, than inside this group. Second, the best performance among the four simple policies introduced in Section 7.3 was achieved by marking the largest subtree, for all values of p tested. Picking the smallest subtree had the worst performance compared to picking the leftmost/largest subtree. Finally, choosing arbitrarily the leftmost (or, for that matter, the rightmost or a random) subtree resulted in a performance between these extremes.
We used the greedy largest policy as a starting candidate policy for the policy iteration algorithm for p = 3 , 4 , 5 , since it gave the best performance among the simple policies. In each case, the algorithm converged in a few iterations, and the resulting policy was the greedy finite optimal policy. The performance of the policy iteration algorithm is summarized in Table 4. The execution time figures correspond to an implementation in Python 3.9 running on a 1.4 GHz quad-core processor with 8 GB memory and a 1536 MB graphic card.
We could have also started the algorithm with the greedy finite optimal policy and checked that it solves the optimality equations. Selecting another policy was also a way of checking that our implementation of policy iteration works properly, as well as measuring “how far” from the optimum the greedy largest policy is.
Observe in Table 3 that the relative performance of the greedy finite optimal policy was more pronounced for p = 5 as compared to p = 4 . The former MDP had a larger number of Type 2 trees. The greedy largest policy prescribed a suboptimal marking scheme for such trees, which explains the greater cost reduction by switching to the greedy finite optimal policy for p = 5 . The greedy finite optimal policy for p = 3 coincided with the greedy largest policy, and hence, the costs were identical.

8. Discussion

We proposed a stochastic dynamic decision model for prefetching problems, which is simple in the sense that it has only three integer parameters, and yet can help conceive of optimal strategies in practical situations. The simplicity of the model lies in several assumptions that we discuss now.
We first observe that the modeling we proposed does not look practical for large values of parameters d and p. Indeed, Table 1 clearly shows that the state space sizes that could be handled numerically corresponded to small values of these parameters. On the other hand, the formal results obtained so far suggest that such a numerical solution would be needed in practice. We argue, consistent with our introduction, that large values of d are not desirable in practice: it may be better not to know the graph at a larger distance, as long as the complexity of the decision grows exponentially with the amount of information. A value d = 2 may be a good compromise between excessive shortsightedness and an excess of information. Concerning the parameters k and p, practical situations should involve cases where these values are not far from each other. Clearly, if k p , the problem is easy, whereas if k p , all policies will be bad because the controller is overwhelmed by the number of nodes to control. The modeling we propose is relevant if the network is a bottleneck of the system: in other words, if k is not very large. Therefore, p should not be very large either.
The next feature departing from practical cases is about the discovery process. In our model, we assumed a uniform distribution for the new generation of nodes. Practical graphs are known to have different node degree distributions. Here, since we identified this mechanism with a Galton–Watson branching process, it seems possible to use other distributions while conserving the possibility of characterizing the distribution of trees as we did in Section 4. Therefore, evaluating the performance of simple greedy policies and obtaining bounds might be possible with this generalization. Furthermore, the results we obtained with budget k = 1 or depth d = 1 were probably insensitive to the distribution of the number of sons.
The assumption we made about the movements of the “surfer” in the graph of documents may also be questioned. We assumed a uniform choice between neighboring documents. In addition to simplicity, we argue that this represents the most difficult situation for the controller, since the amount of information available to it is minimal. In the case where nonuniform movement probabilities are known, they can easily be integrated in the MDP, just as in the models reviewed in the Introduction. It is interesting to note that in such a case, models can be imagined where the optimal policy is not of the greedy type: consider just a case where the probability of moving to some son happens to be zero (or close to zero). The prefetching budget should then be devoted to marking the other sons and their subtrees. Apart from such obvious situations, it is difficult to imagine cases where an optimal policy would be formally identified, since this policy should weight the movement probabilities with the characteristics of the subtrees.
Going back to the simple model we proposed, the results of Section 7 (the case d = 2 and k = 2 ) and their interpretation suggest a general form for a heuristic policy. The first principle is that sons should be marked as priority. The issue is what to do with the remaining budget. On the one hand, the subtrees that have themselves less than k sons are easy to deal with and can be ignored. Among the remaining subtrees, those with the smallest number of sons should be marked first. If budget remains, the principle can be applied recursively.

9. Conclusions and Perspectives

Among the results of this paper, we proved that simple greedy policies are optimal in several situations, suggesting that optimal policies are always greedy. One first obvious step in future research will be to prove the following result, generalizing Proposition 4:
Conjecture 2.
When k = 1 , any greedy policy is optimal.
Next, our research will focus on the more challenging case k = 2 . The first objective will be to prove Conjecture 1 and then determine how to adapt this policy to larger d. Some numerical investigation appears to be possible there for small values of p, despite the large size of the state space. Another line of research on formal solutions will focus on the analysis of Markov chains defined by simple policies with the purpose of providing bounds tighter than that of Proposition 3.
On the practical side, several issues need further investigation. The principal one is to efficiently identify the model and the optimal control from practical data. We plan to develop algorithms that would leverage the knowledge gained with the exact solution of simple models, with the objective of reducing learning time and learning errors.

Author Contributions

Formal analysis, K.K., A.J.-M., and S.A.; investigation, K.K., A.J.-M., and S.A.; supervision, A.J.-M. and S.A.; writing—original draft, K.K., A.J.-M., and S.A.; writing—review and editing, K.K., A.J.-M., and S.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data is contained within the article.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
CPUCentral Processing Unit
MDPMarkov Decision Process

Appendix A. Notation

Table A1. Main notation used in the paper.
Table A1. Main notation used in the paper.
NotationDefinition
ddepth of a tree
pfanout of a tree; maximal number of sons of any node in a tree
kprefetching budget; maximal number of nodes the controller can mark
S E Q 1 . . p ( A ) sequence of objects in set A, with length between 1 and p
T p , d set of trees of depth d with fanout p
T p , d cardinal of T p , d
M p , d set of trees of depth d with fanout p and marks in { 0 , 1 }
M p , d cardinal of M p , d
M p , d + set of trees of depth d with fanout p and marks in { 0 , 1 } except for leaves
that have mark 0
M p , d + cardinal of M p , d +
μ ( t ) mark (in { 0 , 1 } ) of the root of tree t
s ( t ) list of subtrees of tree t
leaves ( t ) number of leaves in tree t
inode ( t ) number of internal nodes in tree t (all nodes except leaves)
Sstate space of MDP
tstate of MDP; it takes values in M p , d +
A t action space of MDP when state is t
t a tree after the marking action; it takes values in M p , d
t b tree after the surfer’s movement; it takes values in M p , d 1
D ( t ) set of trees that can be discovered from tree t
SD ( t ) set of trees after surfer movement and the discovery of new leaves
when initial tree is t
a ( t ) tree obtained after marking tree t according to action a
P ( t , a , t ) transition probability of moving from a tree t to a tree t under action a
c ( t , a ) expected immediate cost incurred when applying action a to tree t
U set of usable states (states attainable through transitions given
specific budget k)
H R class of policies corresponding to history-dependent randomized strategies
γ a marking policy
γ n ( t ) action prescribed by policy γ at time step n on tree t
γ ( t ) stationary version of γ n ( t )
M γ Markov chain generated by policy γ
σ p , d ( t ) shape of tree t; a tree with the same shape as tree t and all nodes’ marks being 0
M Markov chain of tree shapes with depth d and fanout p
M b Markov chain of tree shapes with depth d 1 and fanout p
π p , d stationary distribution of the Markov chain of tree shapes M
E C upper bound on the optimal expected cost of any greedy policy
H p harmonic number; finite partial sum of the harmonic series
g * optimal expected cost of the MDP
N p , d average number of nodes in trees distributed according to π p , d
L p , d average number of leaves in trees distributed according to π p , d
C p , d average number of nodes created in Markov chain M
D p , d average number of nodes deleted in Markov chain M
g, f ( s ) constant and bounded function in optimality equation
P ¯ ( t , a , t b ) transition probability of moving from a tree t to a tree t b under action a
J γ expected cost for policy γ

Appendix B. MDP Facts

We borrow the below existential theorem from [17] (Theorem 2.1, Chapter V).
Theorem A1.
If there exists a bounded function f ( s ) for every s S and a constant g such that:
g + f ( s ) = max a r ( s , a ) + s S P ( s , a , s ) f ( s )
then there exists a stationary policy γ * such that:
g = max γ ϕ γ ( s ) = ϕ γ * ( s ) .
Theorem A1 guarantees the existence of an optimal policy given that there exists a bounded function f and constant g that satisfies (A1). We use the following theorem from [17] (Theorem 2.2, Chapter V), which proves the existence of such a function and constant.
Theorem A2.
Let { s n } n 0 and { a n } n 0 be the sequence of states and actions of the MDP when a policy γ α is followed. Define:
W γ α ( s ) : = E n = 0 α n r ( s n , a n ) w h e r e   0 < α < 1
and W γ α * ( s ) : = max γ α W γ α ( s ) . For some fixed s 0 S , if there exists a B < such that | W γ α * ( s ) W γ α * ( s 0 ) | < B for all α and s, then there exist a bounded function f and a constant g satisfying (A1).
The uniform boundedness property on W γ α * exists if the expected time to go from any state s to the fixed state s 0 is bounded by a finite value while using the optimal policy γ α * . The reader may refer to Theorem 2.4, Chapter V, from [17] for a proof of the same. A sufficient condition for the bounded expected time is that every stationary policy in the MDP yields a unichain.

Appendix C. Finite Horizon MDP for Trees of Depth d = 2 with Budget k = 2

This section is devoted to the findings from the study of the finite-horizon prefetching MDP, in the cases d = 2 and k = 2 and general p. These results are quoted from the unpublished report [16]. Other results for finite-horizon MDP are quoted in Remarks 3 and 4.
The optimal actions for all tree types for the finite horizons n = 3 , 4 are specified in Table A2. The optimal actions for the n = 3 horizon were computed analytically and confirmed through numerical simulations. For the n = 4 horizon, the optimal actions in Table A2 are the numerical results. The same shorthands for marking actions from Section 7 are used here.
Table A2. Comparison of optimal actions at n = 3 and n = 4 .
Table A2. Comparison of optimal actions at n = 3 and n = 4 .
Tree TypeSpecificationsOptimal Action at n = 3 Optimal Action at n = 4
Type 1 a ( d 1 , d 1 ) a ( d 1 , d 1 )
Type 2a j r 3 for some r a ( d 1 , l j c ) , j c = min r j r 3 a ( d 1 , l j c ) , j c = min r j r 3
j r < 3 for all r a ( d 1 , l j c ) , j c = max r j r a ( d 1 , l j c ) , j c = max r j r
Type 2b j 1 > 3 j 2 > 3 a ( l j 1 , l j 1 ) assuming j 1 j 2 a ( l j 1 , l j 1 ) assuming j 1 j 2
j 1 = 3 j 2 3 a ( l j 1 , l j 1 ) when (A3)
holds true and a ( l j 1 , l j 2 ) otw.
a ( l j 1 , l j 2 ) for p < 13 ,
a ( l j 1 , l j 1 ) when p 13
for large j 2
j 1 = 2 j 2 > 3 a ( l j 1 , l j 1 ) when (A4) holds true,
a ( l j 1 , l j 2 ) when (A5) holds true,
but (A4) does not, and a ( l j 2 , l j 2 ) otw.
a ( l j 1 , l j 2 ) for p < 52 ,
a ( l j 1 , l j 1 ) when p 52
for large j 2
j 2 = 3 a ( l j 2 , l j 2 ) a ( l j 1 , l j 2 )
j 2 = 2 a ( l j 1 , l j 2 ) a ( l j 1 , l j 2 )
j 1 = 1 j 2 4 a ( l j 2 , l j 2 ) when (A6)
holds true and a ( l j 1 , l j 2 ) otw.
a ( l j 1 , l j 2 )
j 2 = 3 a ( l j 2 , l j 2 ) for p < 5 , a ( l j 1 , l j 2 ) for p 5 a ( l j 1 , l j 2 )
j 2 < 3 a ( l j 1 , l j 2 ) a ( l j 1 , l j 2 )
Type 2c a ( l j 1 , l j 1 ) a ( l j 1 , l j 1 )
The policy for Type 2 trees depends on the exact size of the subtrees, unlike the policy for Type 1 trees, where the sizes of the subtrees are irrelevant. There are thresholds that are functions of p, which decide the optimal action for Type 2b trees. The thresholds on j 2 , with the obvious constraint j 2 p , that decide the optimal action for certain specifications of Type 2b trees are below.
j 2 6 p 3 6 p 2 ( H p 3 ) + 6 p ( 5 H p 8 ) + 38 H p 15
j 2 12 p 2 12 p H p + 8 H p 39
j 2 4 p 2 2 p ( 2 H p 5 ) + 10 H p 7
j 2 6 p 6 H p 11 .
We observed a change in optimal actions for Type 2 trees for considerably large fanouts in the numerical experiments for n = 4 . Since the results of n = 4 were obtained numerically, we could identify the critical p values after which there would be a change in the optimal action for some j 2 . The precise threshold on j 2 for which the optimal action changes was not found due to the complex calculations involved. Comparing the optimal actions, we note a simplification when going from n = 3 to n = 4 . The policy for trees of Types 1, 2a, and 2c is the same. For trees of Type 2b, the number of cases where a switch in optimal actions occurs is smaller in the n = 4 than in the n = 3 case. When a switch occurs in both cases, the threshold on the value j 2 is observed to be larger in the case n = 4 . Naturally, we could expect that for a given specification of Type 2b trees, the optimal action would remain the same for most of the trees as the horizon increases. With this line of thought, we may conjecture that the threshold values disappear as n and the optimal actions are the same for all trees of a given type and specification. This is the principle that led to the definition of the greedy finite optimal policy in Table 2.

References

  1. Grigoraş, R.; Charvillat, V.; Douze, M. Optimizing hypervideo navigation using a Markov decision process approach. In Proceedings of the ACM Multimedia, Juan-les-Pins, France, 1–6 December 2002; pp. 39–48. [Google Scholar]
  2. Charvillat, V.; Grigoraş, R. Reinforcement learning for dynamic multimedia adaptation. J. Netw. Comput. Appl. 2007, 30, 1034–1058. [Google Scholar] [CrossRef]
  3. Morad, O.; Jean-Marie, A. Prefetching Control for On-Demand Contents Distribution: A Markov Decision Process Model. In Proceedings of the 22nd IEEE International Symposium on Modeling, Analysis and Simulation of Computer and Telecommunication Systems (MASCOTS 2014), Paris, France, 9–11 September 2014; pp. 421–426. [Google Scholar] [CrossRef]
  4. Morad, O.; Jean-Marie, A. On-Demand Prefetching Heuristic Policies: A Performance Evaluation. In Proceedings of the 29th International Symposium on Computer and Information Sciences (ISCIS 2014), Krakow, Poland, 27–28 October 2014; pp. 317–324. [Google Scholar] [CrossRef]
  5. Pleşca, C.; Charvillat, V.; Ooi, W.T. Multimedia prefetching with optimal Markovian policies. J. Netw. Comput. Appl. 2016, 69, 40–53. [Google Scholar] [CrossRef]
  6. Joseph, D.; Grunwald, D. Prefetching using Markov predictors. IEEE Trans. Comput. 1999, 48, 121–133. [Google Scholar] [CrossRef]
  7. Dongshan, X.; Junyi, S. A new Markov model for Web access prediction. Comput. Sci. Eng. 2002, 4, 34–39. [Google Scholar] [CrossRef]
  8. Jin, X.; Xu, H. An approach to intelligent Web pre-fetching based on hidden Markov model. In Proceedings of the 42nd IEEE International Conference on Decision and Control (IEEE Cat. No.03CH37475), Maui, HI, USA, 9–12 December 2003; Volume 3, pp. 2954–2958. [Google Scholar] [CrossRef]
  9. Gellert, A.; Florea, A. Web Page Prediction Enhanced with Confidence Mechanism. J. Web Eng. 2014, 13, 507–524. [Google Scholar]
  10. Kawazu, H.; Toriumi, F.; Takano, M.; Wada, K.; Fukuda, I. Analytical method of web user behavior using Hidden Markov Model. In Proceedings of the 2016 IEEE International Conference on Big Data (Big Data), Washington, DC, USA, 5–8 December 2016; pp. 2518–2524. [Google Scholar] [CrossRef]
  11. Gupta, S.; Moharir, S. Request patterns and caching for VoD services with recommendation systems. In Proceedings of the 2017 9th International Conference on Communication Systems and Networks (COMSNETS), Bengaluru, India, 4–8 January 2017; pp. 31–38. [Google Scholar] [CrossRef] [Green Version]
  12. Fomin, F.; Giroire, F.; Jean-Marie, A.; Mazauric, D.; Nisse, N. To satisfy impatient Web surfers is hard. J. Theor. Comput. Sci. 2014, 526, 1–17. [Google Scholar] [CrossRef] [Green Version]
  13. Giroire, F.; Lamprou, I.; Mazauric, D.; Nisse, N.; Pérennes, S.; Soares, R. Connected surveillance game. Theor. Comput. Sci. 2015, 584, 131–143. [Google Scholar] [CrossRef]
  14. Flajolet, P.; Sedgewick, R. Analytic Combinatorics; Cambridge University Press: Cambridge, UK, 2009. [Google Scholar]
  15. Puterman, M. Markov Decision Processes—Discrete Stochastic Dynamic Programming; Wiley: New York, NY, USA, 1994. [Google Scholar]
  16. Keshava, K. Prefetching: A Markov Decision Process Model. BS-MS Degree Dissertation, Indian Institute of Science Education and Research, Mohali, India, 2021. [Google Scholar]
  17. Ross, S. Introduction to Stochastic Dynamic Programming; Academic Press: Cambridge, MA, USA, 1995. [Google Scholar]
Figure 1. A tree t of depth d = 3 with fanout p = 2 . There are 2 subtrees ( | s ( t ) | = 2 ), 6 internal nodes ( inode ( t ) = 6 ), and 5 leaves ( leaves ( t ) = 5 ).
Figure 1. A tree t of depth d = 3 with fanout p = 2 . There are 2 subtrees ( | s ( t ) | = 2 ), 6 internal nodes ( inode ( t ) = 6 ), and 5 leaves ( leaves ( t ) = 5 ).
Mathematics 09 02437 g001
Figure 2. There are 6 trees in T 2 , 2 , the set of rooted trees with depth 2 and fanout 2.
Figure 2. There are 6 trees in T 2 , 2 , the set of rooted trees with depth 2 and fanout 2.
Mathematics 09 02437 g002
Table 1. Instances of the cardinals of the sets of rooted trees of depth d with fanout p, when marks are ignored ( T p , d ), when marks are in { 0 , 1 } ( M p , d ), and when leaves have mark 0 and other nodes have marks in { 0 , 1 } ( M p , d + ), for different d and p.
Table 1. Instances of the cardinals of the sets of rooted trees of depth d with fanout p, when marks are ignored ( T p , d ), when marks are in { 0 , 1 } ( M p , d ), and when leaves have mark 0 and other nodes have marks in { 0 , 1 } ( M p , d + ), for different d and p.
FanoutCardinals of Sets d = 1 Cardinals of Sets d = 2 Cardinals of Sets d = 3 Cardinals of Sets d = 4
p T p , d M p , d M p , d + T p , d M p , d M p , d + T p , d M p , d M p , d + T p , d M p , d M p , d +
1142184116813216
2212463124042195,31232801806 > 10 10 > 10 7
332863945,52851660,879 > 10 14 > 10 8 > 10 14 > 10 43 > 10 25
44608340 > 10 7 9360 > 10 10 > 10 29 > 10 16 > 10 40 > 10 120 > 10 65
Table 2. Specification of the greedy finite optimal policy.
Table 2. Specification of the greedy finite optimal policy.
Tree TypeUnmarked SonsSpecification of Subtree SizeOptimal Action
Type 1 2 a ( d 1 , d 1 )
Type 2a1 j r 3 for some r a ( d 1 , l j c ) , j c = min r j r 3
j r < 3 for all r a ( d 1 , l j c ) , j c = max r j r
Type 2b0 j 1 > 3 j 2 j 1 a ( l j 1 , l j 1 )
j 1 3 j 2 j 1 a ( l j 1 , l j 2 )
Type 2c0 a ( l j 1 , l j 1 )
Table 3. Average cost of different greedy policies.
Table 3. Average cost of different greedy policies.
PolicyCost for p = 3 Cost for p = 4 Cost for p = 5
Greedy depth 10.1111110.2083330.286667
Greedy smallest0.0679120.1615680.229741
Greedy leftmost0.0628020.1602270.226289
Greedy largest0.0543690.1569070.217443
Greedy finite optimal0.0543690.1544010.208282
Table 4. Policy iteration performance.
Table 4. Policy iteration performance.
Fanout pNo. of IterationsSize of MatricesTime to Converge
3 1231 × 231  11 s
4 53336 × 3336 632 s
51457,860 × 57,8607898 s
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Keshava, K.; Jean-Marie, A.; Alouf, S. Optimal Prefetching in Random Trees. Mathematics 2021, 9, 2437. https://doi.org/10.3390/math9192437

AMA Style

Keshava K, Jean-Marie A, Alouf S. Optimal Prefetching in Random Trees. Mathematics. 2021; 9(19):2437. https://doi.org/10.3390/math9192437

Chicago/Turabian Style

Keshava, Kausthub, Alain Jean-Marie, and Sara Alouf. 2021. "Optimal Prefetching in Random Trees" Mathematics 9, no. 19: 2437. https://doi.org/10.3390/math9192437

APA Style

Keshava, K., Jean-Marie, A., & Alouf, S. (2021). Optimal Prefetching in Random Trees. Mathematics, 9(19), 2437. https://doi.org/10.3390/math9192437

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop