Next Article in Journal
The Spatiotemporal Dynamics of Insect Predator–Prey System Incorporating Refuge Effect
Previous Article in Journal
Principled Limitations on Self-Representation for Generic Physical Systems
Previous Article in Special Issue
Blahut–Arimoto Algorithms for Inner and Outer Bounds on Capacity Regions of Broadcast Channels
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Hierarchical Cache-Aided Networks for Linear Function Retrieval

1
Guangxi Key Laboratory of Multi-Source Information Mining & Security, Guangxi Normal University, Guilin 541004, China
2
The Department of Electrical Engineering, University of North Texas, Denton, TX 76207, USA
3
The School of Information Science and Technology, ShanghaiTech University, Shanghai 201210, China
*
Author to whom correspondence should be addressed.
Entropy 2024, 26(3), 195; https://doi.org/10.3390/e26030195
Submission received: 23 January 2024 / Revised: 14 February 2024 / Accepted: 19 February 2024 / Published: 25 February 2024
(This article belongs to the Special Issue Advances in Multiuser Information Theory)

Abstract

:
In a hierarchical caching system, a server is connected to multiple mirrors, each of which is connected to a different set of users, and both the mirrors and the users are equipped with caching memories. All the existing schemes focus on single file retrieval, i.e., each user requests one file. In this paper, we consider the linear function retrieval problem, i.e., each user requests a linear combination of files, which includes single file retrieval as a special case. We propose a new scheme that reduces the transmission load of the first hop by jointly utilizing the two layers’ cache memories, and we show that our scheme achieves the optimal load for the second hop in some cases.

1. Introduction

In order to reduce the transmission pressure of wireless networks during peak traffic times, Maddah-Ali and Niesen in [1] provided a ( K ,   M ,   N ) coded caching scheme (MN Scheme) where a single server has N files and connects K cache-aided users with the cache memories of M files through an error-free shared link. A coded caching scheme consists of two phases: (1) the placement phase, where the server is equipped with the data and each user’s cache is also equipped with the size of at most M files without knowledge of the users’ future demands; (2) the delivery phase, where each user randomly requests one file and then the server sends the coded signal to the users such that each user can decode its requested file with the help of its cached packets. It is shown that the MN Scheme is generally order-optimal within a factor of 2 [2] and optimal under the uncoded data placement when K N [3]. The MN Scheme is also widely used in different networks, such as combination networks [4,5], device-to-device networks [6], etc.
In practical scenarios, caching systems are transformed into multiple layers in order to make transmission more efficient, such as the hierarchical edge caching architecture for Internet of Vehicles [7], the three-tier mobile cloud-edge computing structure [8], and so on. In this paper, we particularly study the hierarchical caching system [9], a two-layer network as illustrated in Figure 1. A ( K 1 , K 2 ; M 1 , M 2 ; N ) hierarchical caching system consists of a single server with a library of N files, K 1 cache-aided mirror sites, and K 1 K 2 cache-aided users. For the first layer, the K 1 mirror sites are connected to the server through an error-free shared link, and for the second layer, each user connects to only one mirror. Our goal is to design a scheme to decrease the first load R 1 in the first hop (i.e., from the server to all the mirror sites) and the second load R 2 in the second hop (i.e., from each mirror site to its connected users).
The authors in [9] proposed the first hierarchical coded caching scheme (KNMD Scheme). The MN Scheme is applied two times in two layers consecutively. Although the KNMD Scheme achieves the optimal transmission load for the second hop, it involves a significant increase in R 1 since it ignores the users’ cache memory when designing the multicast message sent from the server. To improve the first load R 1 , the authors in [10,11] proposed new schemes that jointly use the two types of the MN Scheme together for the mirror sites and users, respectively.
It is worth noting that all the schemes consider the single file retrieval case, i.e., each user requests one file. The authors in [12] first considered the linear function retrieval scheme (WSJT Scheme), i.e., a linear combination of files is requested from each user through the shared link broadcast network. Clearly, linear function retrieval includes the single file retrieval case. In this paper, we study the linear function retrieval scheme for hierarchical networks and obtain the following results.
  • We first propose a baseline scheme via the WSJT Scheme and KNMD Scheme where the second-layer load achieves the optimal transmission load. However, we achieve this by sacrificing the first-layer load.
  • Then, in order to reduce the first-layer load, we propose another scheme whose second load also achieves optimality at the expense of increased subpacketization. Our scheme also aids in reducing the redundancy for some special demand distributions.
The rest of this paper is organized as follows. Section 2 formally introduces the system model and some existing schemes. Section 3 presents the main results. Section 4 gives an example and the general description of our scheme, i.e., the scheme for Theorem 2. The conclusion of this paper is given in Section 5.
Notations: For any positive integers a and b with a < b , let [ a : b ] { a , , b } and [ a ] [ 1 : a ] . Let [ b ] t { V | V [ b ] , | V | = t } , for any positive integer t b . For a positive integer n, the n-dimensional vector space over the field F q is denoted by F q n . For a given matrix P with row size X 1 , we divide it into X 1 parts by row, which is represented by P = { P ( x 1 ) | x 1 [ X 1 ] } . For any integer set T , define P T as the sub-matrix of P by selecting some rows from P , where the rows have indices in T . The rank of matrix P is denoted as rank ( P ) . The transpose of P is represented by P .

2. Preliminary

In this section, we give a formal description of the hierarchical caching system and review some existing related schemes for the hierarchical caching problem.

2.1. System Model

Consider a hierarchical network as shown in Figure 1. It consists of a single server with a library of N files, K 1 cache-aided mirror sites, and K 1 K 2 cache-aided users. For the first layer, the K 1 mirror sites are connected to the server through an error-free shared link, and for the second layer, each user connects to only one mirror. Mk1 represents the k 1 -th mirror and the k 2 -th user attached to M k 1 as U k 1 , k 2 , k 1 [ K 1 ] ,   k 2 [ K 2 ] , and the set of users attached to M k 1 as U k 1 . The server contains a collection of N files, denoted by W = { W ( 1 ) , W ( 2 ) , , W ( N ) } , each of which is uniformly distributed over F 2 B , where B N + . Each mirror and user is equipped with M 1 and M 2 files, respectively, where M 1 , M 2 0 . A ( K 1 ,   K 2 ;   M 1 ,   M 2 ;   N ) hierarchical caching system contains two phases.
  • Placement phase: The mirror site Mk1 caches some parts of the files by using a cache function φ k 1 : F 2 N B F 2 M 1 B , where M 1 N is the memory ratio of the mirror in the first layer, M 1 [ 0 : N ] . The cache contents of mirror M k 1 are
    Z k 1 = φ k 1 ( W ) ,   k 1 [ K 1 ] .
    The user U k 1 , k 2 caches some parts of the files by using a cache function ϕ k 1 , k 2 : F 2 N B F 2 M 2 B , where M 2 N is the memory ratio of users in the second layer, M 2 [ 0 : N ] . Then, the cache contents of U k 1 , k 2 are
    Z ˜ k 1 , k 2 = ϕ k 1 , k 2 ( W ) , k 1 [ K 1 ] ,   k 2 [ K 2 ] .
  • Delivery phase: Each user U k 1 , k 2 randomly requests a linear combination of the files
    L k 1 , k 2 = d k 1 , k 2 ( 1 ) W ( 1 ) + d k 1 , k 2 ( 2 ) W ( 2 ) + + d k 1 , k 2 ( N ) W ( N ) .
    for any k 1 [ K 1 ] , k 2 [ K 2 ] , where d k 1 , k 2 = ( d k 1 , k 2 ( 1 ) , , d k 1 , k 2 ( N ) ) F 2 N denotes the demand vector of user U k 1 , k 2 . When each user requests a single file, the demand vector d k 1 , k 2 is a N-length unit vector. For example, if user U k 1 , k 2 only requests the 1-st file, then the demand vector is set as d k 1 , k 2 = ( 1 , , 0 ) F 2 N , k 1 [ K 1 ] , k 2 [ K 2 ] , which is a special case of our proposed scheme. We can obtain the demand matrices of all users as follows:
    D ( k 1 ) = d k 1 , 1 d k 1 , K 2 , D = D ( 1 ) D ( K 1 ) .
    where D ( k 1 ) represents the demand vectors of U k 1 . Given the demand matrix D , we should consider the following two types of messages.
    The messages sent by the server: The server generates signal X server by using an encoding function χ : F 2 K 1 K 2 N × F 2 N B F 2 R 1 B , where
    X server = χ ( D , W ) .
    and then the server sends X server to the mirrors. The normalized number of transmissions R 1 is called the transmission load for the first layer.
    The messages sent by the mirror: Based on X server , Z k 1 , and D , each mirror M k 1 generates a signal X k 1 mirror by using the encoding function κ : F 2 K 2 N × F 2 M 1 B × X server F 2 R 2 B , where
    X k 1 mirror = κ ( D ( k 1 ) , Z k 1 , X server ) .
    and then mirror M k 1 sends X k 1 mirror to its connected users. The normalized number of transmissions R 2 is called the transmission load for the second layer.
    For the retrieval process, each user U k 1 , k 2 can decode its required linear combination of files from ( D , Z ˜ k 1 , k 2 , X k 1 mirror ) , which means that there exist decoding functions ξ k 1 , k 2 : F 2 K 1 K 2 N × F 2 M 2 B × F 2 R 2 B F 2 B , k 1 [ K 1 ] , k 2 [ K 2 ] , such that
    ξ k 1 , k 2 ( D , Z ˜ k 1 , k 2 , X k 1 mirror ) = d k 1 , k 2 ( 1 ) W ( 1 ) + + d k 1 , k 2 ( N ) W ( N ) .
We define the optimal transmission loads for the two layers as R 1 * and R 2 * separately.
R 1 * = inf χ , κ , ( ξ k 1 , k 2 ) k 1 [ K 1 ] , k 2 [ K 2 ] { R 1 } , R 2 * = inf χ , κ , ( ξ k 1 , k 2 ) k 1 [ K 1 ] , k 2 [ K 2 ] { R 2 } .
Our goal is to design schemes in which the transmission loads R 1 and R 2 are as small as possible.

2.2. Existing Schemes

In the following, we review the KNMD Scheme for the hierarchical caching problem and the WSJT Scheme over the binary field F 2 , which will be useful for the hierarchical caching system with linear function retrieval. First, let us outline the MN Scheme.
(1) MN Scheme [1]: Set t M K / N , when t [ 0 : K ] , N K , each file is partitioned into F = K t packets, i.e., for each n [ N ] , W ( n ) = ( W T ( n ) ) , where T [ K ] t . In the placement phase, for each user Uk, k [ K ] . The cache content of user Uk is Z k = { W T ( n ) | n [ N ] , k T , T [ K ] t } . In the delivery phase, the file W d k is requested by each user Uk, where d k [ N ] . Fixing a user k S , the user k requests the subfiles W d k , S { k } when it is presented in the cache of any user k S { k } . Then, the server transmits the coded signal k S W d k , S { k } , where S [ K ] of | S | = t + 1 . The transmission load R MN = K ( 1 M / N ) K M / N + 1 .
(2) KNMD Scheme [9]: This scheme uses the MN Scheme in each layer of the hierarchical network. More specifically, for the first layer between the server and K 1 mirrors, it uses the ( K 1 , M 1 , N ) MN Scheme K 2 times to recover all K 1 K 2 requested files, and then each mirror M k 1 , k 1 [ K 1 ] works as a server whose library contains K 2 files that are requested by users in U k 1 , and finally it utilizes the ( K 2 , M 2 , N ) MN Scheme between M k 1 and U k 1 . Then, each user can retrieve its requested file with the transmission load as follows.
R 1 = K 2 K 1 K 1 M 1 / N K 1 M 1 / N + 1 , R 2 = K 2 K 2 M 2 / N K 2 M 2 / N + 1 .
However, the MN Scheme only works for single file retrieval. The authors in [12] proposed a scheme (WSJT Scheme) that is suitable for the linear function retrieval problem.
(3) WSJT Scheme [12]: Using the placement strategy of the MN Scheme, each user Uk where k [ K ] requests a linear combination of files with demand vector d k F 2 N . After revealing the demand matrix D = ( d 1 , , d K ) with dimension K × N , the server broadcasts some coded packets by modifying the transmission strategy of the MN Scheme such that each user is able to recover its demanded linear combination of files with the transmission load
R WSJT = K t + 1 K rank ( D ) t + 1 K t , t [ 0 : K ] .
It is worth noting that when D is row full rank, R WSJT is optimal under the uncoded placement.

3. Main Results

In this section, we first propose a baseline scheme via the WSJT Scheme where R 2 achieves optimality when the sub-matrix D ( k 1 ) , k 1 [ K 1 ] , is full rank. Then, we propose another scheme that improves R 1 while the R 2 remains unchanged compared with the Baseline Scheme. Finally, some theoretical and numerical comparisons are provided.
For the sake of convenience in proposing another scheme for some special demand distributions, the following definitions of the leader mirror and user sets are necessary.
Definition 1
(Leader mirror set). For a K 1 K 2 × N demand matrix D in (1), we call a subset of mirrors the leader mirror set, which is represented by L M = { l 1 , , l | L M | } , L M [ K 1 ] , if it satisfies the following condition for k 1 [ K 1 ] , k 2 [ K 2 ]
d k 1 , k 2 = α 1 ( k 1 ) d l 1 , k 2 + + α | L M | ( k 1 ) d l | L M | , k 2 .
and it has the minimum cardinality among all the subsets satisfying (2), where ( α 1 ( k 1 ) , , α | L M | ( k 1 ) ) F 2 | L M | .
Definition 2
(Leader user set). For a K 2 × N demand matrix D ( k 1 ) in (1), we call a subset of users the leader user set, which is represented by L k 1 = { l 1 , , l | L k 1 | } , L k 1 [ K 2 ] , if, for any k 1 [ K 1 ] , k 2 [ K 2 ] , it satisfies the condition (3) and it has the minimum cardinality among all the subsets satisfying (3), where ( α 1 , , α | L k 1 | ) F 2 | L k 1 | :
d k 1 , k 2 = α 1 d k 1 , 1 + + α | L k 1 | d k 1 , l | L k 1 | .
Now, we introduce the Baseline Scheme, which is generated by using the KNMD Scheme in [9] and the WSJT Scheme in [12]. We utilize the WSJTC Scheme to replace the MN Scheme in the KNMD Scheme, and then we obtain the Baseline Scheme, which is suitable for the linear function retrieval problem in the hierarchical network.
Theorem 1
(Baseline Scheme). For any positive integers K 1 , K 2 , t 1 [ K 1 ] , t 2 [ K 2 ] and the demand matrix D in (1), there exists a ( K 1 ,   K 2 ;   M 1 ,   M 2 ;   N ) hierarchical coded caching scheme for a linear function retrieval problem with memory ratios M 1 N = t 1 K 1 , M 2 N = t 2 K 2 and transmission loads
R base 1 = K 2 K 1 t 1 + 1 K 1 | L M | t 1 + 1 / K 1 t 1 .
R base 2 = max k 1 [ K 1 ] K 2 t 2 + 1 K 2 rank ( D ( k 1 ) ) t 2 + 1 / K 2 t 2 .
where L M is defined in Definition 1.
In fact, the transmission loads are related to the placement strategy and demand distribution, respectively. The KNMD Scheme considers the first and second layers separately and ignores the users’ and mirrors’ cache memories, which leads to good performance on R 2 but results in a large transmission load R 1 . For the second layer, it can be regarded as a ( K 2 , M 2 , N ) shared link caching problem in which the WSJT Scheme achieves the optimal transmission load under certain circumstances, i.e., when the sub-matrix D ( k 1 ) , k 1 [ K 1 ] , is full rank, R base 2 = R 2 * . For the purpose of improving R 1 , we propose another scheme, stated below, and the proof is included in Section 4.
Theorem 2.
For any positive integers K 1 , K 2 , t 1 [ K 1 ] , t 2 [ K 2 ] and the demand matrix D in (1), there exists a ( K 1 ,   K 2 ;   M 1 ,   M 2 ;   N ) hierarchical coded caching scheme for a linear function retrieval problem with memory ratios M 1 N = t 1 K 1 , M 2 N = t 2 K 2 and transmission loads
R 1 = K 1 t 1 + 1 K 1 | L M | ) t 1 + 1 K 2 t 2 + 1 / K 1 t 1 K 2 t 2 , R 2 = max k 1 [ K 1 ] K 2 t 2 + 1 K 2 rank ( D ( k 1 ) ) t 2 + 1 / K 2 t 2 .
Now, let us consider the performance of our two schemes. For the first layer, we claim that R 1 < 1 t 2 + 1 R base 1 , where t 2 0 since
R 1 R base 1 = K 1 t 1 + 1 K 1 | L M | ) t 1 + 1 K 2 t 2 + 1 K 1 t 1 K 2 K 1 t 1 + 1 K 1 | L M | t 1 + 1 K 1 t 1 K 2 t 2 = K 2 t 2 + 1 K 2 K 2 t 2 = K 2 ! t 2 ! ( K 2 t 2 ) ! K 2 K 2 ! ( t 2 + 1 ) ! ( K 2 t 2 1 ) ! = t 2 ! ( K 2 t 2 ) ( K 2 t 2 1 ) ! K 2 t 2 ! ( t 2 + 1 ) ( K 2 t 2 1 ) ! = K 2 t 2 K 2 · 1 t 2 + 1 1 t 2 + 1 .
Obviously, this scheme has the same performance as the Baseline Scheme, i.e., R 2 = R base 2 , which also achieves the optimal transmission load when the demand matrix D ( k 1 ) , k 1 [ K 1 ] , is full rank.
Finally, we perform a numerical comparison to further show the performance of our scheme. In Figure 2, we compare the Baseline Scheme with the scheme for Theorem 2 with fixed parameters ( K 1 ,   K 2 ,   | L M | ,   N ) = ( 20 ,   10 ,   10 ,   200 ) and varying the memory ratio M 1 / N from 0 to 1 with a step size 0.1 . As seen in Figure 2, compared to the Baseline Scheme, the scheme for Theorem 2 can reduce the transmission load R 1 significantly, as this scheme utilizes both the user’s cache and the mirror’s cache when constructing the multicast message sent by the server. The scheme for Theorem 2 achieves the same R 2 as the Baseline Scheme, while our scheme has a lower transmission load R 1 .

4. Scheme for Theorem 2

In this section, we first give an illustrative example of our scheme. Then, the general description of the scheme is provided. Before the description, we first introduce the following lemmas regarding the message sent by the server and mirrors, whose proofs are included in Appendix A and Appendix B, respectively.
Lemma 1
(The messages sent by the server). Given a demand matrix D in (1), the leader mirror set L M , and a user set B ( [ K 2 ] t 2 + 1 ) , if there exists a mirror set C ( [ K 1 ] | L M | + t 1 + 1 ) , where L M C , let V C be the family of mirror set V , V C , where each V satisfies Definition 1. Then, we have V V C X C V , B = 0 , where X C V , B represents the message sent by the server, which is defined in (8).
Lemma 2
(The messages sent by the mirror). Given a sub-matrix D ( k 1 ) , k 1 [ K 1 ] of D , the leader user set L k 1 , and a mirror set T 1 ( [ K 1 ] t 1 ) , if there exists a user set C ( [ K 2 ] | L k 1 | + t 2 + 1 ) , where L k 1 C , let V C be the family of all set V , V C , where each V satisfies Definition 2. Then, we have V V C X T 1 , C V ( k 1 ) = 0 , where X T 1 , C V ( k 1 ) represents the message sent by the mirror M k 1 , which is defined in (9).
By Lemma 1, for any mirror set A [ K 1 ] t 1 + 1 , L M A = , and the message X A , B , B [ K 2 ] t 2 + 1 can be computed directly from the broadcast messages by using the following equation
X A , B = V V C L M X C V , B
where C = A L M .
By Lemma 2, for any user set B = [ K 2 ] t 2 + 1 , L k 1 B = , the message X T 1 , B ( k 1 ) can be computed directly from the broadcast messages by using the equation
X T 1 , B ( k 1 ) = V V C L k 1 X T 1 , C V ( k 1 )
where C = B L k 1 . After receiving the messages sent by the mirror M k 1 , user Uk1,k2 is able to recover its desired linear combination of files.

4.1. An Example for Theorem 2

When K 1 = 3 , K 2 = 2 , t 1 = t 2 = 1 , we can obtain an F- ( K 1 , K 2 ; M 1 , M 2 ; N ) = 6 ( 3 , 2 ; 2 , 3 ; 6 ) coded caching scheme as follows.
  • Placement phase: Each file from F 2 B is divided into 3 1 2 1 = 6 subfiles with equal size, i.e., W ( n ) = { W 1 , 1 ( n ) , W 1 , 2 ( n ) , , W 3 , 1 ( n ) , W 3 , 2 ( n ) } , n [ 6 ] . For simplicity, we represent a set that is the subscript of some studied object by a string. For example, T { 1 , 2 } is represented by T 12 . The contents cached by the mirrors are as follows:
    Z 1 = { W 1 , 1 ( n ) , W 1 , 2 ( n ) | n [ 6 ] } , Z 2 = { W 2 , 1 ( n ) , W 2 , 2 ( n ) | n [ 6 ] } , Z 3 = { W 3 , 1 ( n ) , W 3 , 2 ( n ) | n [ 6 ] } .
    The subfiles cached by the users are as follows:
    Z ˜ 1 , 1 = Z ˜ 2 , 1 = Z ˜ 3 , 1 = { W 1 , 1 ( n ) , W 2 , 1 ( n ) , W 3 , 1 ( n ) | n [ 6 ] } , Z ˜ 1 , 2 = Z ˜ 2 , 2 = Z ˜ 3 , 2 = { W 1 , 2 ( n ) , W 2 , 2 ( n ) , W 3 , 2 ( n ) | n [ 6 ] } .
  • Delivery phase: In the delivery phase, the demand vectors with length 12 are d k 1 , 1 = ( 1 , 1 , 0 , 0 , 0 , 0 ) , d k 1 , 2 = ( 0 , 0 , 1 , 1 , 0 , 0 ) , k 1 [ 3 ] . As we can see, D ( 1 ) = D ( 2 ) = D ( 3 ) . Without loss of generality, we set L M = { 1 } . We denote a linear combination of subfiles as
    L d k 1 , k 2 T 1 , T 2 = n [ N ] d k 1 , k 2 ( n ) · W T 1 , T 2 ( n ) .
    where T 1 { { 1 } , { 2 } , { 3 } } and T 2 { { 1 } , { 2 } } , k 1 [ 3 ] , k 2 [ 2 ] . Then, the messages sent in this hierarchical system consist of the following two parts.
    The messages sent by the server: The server generates signal X A , B satisfying A [ 3 ] 2 , B [ 2 ] 2 and A L M as follows:
    X 12 , 12 = L d 1 , 1 , 2 , 2 L d 1 , 2 , 2 , 1 L d 2 , 1 , 1 , 2 L d 2 , 2 , 1 , 1 = ( i [ 2 ] ( W 2 , 2 ( i ) W 1 , 2 ( i ) ) ) ( i [ 3 : 4 ] ( W 2 , 1 ( i ) W 1 , 1 ( i ) ) ) , X 13 , 12 = L d 1 , 1 , 3 , 2 L d 1 , 2 , 3 , 1 L d 3 , 1 , 1 , 2 L d 3 , 2 , 1 , 1 = ( i [ 2 ] ( W 3 , 2 ( i ) W 1 , 2 ( i ) ) ) ( i [ 3 : 4 ] ( W 3 , 1 ( i ) W 1 , 1 ( i ) ) ) .
    In this example, we have L M = { 1 } , and A = { 2 , 3 } has no intersection with L M . Here, we have C = L M A = { 1 , 2 , 3 } and V C = { { 1 } , { 2 } , { 3 } } . By Lemma 2, we can generate X 23 , 12 = X 12 , 12 + X 13 , 12 = ( i [ 2 ] ( W 3 , 2 ( i ) W 2 , 2 ( i ) ) ) ( i [ 3 : 4 ] ( W 3 , 1 ( i ) W 2 , 1 ( i ) ) ) . Thus, the transmission load of the first layer is R 1 = 3 1 6 = 1 / 3 .
    The messages sent by mirror M k 1 : Here, we take mirror M 1 as an example. From D ( 1 ) , we have L 1 = { 1 , 2 } , and M 1 transmits X T 1 , B ( 1 ) , where T 1 [ 3 ] 1 , B [ 2 ] 2 , B L 1 , i.e.,
    X 2 , 12 ( 1 ) = X 12 , 12 X 1 , 12 ( 2 ) = ( i [ 2 ] W 2 , 2 ( i ) ) ( i [ 3 : 4 ] W 2 , 1 ( i ) ) , X 3 , 12 ( 1 ) = X 13 , 12 X 1 , 12 ( 3 ) = ( i [ 2 ] W 3 , 2 ( i ) ) ( i [ 3 : 4 ] W 3 , 1 ( i ) ) , X 1 , 12 ( 1 ) = W 1 , 2 ( 1 ) W 1 , 2 ( 2 ) W 1 , 1 ( 3 ) W 1 , 1 ( 4 ) .
    Then, the transmission amount by mirror M 1 is 3 packets, and the transmission load of the second layer is R 2 = 3 6 = 1 2 .
User U 1 , 1 can decode W 1 , 2 ( 1 ) W 1 , 2 ( 2 ) , W 2 , 2 ( 1 ) W 2 , 2 ( 2 ) , W 3 , 2 ( 1 ) W 3 , 2 ( 2 ) , from X 1 , 12 ( 1 ) , X 2 , 12 ( 1 ) , X 3 , 12 ( 1 ) , respectively, as it has cached { W 1 , 1 ( n ) , W 2 , 1 ( n ) , W 3 , 1 ( n ) | n [ 6 ] } .
Compared with the Baseline Scheme, which achieves R base 1 = 4 3 , R base 2 = 1 2 , our scheme has a significant improvement in R 1 .

4.2. General Description of Scheme for Theorem 2

Given a ( K 1 , K 2 ; M 1 , M 2 ; N ) hierarchical caching system, we have an F- ( K 1 , K 2 , M 1 , M 2 , N ) coded caching scheme where F = K 1 t 1 K 2 t 2 , t 1 [ 0 : K 1 ] , t 2 [ 0 : K 2 ] . The scheme consists of two phases.
  • Placement phase: Firstly, we divide each file into K 1 t 1 equal-size subfiles; then, we further divide each subfile into K 2 t 2 sub-subfiles. The index of subfiles consists of two parts, T 1 and T 2 , i.e., W ( n ) = { W T 1 , T 2 ( n ) | T 1 [ K 1 ] t 1 , T 2 [ K 2 ] t 2 } , n [ N ] . Each mirror site M k 1 , k 1 [ K 1 ] caches subfiles W T 1 , T 2 ( n ) according to the following rule, which is mainly related to the first subscript T 1 .
    Z k 1 = W T 1 , T 2 ( n ) | T 1 [ K 1 ] t 1 , T 2 [ K 2 ] t 2 , k 1 T 1 , n [ N ] .
    Similarly, each user U k 1 , k 2 , k 1 [ K 1 ] , k 2 [ K 2 ] caches subfiles W T 1 , T 2 ( n ) according to the following rule, which is mainly related to the second subscript T 2 .
    Z ˜ k 1 , k 2 = W T 1 , T 2 ( n ) | T 1 [ K 1 ] t 1 , T 2 [ K 2 ] t 2 , k 2 T 2 , n [ N ] .
    Under this caching strategy, we can verify that it satisfies the memory constraints stated in Theorem 2. Each mirror caches K 1 1 t 1 1 K 2 t 2 N subfiles and each user caches K 1 t 1 K 2 1 t 2 1 N subfiles, where each subfile is B / K 1 t 1 K 2 t 2 bits. Thus, the memory ratios of the mirror and user are M 1 N = t 1 K 1 and M 2 N = t 2 K 2 , respectively. For any user’s demand vector d k 1 , k 2 = ( d k 1 , k 2 ( 1 ) , , d k 1 , k 2 ( N ) ) of N-length, we use the notation as follows to denote a linear combination of subfiles:
    L d k 1 , k 2 T 1 , T 2 = n [ N ] d k 1 , k 2 ( n ) W T 1 , T 2 ( n ) , T 1 [ K 1 ] t 1 , T 2 [ K 2 ] t 2 .
  • Delivery phase: For the convenience of the subsequent discussion, we first give the following two definitions of the signals transmitted in the first layer, say X A , B , and the second layer, say X T 1 , B ( k 1 ) . For any mirror set containing t 1 + 1 mirrors defined as A [ K 1 ] t 1 + 1 , any mirror site set containing t 1 mirror sites defined as T 1 [ K 1 ] t 1 , and any user set containing t 2 + 1 users defined as B [ K 2 ] t 2 + 1 , we define
    X A , B = k 1 A k 2 B L d k 1 , k 2 , A { k 1 } , B { k 2 } ,
    X T 1 , B ( k 1 ) = k 2 B L d k 1 , k 2 , T 1 , B { k 2 } .
    After the demand matrix D of size K 1 K 2 × N and its sub-matrix D ( k 1 ) of size K 2 × N in (1) are revealed, we have the leader mirror set L M according to Definition 1. For each sub-matrix D ( k 1 ) of D , k 1 [ K 1 ] , we have the leader user set L k 1 , L k 1 [ K 2 ] according to Definition 2. There are two types of messages transmitted by the server and mirror, respectively.
    The messages sent by the server: For each A [ K 1 ] t 1 + 1 , B [ K 2 ] t 2 + 1 , L M A , the server transmits X A , B to the mirror sites.
    The messages sent by the mirror: Mirror site M k 1 transmits X T 1 , B ( k 1 ) via the following rules.
    (1) For each T 1 [ K 1 ] t 1 , k 1 T 1 , A = T 1 { k 1 } , B [ K 2 ] t 2 + 1 , B L k 1 , mirror M k 1 transmits X T 1 , B ( k 1 ) by subtracting k 1 T 1 X A { k 1 } , B ( k 1 ) from X A , B , i.e.,
    X T 1 , B ( k 1 ) = X A , B k 1 T 1 X A { k 1 } , B ( k 1 ) .
    (2) For each T 1 [ K 1 ] t 1 , k 1 T 1 , B [ K 2 ] t 2 + 1 , B L k 1 , mirror M k 1 directly transmits X T 1 , B ( k 1 ) to its connected users generated from its cached content Z k 1 .
    As regards the messages X A , B , A L M = , and X T 1 , B ( k 1 ) , B L k 1 = , which are also necessary for the users, these messages can be computed from the sent messages by using Lemmas 1 and 2. More precisely, X A , B , A L M = can be obtained by (5), and X T 1 , B ( k 1 ) , B L k 1 = can be obtained by (6).
Now, we prove that each message X A , B transmitted by the server is decodable, i.e., after each mirror subtracts some packets from X A , B , the rest of the message only contains coded packets required by the users in U k 1 . Then, we further prove that each message X T , B ( k 1 ) transmitted by M k 1 is decodable, i.e., after user Uk1,k2, k 1 [ K 1 ] , k 2 [ K 2 ] , subtracting some packets from X T , B ( k 1 ) , the rest of the message only contains coded packets required by user Uk1,k2.

4.2.1. Decodability of Mirror

For each mirror M k 1 , k 1 [ K 1 ] , it can receive or recover all the X A , B A [ K 1 ] , B [ K 2 ] , from the server. By (8), we have
X A , B = k 1 A k 2 B L d k 1 , k 2 , A { k 1 } , B { k 2 } = k 1 A k 2 B n [ N ] d k 1 , k 2 ( n ) W A { k 1 } , B { k 2 } ( n ) = k 2 B n [ N ] d k 1 , k 2 ( n ) W A { k 1 } , B { k 2 } ( n )
                              + k 1 A { k 1 } k 2 B n [ N ] d k 1 , k 2 ( n ) W A { k 1 } , B { k 2 } ( n )
                                                            = X A { k 1 } , B ( k 1 ) The coded packets required by users in U k 1 .                             + k 1 A { k 1 } k 2 B n [ N ] d k 1 , k 2 ( n ) W A { k 1 } , B { k 2 } ( n ) The coded packets cached by M k 1 .
where (10) holds directly from (7), and (11) holds by separating k 1 from A . Moreover, (11) holds by (7) and (9). The first term of (11) denotes coded packets that will be transmitted to U k 1 and the second term denotes packets cached by M k 1 because k 1 A { k 1 } .

4.2.2. Decodability of User

For each user U k 1 , k 2 , k 1 [ K 1 ] , k 2 [ K 2 ] , it can receive all the X T 1 , B ( k 1 ) , T 1 [ K 1 ] , B [ K 2 ] , k 2 B , from mirror M k 1 . By (9), we have
X T 1 , B ( k 1 ) = k 2 B L d k 1 , k 2 , T 1 , B { k 2 } = n [ N ] d k 1 , k 2 ( n ) W T 1 , B { k 2 } ( n ) requested by user U k 1 , k 2
+ k 2 B { k 2 } n [ N ] d k 1 , k 2 ( n ) W T 1 , B { k 2 } ( n ) cached by user U k 1 , k 2 .
where (12) holds directly by separating k 2 from B . It is clear that user Uk1,k2 can decode its desired linear combination of packets, i.e., the first term of (12), by subtracting the cached contents, i.e., the second term of (12), as k 2 B { k 2 } , which means that Uk1,k2 has already cached the packets from X T 1 , B ( k 1 ) .

4.2.3. Performance

From the placement phase, each file is firstly divided into K 1 t 1 subfiles and then each subfile is further divided into K 2 t 2 subfiles, so the subpacketization is K 1 t 1 K 2 t 2 . Each subfile is B / K 1 t 1 K 2 t 2 bits, each mirror caches K 1 1 t 1 1 K 2 t 2 N subfiles, and each user caches K 1 t 1 K 2 1 t 2 1 N subfiles. Thus, the memory ratios of the mirror and user are M 1 N = t 1 K 1 and M 2 N = t 2 K 2 , which satisfy the memory constraints in Theorem 2. In total, the server transmits K 1 t 1 + 1 K 1 | L M | t 1 + 1 K 2 t 2 + 1 multicast messages, and the mirror transmits K 2 t 2 + 1 K 2 rank ( D ( k 1 ) ) t 1 + 1 K 1 t 1 multicast messages. Each message contains B / K 1 t 1 K 2 t 2 bits, so the transmission loads of the first layer and the second layer are as illustrated in (4). Although the scheme for Theorem 2 has a higher subpacketization level of K 1 t 1 K 2 t 2 compared with K 1 t 1 + K 2 t 2 of the Baseline Scheme, we achieve a much lower transmission load R 1 under the same transmission load R 2 , where both schemes achieve the optimal transmission load of the second layer when the sub-demand matrix D ( k 1 ) , k 1 [ K 1 ] is full rank.

5. Conclusions

In this paper, we studied the linear function retrieval problem for hierarchical cache-aided networks. We proposed two schemes, where the first scheme achieves the optimal transmission load for the second layer for some demand distribution and our second scheme further reduces the load of the first layer while maintaining the same transmission load in the second layer.

Author Contributions

Conceptualization, L.Z. and Y.K.; methodology, L.Z., Y.K., Y.W. and M.C.; writing—original draft preparation, L.Z.; writing—review and editing, Y.W. and M.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work is in part supported by 2022GXNSFDA035087, NSFC (No. 62061004).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A. Proof of Lemma 1

Proof. 
Without loss of generality, we assume that L M = [ | L M | ] . From (8) and (7), we have
V V C X C V , B = V V C k 1 C V k 2 B L d k 1 , k 2 , C ( V { k 1 } ) , B { k 2 } = V V C k 1 C V k 2 B n [ N ] d k 1 , k 2 ( n ) · W C ( V { k 1 } ) , B { k 2 } ( n ) .
If the occurrence number of each subfile W T 1 , T 2 ( n ) , T 1 [ K 1 ] t 1 , T 2 [ K 2 ] t 2 , n [ N ] is even in V V C X C V , B , then the coefficient of W T 1 , T 2 ( n ) in the summation is 0. Note that if we focus on W T 1 , T 2 ( n ) , then k 2 = B T 2 , which is a fixed user label. W T 1 , T 2 ( n ) appears in V V C X C V , B if and only if there exist some mirrors M k 1 , k 1 C T 1 , which satisfy the two conditions: d k 1 , k 2 ( n ) 0 , C ( { k 1 } T 1 ) V C . Moreover, for each k 1 C T 1 satisfying the two conditions, there exists one coded message that contains W T 1 , T 2 ( n ) . Thus, we only need to prove that the number of mirrors M k 1 , k 1 C T 1 satisfying the two conditions is even.
Assume that mirror k 1 satisfies the two conditions; then, we have d k 1 , k 2 ( n ) 0 and C ( { k 1 } T 1 ) V C . Let L M = C ( { x } T 1 ) = { l 1 , , l | L M | } , and L M is also a leader mirror set satisfying Definition 1. Then, by (2), we have d k 1 , k 2 = α 1 ( k 1 ) d l 1 , k 2 + + α | L M | ( k 1 ) d l | L M | , k 2 . Then, there must be k 1 mirrors in α 1 ( k 1 ) d l 1 , k 2 + + α | L M | ( k 1 ) d l | L M | , k 2 satisfying d k 1 , k 2 ( n ) 0 and the corresponding coefficient α l M ( k 1 ) 0 , l M [ | L M | ] . k 1 is an odd number; otherwise, d k 1 , k 2 ( n ) = 0 . It is easy to check that these k 1 mirrors also satisfy constraints d k 1 , k 2 ( n ) 0 , C ( { k 1 } T 1 ) V C . Thus, there are in total k 1 + 1 mirrors in C T 1 , which is an even number, satisfying the two constraints. □

Appendix B. Proof of Lemma 2

Proof. 
Without loss of generality, we assume that L k 1 = [ | L k 1 | ] . From (9) and (7), we have
V V C X T 1 , C V ( k 1 ) = V V C k 2 C V L d k 1 , k 2 , T 1 , C ( V { k 2 } ) = V V C k 2 C V n [ N ] d k 1 , k 2 ( n ) · W T 1 , C ( V { k 2 } ) ( n ) .
If the occurrence number of subfile W T 1 , T 2 ( n ) is even in V V C X T 1 , C V ( k 1 ) , then the coefficient of W T 1 , T 2 ( n ) in the summation is 0. W T 1 , T 2 ( n ) appears in V V C X T 1 , C V ( k 1 ) if and only if there exist some users Uk1,x, x C T 2 , which satisfy the following two conditions: d k 1 , x ( n ) 0 and C ( { x } T 2 ) V C .
Moreover, for each x C T 2 satisfying the two conditions, there exists one coded message that contains W T 1 , T 2 ( n ) . Thus, we only need to prove that the number of users Uk1,x, x C T 2 satisfying the two conditions is even.
Assume that user Uk1,x satisfies the two conditions; then, we have d k 1 , x ( n ) 0 and C ( { x } T 2 ) V C . Let L k 1 = C ( { x } T 2 ) = { l 1 , , l | L k 1 | } , and L k 1 is also a leader user set among users in U k 1 , where rank ( D ( k 1 ) ) = rank ( D L k 1 ) . Then, we have
d k 1 , k 2 = α 1 d k 1 , l 1 + + α | L k 1 | d k 1 , l | L k 1 | , ( α 1 , , α | L k 1 | ) [ F 2 ] | L k 1 | .
Then, there must be x users in α 1 d k 1 , l 1 + + α | L k 1 | d k 1 , l | L k 1 | , ( α 1 , , α | L k 1 | ) [ F 2 ] | L k 1 | satisfying d k 1 , x ( n ) 0 and the corresponding coefficient α l k 1 0 , k 1 [ | L k 1 | ] . x is an odd number; otherwise, d k 1 , x ( n ) = 0 . It is easy to check that these x users also satisfy C ( { x } T 2 ) V C . Thus, there are in total x + 1 users in C T 1 , which is an even number, satisfying the two constraints. □

References

  1. Maddah-Ali, M.A.; Niesen, U. Fundamental limits of caching. IEEE Trans. Inf. Theory 2014, 60, 2856–2867. [Google Scholar] [CrossRef]
  2. Yu, Q.; Maddah-Ali, M.A.; Avestimehr, A.S. Characterizing the rate-memory tradeoff in cache networks within a factor of 2. IEEE Trans. Inf. Theory 2019, 65, 647–663. [Google Scholar] [CrossRef]
  3. Wan, K.; Tuninetti, D.; Piantanida, P. An index coding approach to caching with uncoded cache placement. IEEE Trans. Inf. Theory 2020, 66, 1318–1332. [Google Scholar] [CrossRef]
  4. Wan, K.; Ji, M.; Piantanida, P.; Tuninetti, D. Caching in combination networks: Novel multicast message generation and delivery by leveraging the network topology. In Proceedings of the 2018 IEEE International Conference on Communications, Kansas City, MO, USA, 20–24 May 2018; pp. 1–6. [Google Scholar]
  5. Ji, M.; Tulino, A.M.; Llorca, J.; Caire, G. Caching in combination networks. In Proceedings of the 2015 49th Asilomar Conference on Signals, Systems & Computers, Pacific Grove, CA, USA, 8–11 November 2015; pp. 1269–1273. [Google Scholar]
  6. Ji, M.; Caire, G.; Molisch, A.F. Fundamental limits of caching in wireless D2D networks. IEEE Trans. Inf. Theory 2016, 62, 849–869. [Google Scholar] [CrossRef]
  7. Zhou, H.; Jiang, K.; He, S.B.; Min, G.Y.; Wu, J. Distributed Deep Multi-Agent Reinforcement Learning for Cooperative Edge Caching in Internet-of-Vehicles. IEEE Trans. Wireless Commun. 2023, 22, 9595–9609. [Google Scholar] [CrossRef]
  8. Zhou, H.; Wang, Z.; Zheng, H.T.; He, S.B.; Dong, M.X. Cost Minimization-Oriented Computation Offloading and Service Caching in Mobile Cloud-Edge Computing: An A3C-Based Approach. IEEE Trans. Netw. Sci. Eng 2023, 10, 1326–1338. [Google Scholar] [CrossRef]
  9. Karamchandani, N.; Niesen, U.; Maddah-Ali, M.A.; Diggavi, S.N. Hierarchical coded caching. IEEE Trans. Inf. Theory 2016, 62, 3212–3229. [Google Scholar] [CrossRef]
  10. Wang, K.; Wu, Y.; Chen, J.; Yin, H. Reduce transmission delay for caching-aided two-layer networks. In Proceedings of the 2023 IEEE International Symposium on Information Theory, Paris, France, 7–12 July 2019; pp. 2019–2023. [Google Scholar]
  11. Kong, Y.; Wu, Y.; Cheng, M. Combinatorial designs for coded caching on hierarchical networks. In Proceedings of the IEEE Conference on Wireless Communications and Networking, Glasgow, UK, 26–29 March 2023; pp. 1–6. [Google Scholar]
  12. Wan, K.; Sun, H.; Ji, M.; Tuninetti, D.; Caire, G. On the optimal load-memory tradeoff of cache-aided scalar linear function retrieval. IEEE Trans. Inf. Theory 2021, 67, 4001–4018. [Google Scholar] [CrossRef]
Figure 1. The ( K 1 ,   K 2 ;   M 1 ,   M 2 ;   N ) hierarchical caching system with N = 4 , K 1 = K 2 = 2 , M 1 = 2 , and M 2 = 1 .
Figure 1. The ( K 1 ,   K 2 ;   M 1 ,   M 2 ;   N ) hierarchical caching system with N = 4 , K 1 = K 2 = 2 , M 1 = 2 , and M 2 = 1 .
Entropy 26 00195 g001
Figure 2. R 1 on N = 200 , K 1 = 20 , K 2 = 10 .
Figure 2. R 1 on N = 200 , K 1 = 20 , K 2 = 10 .
Entropy 26 00195 g002
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, L.; Kong, Y.; Wu, Y.; Cheng, M. Hierarchical Cache-Aided Networks for Linear Function Retrieval. Entropy 2024, 26, 195. https://doi.org/10.3390/e26030195

AMA Style

Zhang L, Kong Y, Wu Y, Cheng M. Hierarchical Cache-Aided Networks for Linear Function Retrieval. Entropy. 2024; 26(3):195. https://doi.org/10.3390/e26030195

Chicago/Turabian Style

Zhang, Lingyu, Yun Kong, Youlong Wu, and Minquan Cheng. 2024. "Hierarchical Cache-Aided Networks for Linear Function Retrieval" Entropy 26, no. 3: 195. https://doi.org/10.3390/e26030195

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop