Next Article in Journal
Weighted Sobolev–Morrey Estimates for Nondivergence Degenerate Operators with Drift on Homogeneous Groups
Previous Article in Journal
Aboodh Transform Iterative Method for Solving Fractional Partial Differential Equation with Mittag–Leffler Kernel
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Information Limits for Community Detection in Hypergraph with Label Information

1
School of Mathematics and Statistics, North China University of Water Resources and Electric Power, Zhengzhou 450045, China
2
Department of Statistics, North Dakota State University, Fargo, ND 58103, USA
*
Author to whom correspondence should be addressed.
Symmetry 2021, 13(11), 2060; https://doi.org/10.3390/sym13112060
Submission received: 26 September 2021 / Revised: 11 October 2021 / Accepted: 21 October 2021 / Published: 1 November 2021

Abstract

:
In network data mining, community detection refers to the problem of partitioning the nodes of a network into clusters (communities). This is equivalent to identifying the cluster label of each node. A label estimator is said to be an exact recovery of the true labels (communities) if it coincides with the true labels with a probability convergent to one. In this work, we consider the effect of label information on the exact recovery of communities in an m-uniform Hypergraph Stochastic Block Model (HSBM). We investigate two scenarios of label information: ( 1 ) a noisy label for each node is observed independently, with 1 α n as the probability that the noisy label will match the true label; ( 2 ) the true label of each node is observed independently, with the probability of 1 α n . We derive sharp boundaries for exact recovery under both scenarios from an information-theoretical point of view. The label information improves the sharp detection boundary if and only if α n = n β + o ( 1 ) for a constant β > 0 .

1. Introduction

A graph or network consists of a set of nodes (vertices) and an edge set. Graphs have been used extensively to model a variety of systems in many fields [1,2,3,4,5]. Due to the widespread application, network data analysis has drawn a lot of attention in both statistical and machine learning communities [6,7,8,9,10,11,12,13]. Real-world networks are usually more complex than ordinary graphs, and in this case, a hypergraph is a popular alternative model [8,14,15,16,17,18,19,20,21]. Given a positive integer n, let V = [ n ] : = { 1 , 2 , , n } . An undirected m-uniform hypergraph on V is a pair H m = ( V , E ) in which E is a set of subsets of V such that | e | = m for every e E , and each element in E is called a hyperedge. That is, in H m , each hyperedge consists of exactly m distinct nodes. For i 1 < i 2 < < i m , we denote A i 1 i 2 i m = 1 if { i 1 , i 2 , , i m } is a hyperedge, and A i 1 i 2 i m = 0 if otherwise. Suppose A i 1 i 2 i m = A j 1 j 2 j m if { i 1 , i 2 , , i m } = { j 1 , j 2 , , j m } , and A i 1 i 2 i m = 0 if | { i 1 , i 2 , , i m } | m . That is, a hypergraph that is symmetric and a self-loop is not allowed. Then, the m-dimensional symmetric array A = ( A i 1 , , i m ) { 0 , 1 } n m is called the adjacency tensor of hypergraph H m . When m = 2 , H 2 is just the usual graph that has been widely used in community detection problems [9].
A hypergraph is random if A i 1 i 2 i m is a random variable for i 1 < i 2 < < i m . The binary m-uniform Hypergraph Stochastic Block Model (HSBM) H m , n , p , q is defined as follows: Each node i [ n ] is independently assigned a community label σ i , and randomly with P ( σ i = 1 ) = P ( σ i = 1 ) = 1 2 . We denote σ = ( σ 1 , σ 2 , , σ n ) and I + = I + ( σ ) = { i | σ i = + 1 } , I = I ( σ ) = { i | σ i = 1 } . Then for 1 i 1 < i 2 < < i m n ,
A i 1 , i 2 , , i m B e r n ( p ) , if { i 1 , i 2 , , i m } I + or I , B e r n ( q ) , o . w . .
In addition, A i 1 , i 2 , i m ( i 1 < i 2 < < i m ) are assumed to be independent and conditional on σ . Here, + 1 , 1 represent two communities and I + and I denote the nodes that belong to the + 1 and 1 communities, respectively. The subset { i 1 , i 2 , , i m } forms a hyperedge with a probability p if the distinct nodes i 1 , i 2 , , i m are in the same community. Otherwise, { i 1 , i 2 , , i m } forms a hyperedge with a probability q. Throughout this paper, the community structure is assumed to be balanced; that is, the number of nodes in each community is n 2 , as in [6,22,23]. Moreover, we focus on the case p = a log n n m 1 and q = b log n n m 1 with a b > 0 , since this is the smallest order of hyperedge probability where exact recovery is possible. We denote the binary m-uniform Hypergraph of Stochastic Block Model (HSBM) as H m , n , a , b = H m ( n , a log ( n ) n m 1 , b log ( n ) n m 1 ) .
Community detection refers to the problem of identifying the true label σ based on an observation of a hypergraph A. Let σ ^ be an estimator of σ . We say σ ^ is an exact recovery of σ or σ ^ exactly recovers σ if:
P ( s { ± 1 } : σ ^ = s σ ) = 1 o ( 1 ) .
In words, exact recovery means that the estimated label σ ^ is equal to the true label σ , with a probability convergent to one as the number of nodes goes to infinity. We say exact recovery is possible if there is an estimator σ ^ that exactly recovers σ , and exact recovery is impossible if any estimator σ ^ does not exactly recover σ .
In practice, along with the hypergraph A, side information about node labels is usually available [23,24,25,26,27,28,29,30]. For example, in a co-authorship or co-citation network, the cluster labels of some authors are known [28]. In a Web query network, some queries are labeled [30]. In student relational networks, the dorms in which students live can serve as label information [29]. Various algorithms have been developed to incorporate label information in community recovery in hypergraphs [28,29,30], and incorporating side information has been shown to improve clustering performance [23,24,25,26,27,28,29,30]. The sharp recovery boundary with a label or side information were given by [23,27] in the graph case. However, to the best of our knowledge, the sharp recovery boundary for hypergraphs is still unknown. In this paper, we study the effect of label information on the boundary of exact recovery for hypergraphs and consider two types of label information: ( 1 ) a noisy label for each node is observed independently, with 1 α n as the probability that the noisy label will match the true label; ( 2 ) the true label of each node is observed independently, with the probability of 1 α n . Let α n = n β + o ( 1 ) with a constant β 0 . From an information-theoretical point of view, we derive sharp boundaries of exact recovery in terms of m , a , b , β . Interestingly, label information is useful if and only if β > 0 . The main result is summarized in Table 1, where η m , a , b ( β ) , C m , a , b are defined in Equations (1) and (2). In both cases, for a fixed m, the region (in terms of a , b ) where exact recovery is impossible shrinks as β gets larger. The label information is helpful if and and only if β > 0 ; that is, α n has to converge to zero at the rate of n β for β > 0 . The visualization of the regions in Table 1 can be found in Figures 1 and 2 in Section 2.

2. Main Result

In this section, we consider community detection in hypergraphs through an observation of noisy labels or a proportion of the true labels from an information-theoretical point of view. We derive sharp boundaries of exact recovery, which provides a benchmark for developing practical community detection algorithms.

2.1. Detection with Noisy Label Information

In this subsection, we consider community detection in hypergraphs through a noisy version of node labels available. In the graph regime, community detection with noisy label information was proposed in [27] and extensively studied in [23]. Here, we focus on an m-uniform hypergraph with an arbitrary fixed m 2 . Given a true label vector σ = ( σ 1 , , σ n ) , for each node i, a noisy label Y i is independently observed and Y i coincides with the true label σ i with a probability of 1 α n . More specifically, P ( Y i = σ i | σ i ) = 1 α n and P ( Y i = σ i | σ i ) = α n , α n [ 0 , 1 2 ] and Y i ( 1 i n ) are independent and conditional on σ . If α n = 0 , the true label for each node is fully known. If α n = 1 2 , the noisy labels Y = ( Y 1 , Y 2 , , Y n ) do not provide any information about the true labels. The hypergraph A and Y are assumed to be independent and conditional on σ . In this subsection, we focus on the effect of a noisy label Y on community detection in a hypergraph.
Assume α n = n β + o ( 1 ) with a constant β 0 and define the following:
η m , a , b ( β ) = 1 2 m 1 ( m 1 ) ! a + b γ m , a , b ( β ) C m , a , b + β 2 C m , a , b log γ m , a , b ( β ) + β γ m , a , b ( β ) β + β 2 ,
where
C m , a , b = log ( a ) log ( b ) 2 m 1 ( m 1 ) ! , γ m , a , b ( β ) = β 2 + 4 a b C m , a , b 2 .
Here, C m , a , b and γ m , a , b are defined just for notation convenience without any practical meaning. The quantity η m , a , b ( β ) can be considered the signal contained in the model. The larger η m , a , b ( β ) is, the easier exact recovery becomes. It is clearer to see this in the special case of β = 0 :
η m , a , b ( 0 ) = a b 2 2 m 1 ( m 1 ) ! .
For a fixed m, a large η m , a , b ( 0 ) implies that the difference of a b is large. The within-community nodes are also more densely connected than between-community nodes. Hence, it gets easier to cluster the nodes into groups. Note that η m , a , b ( 0 ) was used to characterize the sharp detection boundary in [6]. For an arbitrary β 0 , we provide the necessary and sufficient conditions for the exact recovery of the community structure. To this end, we firstly investigate the maximum likelihood estimator (MLE) of true labels. The region where exact recovery is impossible corresponds to the region where MLE fails. Then, based on the noisy labels, we construct an estimator that exactly recovers the community structure. The result is summarized in the following theorem.
Theorem 1.
Assume α n = n β + o ( 1 ) with a constant β 0 , then exact recovery in HSBM H m , n , a , b is impossible if
η m , a , b ( β ) < 1 , when β < C m , a , b ( a b ) , β < 1 , when β > C m , a , b ( a b ) .
Exact recovery is possible if
η m , a , b ( β ) > 1 , when β < C m , a , b ( a b ) , β > 1 , when β > C m , a , b ( a b ) .
Here, η m , a , b ( β ) and C m , a , b are defined in Equations (1) and (2).
Based on Theorem 1, there is a phase transition phenomenon for exact recovery in H m , n , a , b . In the region β < C m , a , b ( a b ) , exact recovery is possible if η m , a , b ( β ) > 1 , and not possible if η m , a , b ( β ) < 1 . In the region β > C m , a , b ( a b ) , exact recovery is possible if β > 1 , and exact recovery is impossible if β < 1 . In this sense, phase transition occurs at 1, and 1 is the sharp boundary for exact recovery. When α n is bounded away from zero, β = 0 and C m , a , b ( a b ) > 0 trivially hold. Then, Theorem 1 recovers Theorem 4 and Theorem 5 in [6]. This shows that the noisy label is useful if and only if α n converges to zero at a rate of n β for β > 0 . Furthermore, exact recovery with a fixed m , a , b becomes easier as β increases, since η m , a , b ( β ) is increasing in β (see Lemma A1). When m = 2 , Theorem 1 is reduced to Theorem 1 and Theorem 2 in [23]. Note that C m , a , b is decreasing in m. Then, given a fixed β , the region of exact recovery for m = 2 contains the region of m 3 as a proper subset. These findings can be summarized in Figure 1, where we visualize the regions characterized by (4) and (5) with m = 2 , 3 and β = 0 , 0.4 , 0.8 . In Figure 1, the red (green) region represents exact recovery as impossible (possible). We point out that the time complexity of our estimator for exact recovery is O ( n m ) . Since our focus in this paper is to derive the sharp boundary of exact recovery as in [23,27] and not to propose algorithms with optimal performance, our estimator may or may not outperform existing algorithms.

2.2. Detection with Partially Observed Labels

In this subsection, we consider the community detection problem with true labels partially observed. This type of side information was considered in [23,24,25,26] in the context of graphs, and in [28,29,30] for hypergraphs. Here, we focus on an m-uniform hypergraph with an arbitrary fixed m 2 . Given the true labels σ = ( σ 1 , , σ n ) , for each node i, the true label is independently observed with a probability of 1 α n . More specifically, we define a random variable Y i with P ( Y i = σ i | σ i ) = 1 α n and P ( Y i = 0 | σ i ) = α n , α n [ 0 , 1 ] . Here, Y i = 0 indicates that the true label for node i is not observed. If α n = 1 , no label information is observed. If α n = 0 , all the true labels are observed and community detection is not necessary. We study how α n changes the sharp detection boundary from the information-theoretical point of view. To this end, we investigate the maximum likelihood estimator (MLE) of true labels. The region where exact recovery is impossible corresponds to the region where MLE fails. The exact recovery estimator is constructed based on the partially observed labels. The result is summarized in the following theorem.
Theorem 2.
Assume α n = n β + o ( 1 ) with a constant β 0 , then exact recovery in HSBM H m , n , a , b is impossible if
a b 2 2 m 1 ( m 1 ) ! + β < 1 .
Exact recovery is possible if
a b 2 2 m 1 ( m 1 ) ! + β > 1 .
Theorem 2 clearly characterizes how the partially observed labels affect the boundary for exact recovery. A phase transition phenomenon of exact recovery exists at 1, since exact recovery is possible if η m , a , b ( 0 ) + β > 1 , but not possible if η m , a , b ( 0 ) + β < 1 . When β = 0 , Theorem 2 recovers Theorem 4 and Theorem 5 in [6]. If β > 0 , the region (6) where exact recovery is impossible is smaller than that in [6]. The side information of partially known labels makes exact recovery easier if and only if β > 0 . When m = 2 , Theorem 2 is reduced to Theorem 1 and Theorem 2 in [23]. For a fixed β , the region of exact recovery of m 3 is smaller than that of m = 2 . These findings can be verified in Figure 2, where we visualize the regions characterized by (6) and (7) with m = 2 , 3 and β = 0 , 0.4 , 0.8 . In Figure 2, the red (green) region represents exact recovery as impossible (possible). Finally, we point out that the time complexity of our estimator achieving exact recovery is O ( n m ) . Again, our focus in this paper is to derive the sharp boundary of exact recovery as in [23,27], not to propose algorithms with optimal performance; hence, our estimator may or may not outperform existing algorithms.

3. Proof of Main Result

In this section, we provide detailed proof of Theorems 1 and 2.
To start with, we derive the explicit expression of the likelihood function. The likelihood function of hypergraph A given the node label σ is:
P ( A | σ ) = 1 i 1 < < i m n p A i 1 , , i m ( 1 p ) 1 A i 1 , , i m I [ σ i 1 = = σ i m ] 1 i 1 < < i m n q A i 1 , , i m ( 1 q ) 1 A i 1 , , i m 1 I [ σ i 1 = = σ i m ] = 1 i 1 < < i m n p ( 1 q ) q ( 1 p ) A i 1 , , i m I [ σ i 1 = = σ i m ] q 1 q A i 1 , , i m 1 p 1 q I [ σ i 1 = = σ i m ] ( 1 q ) .
Then, the log-likelihood function can be written as follows:
log ( P ( A | σ ) ) = log p ( 1 q ) q ( 1 p ) 1 i 1 < < i m n A i 1 , , i m I [ σ i 1 = = σ i m ] + log q 1 q 1 i 1 < < i m n A i 1 , , i m + log 1 p 1 q 1 i 1 < < i m n I [ σ i 1 = = σ i m ] + log ( 1 q ) 1 i 1 < < i m n 1 : = I + II ,
where
I = log p ( 1 q ) q ( 1 p ) 1 i 1 < < i m n A i 1 , , i m I [ σ i 1 = = σ i m ] = log p q + log 1 q 1 p 1 i 1 < < i m n A i 1 , , i m I [ σ i 1 = = σ i m ] : = C a , b + o ( 1 ) e + + e C a , b e + + e , C a , b = log ( a ) log ( b ) , e + = 1 i 1 < < i m | I + | A i 1 , , i m I [ σ i 1 = = σ i m = + 1 ] , e = 1 i 1 < < i m | I | A i 1 , , i m I [ σ i 1 = = σ i m = 1 ] ,
and
II = log q 1 q 1 i 1 < < i m n A i 1 , , i m + log 1 p 1 q 1 i 1 < < i m n I [ σ i 1 = = σ i m ] + log ( 1 q ) 1 i 1 < < i m n 1 = e [ n ] log q 1 q + 2 n 2 m log 1 p 1 q + n m log ( 1 q ) e [ n ] = 1 i 1 < < i m n A i 1 , , i m .
Note that II is independent of σ .

Proof of Theorem 1

The likelihood function of the vector of noisy labels Y given node lable σ is the following:
P ( Y | σ ) = i = 1 n ( 1 α n ) I [ y i σ i = + 1 ] α n I [ y i σ i = 1 ] = i = 1 n ( 1 α n ) I [ y i σ i = + 1 ] α n 1 I [ y i σ i = + 1 ]
Then, the log-likelihood function can be written as follows:
log ( P ( Y | σ ) ) = log 1 α n α n i = 1 n I [ y i σ i = 1 ] + log ( α n ) i = 1 n 1 : = I s + II s ,
where
I s = log 1 α n α n i = 1 n I [ y i σ i = 1 ] = log 1 α n + log 1 α n i = 1 n I [ y i σ i = 1 ] C α n ( s + + s ) , C α n = log 1 α n α n , s + = i = 1 n I [ y i = + 1 , σ i = + 1 ] , s = i = 1 n I [ y i = 1 , σ i = 1 ] ,
and II s = n log ( α n ) .
Noting that A and Y are independent given σ , the joint log-likelihood of A , Y given σ  is the following:
log ( P ( A , Y | σ ) ) = log ( P ( A | σ ) ) + log ( P ( Y | σ ) ) = I + II + I s + II s C a , b ( e + + e ) + C α n ( s + + s ) + II + II s .
where C a , b = log ( a ) log ( b ) , C α n = log ( 1 α n ) log ( α n ) , and II + II s consist of the terms that are independent of σ .
Denote e S 1 , S 2 as the number of edges between two sets of nodes, say S 1 and S 2 . Then, define the following events as follows:
F = { maximum likelihood fails } , F i , + = { i I + : e i , I e i , I + + C α n C a , b y i + 1 } , F i , = { i I : e i , I + e i , I C α n C a , b y i + 1 } , F + = i I + F i , + , F = i I F i , .
Lemma 1.
P ( F ) = 1 o ( 1 ) , if F + and F .
Proof. 
Take i 0 F + and j 0 F . Then, define the following:
I ˜ + = I + \ { i 0 } { j 0 } , I ˜ = I \ { j 0 } { i 0 } .
Denote Λ = P ( A , Y | σ ˜ ) P ( A , Y | σ ) , and we need to show that P ( Λ > 1 ) = 1 o ( 1 ) .
By (8), we have the following:
log ( Λ ) C a , b ( e I ˜ + + e I ˜ ) ( e + + e ) + C α n ( s I ˜ + + s I ˜ ) ( s + + s ) .
It is clear that
e I ˜ + e + = e i 0 , I + + e j 0 , I + e j 0 , I + \ { i 0 } , e I ˜ e = e j 0 , I + e i 0 , I e i 0 , I \ { j 0 } , s I ˜ + s + = I [ y i 0 = + 1 ] + I [ y j 0 = + 1 ] , s I ˜ s = I [ y j 0 = 1 ] + I [ y i 0 = 1 ] .
Plugging into (10) yields
log ( Λ ) C a , b ( e j 0 , I + e j 0 , I ) ( e i 0 , I + e i 0 , I ) e j 0 , I + \ { i 0 } e i 0 , I \ { j 0 } + C α n { y j 0 y i 0 } = C a , b ( e j 0 , I + e j 0 , I ) + C α n y j 0 + C a , b ( e i 0 , I e i 0 , I + ) C α n y i 0 C a , b e i 0 , I \ { j 0 } C a , b e j 0 , I + \ { i 0 } 2 C a , b ( 1 e i 0 , I \ { j 0 } ) .
In the last inequality, we assumed that e i 0 , I \ { j 0 } e j 0 , I + \ { i 0 } without loss of generality.
Next, we will show that E ( e i 0 , I \ { j 0 } ) = o ( 1 ) . Rewrite e i 0 , I \ { j 0 } as:
e i 0 , I \ { j 0 } = 1 i 3 < < i m n 2 1 { i 3 , , i m } I \ { j 0 } A i 0 j 0 i 3 i m = k = 1 n 2 1 m 2 Z k ,
where Z k i . i . d . Z = B e r n ( q ) and q = b log ( n ) n m 1 . Then,
E ( e i 0 , I \ { j 0 } ) = n 2 1 m 2 b log ( n ) n m 1 = log ( n ) n O ( 1 ) = o ( 1 ) .
Applying Markov’s inequality, we have
P ( Λ > 1 ) = P ( log ( Λ ) > 0 ) P ( 2 C a , b ( 1 e i 0 , I \ { j 0 } ) > 0 ) = P ( e i 0 , I \ { j 0 } < 1 ) = 1 P ( e i 0 , I \ { j 0 } 1 ) 1 E ( e i 0 , I \ { j 0 } ) = 1 o ( 1 ) ,
which complete the proof.    □
A. Proof of the impossible part of Theorem 1
Let H be a fixed subset of I + of size | H | = n log τ ( n ) , with a constant τ > 1 m 1 and θ n = log ( n ) log log ( n ) . For any i H , define the following events:
Δ i , H = { e i , H θ n } , F i , H = { e i , I e i , I + \ H + θ n + C α n C a , b y i + 1 } , Δ H = i H Δ i , H , F H = i H F i , H .
Lemma 2.
P ( Δ H ) = o ( 1 ) .
Proof. 
Let W k i . i . d . B e r n ( p ) . By definition, for any i H ,
e i , H = 1 i 2 < < i m | H | 1 { i 2 , , i m } H \ { i } A i i 2 i m = k = 1 | H | 1 m 1 W k , E ( e i , H ) = | H | 1 m 1 a log ( n ) n m 1 .
Then, the multiplicative Chernoff bound (see ( i i i ) in Lemma A1) gives the following:
P ( Δ i , H ) = P e i , H θ n = P k = 1 | H | 1 m 1 W k θ n n m 1 log log ( n ) e a | H | 1 m 1 θ n .
By union bound we have the following:
P ( Δ H ) | H | n m 1 log log ( n ) e a | H | 1 m 1 θ n = n 1 ( m 1 ) τ + o ( 1 ) = o ( 1 ) ,
since τ > 1 m 1 by the assumption on | H | .    □
Lemma 3.
P ( F i , H ) 1 | H | log ( 1 δ ) under the condition in (4), for any δ ( 0 , 1 ) .
Proof. 
See Appendices Appendix A.1 and Appendix A.2.    □
Lemma 4.
P ( F H ) = 1 o ( 1 ) under the condition in (4).
Proof. 
Under the condition in (4), Lemma 3 holds. That is,
P ( F i , H ) > 1 | H | log 1 δ
for any δ ( 0 , 1 ) and for a sufficiently large n. By union bound, we have the following:
P ( F H ) = P ( i H F i , H ) = 1 P ( i H ( F i , H ) c ) = 1 ( 1 P ( F i , H ) ) | H | > 1 ( 1 P ( F i , H ) ) log ( δ ) 1 P ( F i , H ) = 1 δ 1 P ( F i , H ) log ( 1 P ( F i , H ) ) 1 δ .
Here, we used (11) in the first inequality.    □
Lemma 5.
P ( F ) = 1 o ( 1 ) under the condition in (4).
Proof. 
By Lemmas 2 and 4, there exists a sufficiently small δ > 0 , such that P ( Δ c ) 1 δ and P ( F H ) 1 δ simultaneously hold. Clearly, Δ c F H F + . Hence,
P ( F + ) P ( Δ c ) + P ( F H ) 1 1 2 δ .
By symmetry, we have
P ( F ) 1 2 δ .
Then, it follows from Lemma 1 that
P ( F ) P ( F I + ) + P ( F I ) 1 1 4 δ ,
which completes the proof.    □
The impossible part of Theorem 1 follows directly from Lemma 5.
B. Proof of the possible part of Theorem 1
Let H 1 be an independently generated random hypergraph, built on the same node set of V = [ n ] , with an edge probability of d n log ( n ) . Denote its complement by H 2 = H 1 c and
H 1 = H H 1 , H 2 = H H 2 .
A weak recovery algorithm [12] is applied to H 1 to return a partition of two communities I ˜ + and I ˜ , which agree with the true communities I + and I on at least ( 1 δ ) n nodes. Here, δ = δ ( d n ) depends on d n such that δ 0 as d n . d n can be taken as O ( log log ( n ) ) [12]. In the next step, we will use H 2 to decide whether to flip a node’s membership or not. More specifically, for a node i I ˜ + , if it has more edges in H 2 going to I ˜ , plus the scaled side information C α n C a , b y i , then we reset i I ˜ . Similarly, for i I ˜ , if it has more edges in H 2 going to I ˜ + , minus the scaled side information C α n C a , b y i , then we reset i I ˜ + . If the number of flips in each community is not the same, then keep the discard change. This algorithm has been summarized in Algorithm 1.
Algorithm 1: Algorithm for exact recovery of community structure in hypergraphs with label information.
1. Input: Hypergraph H and label information y;
2. Partition H = H 1 H 2 , where H 1 = H H 1 , H 2 = H H 1 c , and H 1 make an
    Erd o ¨ s–Renyi graph generated with an edge probability of d n log ( n ) ;
3. Apply weak recovery algorithm [12] to H 1 to return a partition I + , 0 I , 0 ;
4. Initialize I ˜ + I + , 0 , and I ˜ I , 0 ;
5. Flip membership if
      i I ˜ + and e i , I ˜ e i , I ˜ + + C α n C a , b y i in H 2 , or
      i I ˜ and e i , I ˜ + e i , I ˜ C α n C a , b y i in H 2 ;
6. If | I ˜ + | | I + , 0 | , then keep I + , 0 and I , 0 unchanged.
To show (5), the possible part of Theorem 1, we first introduce the following definitions. For any node i [ n ] , it is mis-classified if and only if it belongs to the following:
i I ˜ + : σ i = 1 i I ˜ : σ i = + 1 .
With WLOG, we assume that i I ˜ + . Then, the mis-classification probability of node i is given by the following:
M i = P e i , I ˜ e i , I ˜ + + C α n C a , b y i = P k = 1 n 2 m 1 n 2 δ m 1 Z k + k = 1 n 2 δ m 1 W k k = 1 n 2 m 1 n 2 δ m 1 W k + k = 1 n 2 δ m 1 Z k + C α n C a , b y i .
In the last equation, we assumed H 2 to be a complete graph. Then, the probability of the existence of a mis-classified label is:
M n M i ,
by union bound on all nodes.
Consider a node i H 1 . Its degree is given by the following:
deg ( i ) = 1 i 2 < < i m n 1 { i 2 , , i m } [ n ] \ { i } A i i 2 i m .
Lemma 6.
max i H 1 { deg ( i ) } 2 n 1 m 1 d n log ( n ) .
Proof. 
Note that deg ( i ) B i n o m n 1 m 1 , d n log ( n ) . Then, the multiplicative Chernoff bound (see ( i i i ) in Lemma A1) gives the following:
P deg ( i ) 2 μ e 4 μ < e 1 4 μ ,
where μ = n 1 m 1 d n log ( n ) . By union bound, we have the following:
P max i H 1 { deg ( i ) } 2 μ n P deg ( i ) 2 μ < n e 1 4 μ = o ( 1 ) .
The Lemma above suggests that:
min i H 2 { deg ( i ) } n 1 m 1 1 2 d n log ( n ) .
Therefore, taking into account the incompleteness of H 2 , we will loose the upper bound (12) by removing 2 n 1 m 1 d n log ( n ) terms from both of the summations on the right-hand side of (12). That is,
M i = P k = 1 n 2 m 1 n 2 δ m 1 Z k + k = 1 n 2 δ m 1 W k k = 1 n 2 m 1 n 2 δ m 1 2 n 1 m 1 d n log ( n ) W k + k = 1 n 2 δ m 1 2 n 1 m 1 d n log ( n ) Z k + C α n C a , b y i .
Lemma 7.
M n 1 η m , a , b ( β ) + o ( 1 ) , β < C m , a , b ( a b ) , n 1 β + o ( 1 ) , β > C m , a , b ( a b ) .
Proof. 
See Appendix A.3. □
The possible part of Theorem 1 follows directly from Lemma 7.

4. Conclusions

In this paper, we studied the effect of label information on the exact recovery of communities in uniform hypergraphs from an information-theoretical point of view. Specifically, we considered two types of label information: a noisy label for each node was observed independently, with 1 α n as the probability that the noisy label would match the true label, and the true label of each node was observed independently, with a probability of 1 α n . We used the maximum likelihood method to derive a lower bound for exact recovery and then constructed an estimator that could exactly recover the communities above the lower bound. In this way, we obtained sharp boundaries for exact recovery under both scenarios. We found that the label information improved the sharp detection boundary if and only if α n converges´d to zero at a rate of n β for some positive constant β .
There are several possible future research directions: (I) The sharp recovery boundary for general HSBM with label information is still unknown. Characterizing the boundary in this case is an important problem. (II) In this paper, we focused on the label information. It is important to consider other side information, such as the covariates observed for each node.

Author Contributions

Methodology: X.Z., W.Z. and M.Y.; writing—original draft preparation: W.Z., M.Y.; writing—review and editing: W.Z. and X.Z.; supervision: M.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Acknowledgments

The authors are grateful to Editor and reivewers for helpful comments that significantly improve this manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Appendix A.1. Chernoff Bound

For a random variable X, denote its cumulant generating function (cgf) by ψ X ( t ) = log ( E ( e t X ) ) . Define
ϕ X , D ( t ) = t D ψ X ( t ) , ϕ X , D ( t ) = t D ψ X ( t ) ,
for any fixed D R , where t ranges over R or R + .
Lemma A1.
Assume X i i . i . d . X with cgf ψ X ( t ) , i = 1 , , n .
(i) 
(Lemma 15 in [23]) For any D , ϵ R ,
P 1 n i = 1 n X i D ϵ e n ϕ X , D ( t max ) + | t max | ϵ 1 σ X ^ 2 n ϵ 2
where t max = arg sup t R ( ϕ X , D ( t ) ) , X ^ is a random variable with the same alphabet as X but distributed according to e t max x P ( x ) E X ( e t max X ) , and μ X ^ , σ X ^ 2 are the mean and variance of X ^ , respectively.
(ii) 
(Generic Chernoff bound ) For any D R ,
P 1 n i = 1 n X i D e n ϕ X , D ( t max )
where t max = arg sup t > 0 ( ϕ X , D ( t ) ) , and
P 1 n i = 1 n X i D e n ϕ X , D ( t max )
where t max = arg sup t > 0 ( ϕ X , D ( t ) ) .
(iii) 
(Multiplicative Chernoff bound ) For any t > 1 ,
P i = 1 n X i t μ t e t μ e μ
where μ = E ( X ) .
For later use, we consider X = C a , b ( Z W ) , where Z B e r n ( q ) , W B e r n ( p ) with p = a log ( n ) n m 1 , p = b log ( n ) n m 1 , and C a , b = log ( a ) log ( b ) , a b > 0 . Define
η m , a , b ( Δ ) = 1 2 m 1 ( m 1 ) ! a + b γ m , a , b ( Δ ) C m , a , b + Δ 2 C m , a , b log γ m , a , b ( Δ ) + Δ γ m , a , b ( Δ ) Δ + Δ 2 , γ m , a , b ( Δ ) = Δ 2 + 4 a b C m , a , b 2 ,
where C m , a , b = log ( a ) log ( b ) 2 m 1 ( m 1 ) ! and Δ R is a constant. In the special case Δ = 0 , we have
η m , a , b ( 0 ) = a b 2 2 m 1 ( m 1 ) ! .
Lemma A2.
(i) 
(Lower bound) Assume D m , n = log ( n ) l m , n ( Δ + o ( 1 ) ) with a constant Δ R , and ϵ m , n = log ( n ) l m , n o ( 1 ) , then
P 1 l m , n i = 1 l m , n X i D m , n + ϵ m , n n η m , a , b ( Δ ) + o ( 1 ) ,
where l m , n = n 2 m 1 , and η m , a , b ( Δ ) is given in (A2).
(ii) 
(Upper bound) Assume D m , n = log ( n ) l m , n ( Δ + o ( 1 ) ) with Δ R , then
P 1 l m , n i = 1 l m , n X i D m , n n η m , a , b ( Δ ) + o ( 1 ) ,
when Δ > C m , a , b ( a b ) , and
P 1 l m , n i = 1 l m , n X i D m , n n η m , a , b ( Δ ) + o ( 1 ) ,
when Δ < C m , a , b ( a b ) . Here, l m , n and η m , a , b ( Δ ) are defined as in ( i ) .
Proof. 
We first calculate and approximate ϕ X ( t ) and ψ X ( t ) . Denote s = ( a b ) t . Then direct calculation gives the following:
ψ X ( t ) = log E ( e t C a , b ( Z W ) ) = log E ( s Z W ) = log ( 1 q ( 1 s ) ) + log ( 1 p ( 1 s 1 ) ) .
Define
ψ ˜ X ( t ) : = log ( n ) n m 1 a + b b s a s 1 = ψ X ( t ) + log ( n ) n m 1 o ( 1 ) .
Taking D = D m , n = log ( n ) l m , n ( Δ + o ( 1 ) ) in the first equation of (A1), we have
ϕ ˜ X , D ( t ) = log ( n ) n m 1 Δ log ( s ) C m , a , b + a + b b s a s 1 + o ( 1 ) ,
where l m , n = n 2 m 1 .
Taking the first derivative of ϕ ˜ X , D ( t ) , w.r.t. t yields
ϕ ˜ X , D ( t ) = log ( n ) l m , n Δ b C m , a , b s + a C m , a , b s 1 .
Set ϕ ˜ X , D ( t ) = 0 , and solve
0 = b C m , a , b s 2 Δ s a C m , a , b ,
we have
s * = γ m , a , b ( Δ ) + Δ 2 b C m , a , b ,
where
γ m , a , b ( Δ ) = Δ 2 + 4 a b C m , a , b 2 .
Noting that
log ( s * ) = log γ m , a , b ( Δ ) + Δ 2 b C m , a , b = 1 2 log γ m , a , b ( Δ ) + Δ γ m , a , b ( Δ ) Δ + C m , a , b , b s * + a s * 1 = γ m , a , b ( Δ ) C m , a , b .
Plugging in (A6) yields
e l m , n ϕ ˜ X ( t * ) = n l m , n n m 1 a + b γ m , a , b ( Δ ) C m , a , b + Δ 2 C m , a , b log γ m , a , b ( Δ ) + Δ γ m , a , b ( Δ ) Δ + Δ 2 + o ( 1 ) = n η m , a , b ( Δ ) + o ( 1 ) .
where η m , a , b ( Δ ) and γ m , a , b ( Δ ) are given in (A2).
(1) For part ( i ) , note that s max = s * is the global maximum of ϕ ˜ X , D ( t ) on R , since ϕ ˜ X , D ( t ) < 0 . Therefore, the first part is completed by applying the Chernoff bound (see ( i ) in Lemma A1).
(2) To show (A4) in ( i i ) , we get from t > 0 that
s max = max { s * , 1 } = 1 , if Δ < C m , a , b ( a b ) , s * , if Δ > C m , a , b ( a b ) ,
which is the global maximum of ϕ ˜ X , D ( t ) on ( 0 , ) , since ϕ ˜ X , D ( t ) < 0 .
If Δ < C m , a , b ( a b ) , then s max = 1 leads to a trivial bound since in this case
e l m , n ϕ X , D ( t max = 0 ) = 1 .
If Δ > C m , a , b ( a b ) , then s max = s * and (A6) hold. This completes the proof of (A4) by applying the Chernoff bound (see ( i i ) in Lemma A1).
Now, we are left to show (A5) in part ( i i ) . Consider the random variable X , we have
ψ X ( t ) = log ( 1 p ( 1 s ) ) + log ( 1 q ( 1 s 1 ) ) ,
Define
ψ ˜ X ( t ) : = log ( n ) n m 1 a + b a s b s 1 = ψ ˜ X ( t ) + log ( n ) n m 1 o ( 1 ) .
Taking D = D m , n = log ( n ) l m , n ( Δ + o ( 1 ) ) in the second equation of (A1), we have
ϕ ˜ X , D ( t ) = log ( n ) n m 1 Δ log ( s ) C m , a , b + a + b a s b s 1 ,
where Δ R and C m a , b is defined as above.
Taking the first derivative of ϕ ˜ X , D ( t ) , w.r.t. t yields
ϕ ˜ X , D ( t ) = log ( n ) l m , n Δ a C m , a , b s + b C m , a , b s 1
Set ϕ ˜ X , D ( t ) = 0 , and solve
0 = a C m , a , b s 2 + Δ s b C m , a , b ,
we have
s * = γ m , a , b ( Δ ) Δ 2 a C m , a , b ,
where
γ m , a , b ( Δ ) = Δ 2 + 4 a b C m , a , b 2 .
Noting that t > 0 , we have
s max = max { s * , 1 } = 1 , if Δ > C m , a , b ( a b ) , s * , if Δ < C m , a , b ( a b ) ,
which is the global maximum of ϕ ˜ X , D ( t ) on ( 0 , ) , since ϕ ˜ X , D ( t ) < 0 .
If Δ > C m , a , b ( a b ) , then s max = 1 leads to a trivial bound since
e l m , n ϕ X , D ( t max = 0 ) = 1 .
If Δ < C m , a , b ( a b ) , then s max = s * . Noting that
log ( s * ) = log γ m , a , b ( Δ ) Δ 2 a C m , a , b = 1 2 log γ m , a , b ( Δ ) + Δ γ m , a , b ( Δ ) Δ + C a , b , a s * + b s * 1 = γ m , a , b ( Δ ) C m , a , b .
Then
e l m , n ϕ ˜ X ( t * ) = n l m , n n m 1 a + b γ m , a , b ( Δ ) C m , a , b + Δ 2 C m , a , b log γ m , a , b ( Δ ) + Δ γ m , a , b ( Δ ) Δ + Δ 2 + o ( 1 ) = n η m , a , b ( Δ ) + o ( 1 ) .
This completes the proof of (A5) by applying the Chernoff bound (see ( i i ) in Lemma A1).

Appendix A.2. Proof of Lemma 3

Denote l m , n = n 2 m 1 and recall that θ n = log ( n ) log log ( n ) . Then
P ( F i , H ) = P k = 1 | I | m 1 Z k k = 1 | I + | | H | m 1 W k + C α n C a , b y i + 1 + θ n P k = 1 l m , n ( Z k W k ) C α n C a , b y i + 1 + θ n = P k = 1 l m , n X k C α n y i + C a , b + C a , b θ n : = P 1 l m , n k = 1 l m , n X k D ϵ m , n , y i ϵ m , n ,
where
D ϵ m , n , y i = C a , b l m , n C α n C a , b y i + 1 + θ n + ϵ m , n , ϵ m , n = log ( n ) n m 1 o ( 1 ) .
It follows that Δ = β y i , since
Δ β , y i = l m , n log ( n ) D ϵ m , n , y i = β y i + o ( 1 ) .
Here, we used C α n = ( β + o ( 1 ) ) log ( n ) . Therefore,
  • y i = + 1 . then η m , a , b ( Δ ) = η m , a , b ( β ) as in (1);
  • y i = 1 , then η m , a , b ( Δ ) = η m , a , b ( β ) β .
By (A3) in Lemma A2, (A6) becomes
P ( F i , H ) ( 1 α n ) n η m , a , b ( β ) + o ( 1 ) + α n n η m , a , b ( β ) + β + o ( 1 ) = ( 1 α n ) n η m , a , b ( β ) + o ( 1 ) + n η m , a , b ( β ) + o ( 1 ) = n η m , a , b ( β ) + o ( 1 )
Therefore, if η m , a , b ( β ) 1 ε for some 0 < ε < 1 , then
P ( F i , H ) n 1 ε > 1 | H | log 1 δ
holds for δ ( 0 , 1 ) and a sufficiently large n, since | H | = n log τ ( n ) for τ > 1 m 1 . This completes the first part under condition (4), the impossible part of Theorem 1.
For the second part in (4), recall that β > C m , a , b ( a b ) . Then
P ( F i , H ) = P k = 1 | I | m 1 Z k k = 1 | I + | | H | m 1 W k + C α n C a , b y i + 1 + θ n P k = 1 l m , n ( Z k W k ) C α n C a , b y i + 1 + θ n = 1 P k = 1 l m , n X k C α n y i + C a , b + C a , b θ n : = 1 P 1 l m , n k = 1 l m , n X k D ϵ m , n , y i ,
where
D ϵ m , n , y i = C a , b l m , n C α n C a , b y i + 1 + θ n .
It follows that Δ = β y i , since
Δ β , y i = l m , n log ( n ) D ϵ m , n , y i = β y i + o ( 1 ) .
Here we used C α n = ( β + o ( 1 ) ) log ( n ) . We still have
  • y i = + 1 , then η m , a , b ( Δ ) = η m , a , b ( β ) . Note that C m , a , b ( a b ) > Δ = β always holds, the probability in (A5) is upper-bounded by 1;
  • y i = 1 , then η m , a , b ( Δ ) = η m , a , b ( β ) β . Note that C m , a , b ( a b ) < Δ = β by assumption, and (A5) applies.
Then (A7) becomes
P ( F i , H ) 1 ( 1 α n ) α n n η m , a , b ( β ) + β + o ( 1 ) = α n n η m , a , b ( β ) + o ( 1 ) = n β n η m , a , b ( β ) + o ( 1 )
Therefore, if β 1 ε 1 and η m , a , b ( β ) 1 + ε 2 for two sufficiently small positive constants ε 1 and ε 2 , then
P ( F i , H ) n 1 ( n ε 1 n ε 2 ) > 1 | H | log 1 δ
holds for δ ( 0 , 1 ) and a sufficiently large n, since | H | = n log τ ( n ) for τ > 1 m 1 . This completes the second part under condition (4), the impossible part of Theorem 1.

Appendix A.3. Proof of Lemma 7

M i = P k = 1 n 2 m 1 n 2 δ m 1 Z k + k = 1 n 2 δ m 1 W k k = 1 n 2 m 1 n 2 δ m 1 2 n 1 m 1 d n log ( n ) W k + k = 1 n 2 δ m 1 2 n 1 m 1 d n log ( n ) Z k + C α n C a , b y i P k = 1 n 2 m 1 Z k + k = 1 n 2 δ m 1 W k k = 1 n 2 m 1 n 2 δ m 1 2 n 1 m 1 d n log ( n ) W k C α n C a , b y i P k = 1 n 2 m 1 Z k k = 1 n 2 m 1 W k + k = 1 2 n 2 δ m 1 + 2 n 1 m 1 d n log ( n ) W k C α n C a , b y i
Defining λ = 1 log ( δ ) . Then
M i P k = 1 n 2 m 1 ( Z k W k ) C α n C a , b y i λ log ( n ) + P k = 1 2 n 2 δ m 1 + 2 n 1 m 1 d n log ( n ) W k λ log ( n ) : = I + II .
For II , the multiplicative Chernoff bound (see ( i i i ) in Lemma A1) gives
II λ log ( n ) e n m 1 a log ( n ) 1 2 n 2 δ m 1 + 2 n 1 m 1 d n log ( n ) λ log ( n ) = λ a e n m 1 2 n 2 δ m 1 + 2 n 1 m 1 d n log ( n ) λ log ( n ) = n λ log λ a e 1 2 1 n m 1 n 2 δ m 1 + 2 1 n m 1 n 1 m 1 d n log ( n ) = n λ log 1 δ m 1 1 log 1 + c m d n log ( n ) 1 δ m 1 log 1 δ m 1 + o ( 1 ) = n ( m 1 ) log ( 1 δ ) 1 log 1 + c m d n log ( n ) 1 δ m 1 log 1 δ m 1 + o ( 1 ) n ( m 1 ) ( 1 + Ω ( 1 ) ) ,
where c m depends only on m. We used δ 0 as d n , and log 1 + d n log ( n ) 1 δ m 1 < log 1 δ m 1 for a sufficiently large d n in the last inequality.
For I , we again apply the Chernoff bound (see ( i i ) in Lemma A4).
I = P k = 1 l m , n X k C α n y i C a , b λ log ( n ) = P 1 l m , n k = 1 l m , n X k D λ , y i ,
where
D λ , y i = C a , b l m , n C α n C a , b y i λ log ( n ) .
It follows that Δ = β y i , since
Δ λ , y i = l m , n log ( n ) D ϵ m , n , y i = β y i + o ( 1 ) .
Here, we used C α n = ( β + o ( 1 ) ) log ( n ) . We still have the following:
  • y i = + 1 , then η m , a , b ( Δ ) = η m , a , b ( β ) . Note that C m , a , b ( a b ) > Δ = β always holds, and (A4) applies;
  • y i = 1 , then η m , a , b ( Δ ) = η m , a , b ( β ) β . If C m , a , b ( a b ) > Δ = β , then (A4) applies. If C m , a , b ( a b ) < β , then we take the upper bound to be 1.
To sum up,
(i)
If C m , a , b ( a b ) > β , then (A9) becomes
I ( 1 α n ) n η m , a , b ( β ) + o ( 1 ) + α n n η m , a , b ( β ) + β + o ( 1 ) = ( 1 α n ) n η m , a , b ( β ) + o ( 1 ) + n η m , a , b ( β ) + o ( 1 ) = n η m , a , b ( β ) + o ( 1 ) .
Combining with (A8) yields
M i I + II n η m , a , b ( β ) + o ( 1 ) + n ( m 1 ) + Ω ( 1 ) .
(ii)
If C m , a , b ( a b ) < β , then (A9) becomes
I ( 1 α n ) n η m , a , b ( β ) + o ( 1 ) + α n = ( 1 α n ) n η m , a , b ( β ) + o ( 1 ) + n β + o ( 1 ) = n β + o ( 1 ) .
Here, we used the fact that η m , a , b ( β ) β for any β R (see Lemma A3). Combining with (A8) yields the following:
M i I + II n η m , a , b ( β ) + o ( 1 ) + n ( m 1 ) + Ω ( 1 )
That is,
M i n η m , a , b ( β ) + o ( 1 ) + n ( m 1 ) + Ω ( 1 ) , C m , a , b ( a b ) > β , n β + o ( 1 ) + n ( m 1 ) + Ω ( 1 ) , C m , a , b ( a b ) < β .
By union bound, the probability of failure is
M n M i n 1 η m , a , b ( β ) + o ( 1 ) , C m , a , b ( a b ) > β , n β + o ( 1 ) , C m , a , b ( a b ) < β ,
under condition (5). This complete the proof of Lemma 7.
Denote β m = 2 m 1 ( m 1 ) ! β and rewrite (1) as
η a , b ( β m ) = a + b γ a , b ( β m ) C a , b + β m 2 C a , b log γ a , b ( β m ) + β m γ a , b ( β m ) β m + β m 2 , γ a , b ( β m ) = β m 2 + 4 a b C a , b 2 ,
where C a , b = log ( a ) log ( b ) .
Lemma A3.
η a , b ( β m ) β m .
Proof. 
Let
ξ a , b ( β m ) : = η a , b ( β m ) β m ,
and we will show that it is convex in β m , with a global minimum value of 0. By (1), we have
C a , b ξ a , b ( β m ) = C a , b ( η a , b ( β m ) β m ) = C a , b ( a + b ) γ a , b ( β m ) + β m 2 log γ a , b ( β m ) + β m γ a , b ( β m ) β m C a , b = C a , b ( a + b ) γ a , b ( β m ) β m log γ a , b ( β m ) β m 2 b = C a , b ( a + b ) γ a , b ( β m ) β m log ( γ a , b ( β m ) β m ) + β m log ( 2 b ) .
Here, we used the fact that
log γ a , b ( β m ) β m 2 b = 1 2 log γ a , b ( β m ) + β m γ a , b ( β m ) β m C a , b .
Taking the first two derivatives of ξ a , b ( β m ) w.r.t. β m , and using
γ a , b ( β m ) = γ a , b 1 ( β m ) β m , ( log ( γ a , b ( β m ) β m ) ) = γ a , b 1 ( β m ) ,
we have
C a , b ξ a , b ( β m ) = log γ a , b ( β m ) β m 2 b , C a , b ξ a , b ( β m ) = γ a , b 1 ( β m ) .
Thus, ξ a , b ( β m ) is convex with a unique critical point β m * = C m , a , b ( a b ) . Hence, ξ a , b ( β m ) ξ ( β m * ) = 0 . □
Corollary A1.
η a , b ( β m ) is increasing in β m for any β m 0 .
Proof. 
Taking the first two derivatives of η a , b ( β m ) = ξ a , b ( β m ) + β m w.r.t. β m and using (A11), we have the following:
C a , b η a , b ( β m ) = C a , b ( ξ a , b ( β m ) + 1 ) = log γ a , b ( β m ) β m 2 b + log a b = log γ a , b ( β m ) β m 2 a , C a , b ξ a , b ( β m ) = C a , b η a , b ( β m ) = γ a , b 1 ( β m ) .
Thus, η a , b ( β m ) is convex with a unique critical point β m * = C m , a , b ( a b ) < 0 . Hence, η a , b ( β m ) is increasing in β m for any β m 0 . □

References

  1. Chen, J.; Yuan, B. Detecting functional modules in the yeast proteinprotein interaction network. Bioinformatics 2006, 22, 2283–2290. [Google Scholar] [CrossRef]
  2. Costa, L.F.; Oliveira, O.N., Jr.; Travieso, G.; Rodrigues, F.A.; Villas Boas, P.R.; Antiqueira, L.; Viana, M.P.; Correa Rocha, L.E. Analyzing and modeling real-world phenomena with complex networks: A survey of applications. Adv. Phys. 2011, 60, 329–412. [Google Scholar] [CrossRef] [Green Version]
  3. Fortunato, S. Community detection in graphs. Phys. Rep. 2010, 486, 75–174. [Google Scholar] [CrossRef] [Green Version]
  4. Newman, M.E.J. Coauthorship networks and patterns of scientific collaboration. Proc. Natl. Acad. Sci. USA 2004, 101, 5200–5205. [Google Scholar] [CrossRef] [Green Version]
  5. Ma’ayan, A. Introduction to Network Analysis in Systems Biology. Sci. Signal. 2011, 4, tr5. [Google Scholar] [CrossRef] [Green Version]
  6. Kim, C.; Bandeira, A.; Goemans, M. Stochastic Block Model for Hypergraphs: Statistical limits and a semidefinite programming approach. arXiv 2018, arXiv:1807.02884. [Google Scholar]
  7. Lei, J. A goodness-of-fit test for stochastic block models. Ann. Stat. 2016, 44, 401–424. [Google Scholar] [CrossRef]
  8. Yuan, M.; Nan, Y. Test dense subgraphs in sparse uniform hypergraph. Commun. Stat.-Theory Methods 2020, 1–20. [Google Scholar] [CrossRef]
  9. Abbe, E. Community Detection and Stochastic Block Models: Recent Developments. J. Mach. Learn. Res. 2018, 18, 6446–6531. [Google Scholar]
  10. Agarwal, S.; Branson, K.; Belongie, S. Higher order learning with graphs. In Proceedings of the International Conference on Machine Learning, Pittsburgh, PA, USA, 25–29 June 2006; pp. 17–24. [Google Scholar]
  11. Amini, A.; Chen, A.; Bickel, P. Pseudo-likelihood methods for community detection in large sparse networks. Ann. Stat. 2013, 41, 2097–2122. [Google Scholar] [CrossRef]
  12. Ahn, K.; Lee, K.; Suh, C. Hypergraph Spectral Clustering in the Weighted Stochastic Block Model. IEEE J. Sel. Top. Signal Process. 2018, 12, 959–974. [Google Scholar] [CrossRef] [Green Version]
  13. Bickel, P.J.; Sarkar, P. Hypothesis testing for automated community detection in networks. J. R. Stat. Soc. Ser. B 2016, 78, 253–273. [Google Scholar] [CrossRef] [Green Version]
  14. Ghoshdastidar, D.; Dukkipati, A. Consistency of spectral partitioning of uniform hypergraphs under planted partition model. Adv. Neural Inf. Process. Syst. 2014, 27, 397–405. [Google Scholar]
  15. Ghoshdastidar, D.; Dukkipati, A. Consistency of spectral hypergraph partitioning under planted partition model. Ann. Stat. 2017, 45, 289–315. [Google Scholar] [CrossRef]
  16. Ke, Z.; Shi, F.; Xia, D. Community Detection for Hypergraph Networks via Regularized Tensor Power Iteration. arXiv 2020, arXiv:1909.06503. [Google Scholar]
  17. Kim, S. Higher-order correlation clustering for image segmentation. Adv. Neural Inf. Process. Syst. 2011, 24, 1530–1538. [Google Scholar]
  18. Yuan, M.; Liu, R.; Feng, Y.; Shang, Z. Testing community structures for hypergraphs. arXiv 2018, arXiv:1810.04617. [Google Scholar]
  19. Yuan, M.; Shang, Z. Sharp detection boundaries on testing dense subhypergraph. arXiv 2021, arXiv:2101.04584. [Google Scholar]
  20. Yuan, M.; Shang, Z. Heterogeneous Dense Subhypergraph Detection. arXiv 2021, arXiv:2104.04047. [Google Scholar]
  21. Yuan, M.; Shang, Z. Information Limits for Detecting a Subhypergraph. arXiv 2021, arXiv:2105.02259. [Google Scholar]
  22. Abbe, E.; Banderira, A.; Hall, G. Exact Recovery in the Stochastic Block Model. IEEE Trans. Inf. Theory 2016, 62, 471–487. [Google Scholar] [CrossRef] [Green Version]
  23. Saad, H.; Nosratinia, A. Community detection with side information: Exact recovery under the stochastic block model. IEEE J. Sel. Top. Signal Process. 2018, 12, 944–958. [Google Scholar] [CrossRef] [Green Version]
  24. Cai, T.T.; Liang, T.; Rakhlin, A. Inference via Message Passing on Partially Labeled Stochastic Block Models. arXiv 2016, arXiv:1603.06923. [Google Scholar]
  25. Kanade, V.; Mossel, E.; Schramm, T. Global and Local Information in Clustering Labeled Block Models. IEEE Trans. Inf. Theory 2016, 62, 5906–5917. [Google Scholar] [CrossRef]
  26. Kadavankandy, A.; Avrachenkov, K.; Cottatellucci, L.; Sundaresan, R. The Power of Side-Information in Subgraph Detection. IEEE Trans. Signal Process. 2018, 66, 1905–1919. [Google Scholar] [CrossRef] [Green Version]
  27. Mossel, E.; Xu, J. Local algorithms for block models with side information. In Proceedings of the 2016 ACM Conference on Innovations in Theoretical Computer Science, Cambridge, MA, USA, 14–16 January 2016; pp. 71–80. [Google Scholar]
  28. Tudisco, F.; Prokopchik, K.; Benson, A. A nonlinear diffusion method for semi-supervised learning on hypergraphs. arXiv 2021, arXiv:2103.14867. [Google Scholar]
  29. Tudisco, F.; Benson, A.; Prokopchik, K. Nonlinear Higher-Order Label Spreading. In Proceedings of the Web Conference 2021, Ljubljana, Slovenia, 19–23 April 2021. [Google Scholar]
  30. Whang, J.; Du, R.; Jung, S.; Lee, G.; Drake, B.; Liu, Q.; Kang, S.; Park, H. MEGA: Multi-View Semi-Supervised Clustering of Hypergraphs. Proc. VLDB Endow. 2020, 13, 698–711. [Google Scholar] [CrossRef]
Figure 1. Detection boundary with noisy labels observed for m = 2 , 3 and β = 0 , 0.4 , 0.8 . Red regions: exact recovery is impossible; green regions: exact recovery is possible.
Figure 1. Detection boundary with noisy labels observed for m = 2 , 3 and β = 0 , 0.4 , 0.8 . Red regions: exact recovery is impossible; green regions: exact recovery is possible.
Symmetry 13 02060 g001
Figure 2. Detection boundary with partially observed labels for m = 2 , 3 and β = 0 , 0.4 , 0.8 . Red regions: exact recovery is impossible; green regions: exact recovery is possible.
Figure 2. Detection boundary with partially observed labels for m = 2 , 3 and β = 0 , 0.4 , 0.8 . Red regions: exact recovery is impossible; green regions: exact recovery is possible.
Symmetry 13 02060 g002
Table 1. Regions for the exact recovery of community structure in a hypergraph with label information.
Table 1. Regions for the exact recovery of community structure in a hypergraph with label information.
Region Where Noisy Labels Are ObservedRecovery
( i ) η m , a , b ( β ) < 1 and β < C m , a , b ( a b ) Exact recovery is impossible
( i i ) β < 1 and β > C m , a , b ( a b ) Exact recovery is impossible
( i ) η m , a , b ( β ) > 1 and β m < C m , a , b ( a b ) Exact recovery is possible
( i i ) β > 1 and β > C m , a , b ( a b ) Exact recovery is possible
Region Where True Labels Are Partially ObservedRecovery
a b 2 2 m 1 ( m 1 ) ! + β < 1 Exact recovery is impossible
a b 2 2 m 1 ( m 1 ) ! + β > 1 Exact recovery is possible
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhao, X.; Zhao, W.; Yuan, M. Information Limits for Community Detection in Hypergraph with Label Information. Symmetry 2021, 13, 2060. https://doi.org/10.3390/sym13112060

AMA Style

Zhao X, Zhao W, Yuan M. Information Limits for Community Detection in Hypergraph with Label Information. Symmetry. 2021; 13(11):2060. https://doi.org/10.3390/sym13112060

Chicago/Turabian Style

Zhao, Xiaofeng, Wei Zhao, and Mingao Yuan. 2021. "Information Limits for Community Detection in Hypergraph with Label Information" Symmetry 13, no. 11: 2060. https://doi.org/10.3390/sym13112060

APA Style

Zhao, X., Zhao, W., & Yuan, M. (2021). Information Limits for Community Detection in Hypergraph with Label Information. Symmetry, 13(11), 2060. https://doi.org/10.3390/sym13112060

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop