Next Article in Journal
An Integrative Dynamic Model of Colombian Population Distribution, Based on the Maximum Entropy Principle and Matter, Energy, and Information Flow
Previous Article in Journal
Entropy and Information Theory: Uses and Misuses
Previous Article in Special Issue
On the Optimality of Interference Decoding Schemes for K-User Gaussian Interference Channels
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Strong Converse Theorem for Hypothesis Testing Against Independence over a Two-Hop Network †

1
School of Information Science and Engineering, Southeast University, Nanjing 210000, China
2
Department of Electrical Engineering and Computer Science, University of Michigan, Ann Arbor, MI 48109, USA
3
Department of Electrical and Computer Engineering and Department of Mathematics, National University of Singapore, Singapore 119077, Singapore
*
Author to whom correspondence should be addressed.
This paper was partially presented at ISIT 2019, Paris, France, 7–12 July 2019.
Entropy 2019, 21(12), 1171; https://doi.org/10.3390/e21121171
Submission received: 29 October 2019 / Revised: 18 November 2019 / Accepted: 28 November 2019 / Published: 29 November 2019
(This article belongs to the Special Issue Multiuser Information Theory II)

Abstract

:
By proving a strong converse theorem, we strengthen the weak converse result by Salehkalaibar, Wigger and Wang (2017) concerning hypothesis testing against independence over a two-hop network with communication constraints. Our proof follows by combining two recently-proposed techniques for proving strong converse theorems, namely the strong converse technique via reverse hypercontractivity by Liu, van Handel, and Verdú (2017) and the strong converse technique by Tyagi and Watanabe (2018), in which the authors used a change-of-measure technique and replaced hard Markov constraints with soft information costs. The techniques used in our paper can also be applied to prove strong converse theorems for other multiterminal hypothesis testing against independence problems.

1. Introduction

Motivated by situations where the source sequence is not available directly and can only be obtained through limited communication with the data collector, Ahlswede and Csiszár [1] proposed the problem of hypothesis testing with a communication constraint. In the setting of [1], there is one encoder and one decoder. The encoder has access to one source sequence X n and transmits a compressed version of it to the decoder at a limited rate. Given the compressed version and the available source sequence Y n (side information), the decoder knows that the pair of sequences ( X n , Y n ) is generated i.i.d. from one of the two distributions and needs to determine which distribution the pair of sequences is generated from. The goal in this problem is to study the tradeoff between the compression rate and the exponent of the type-II error probability under the constraint that the type-I error probability is either vanishing or non-vanishing. For the special case of testing against independence, Ahlswede and Csiszár provided an exact characterization of the rate-exponent tradeoff. They also derived the so-called strong converse theorem for the problem. This states that the rate-exponent tradeoff cannot be improved even when one is allowed a non-vanishing type-I error probability. However, the characterization the rate-exponent tradeoff for the general case (even in the absence of a strong converse) remains open till date.
Subsequently, the work of Ahlswede and Csiszár was generalized to the distributed setting by Han in [2] who considered hypothesis testing over a Slepian-Wolf network. In this setting, there are two encoders, each of which observes one source sequence and transmits a compressed version of the source to the decoder. The decoder then performs a hypothesis test given these two compression indices. The goal in this problem is to study the tradeoff between the coding rates and the exponent of type-II error probability, under the constraint that the type-I error probability is either vanishing or non-vanishing. Han derived an inner bound to the rate-exponent region. For the special case of zero-rate communication, Shalaby and Papamarcou [3] applied the blowing-up lemma [4] judiciously to derive the exact rate-exponent region and a strong converse theorem. Further generalizations of the work of Ahlswede and Csiszár can be categorized into two classes: non-interactive models where encoders do not communicate with one another [5,6,7,8] and the interactive models where encoders do communicate [9,10].
We revisit one such interactive model as shown in Figure 1. This problem was considered by Salehkalaibar, Wigger and Wang in [11] and we term the problem as hypothesis testing over a two-hop network. The two-hop model considered here has potential applications in the Internet of Things (IoT) and sensor networks. In these scenarios, direct communication from the transmitter to the receiver might not be possible due to power constraints that result from limited resources such as finite battery power. However, it is conceivable in such a scenario to assume that there are relays—in our setting, there is a single relay—that aid in the communication or other statistical inference tasks (such as hypothesis testing) between the transmitter and receiver.
The main task in this problem is to construct two hypothesis tests between two joint distributions P X Y Z and Q X Y Z . One of these two distributions governs the law of ( X n , Y n , Z n ) where each copy ( X i , Y i , Z i ) is generated independently either from P X Y Z and Q X Y Z . As shown in Figure 1, the first terminal has knowledge of a source sequence X n and sends an index M 1 to the second terminal, which we call the relay; the relay, given side information Y n and compressed index M 1 , makes a guess of the hypothesis H ^ Y and sends another index M 2 to the third terminal; the third terminal makes another guess of the hypothesis H ^ Z based on M 2 and its own side information Z n . The authors in [11] derived an inner bound for the rate-exponent region and showed that the bound is tight for several special cases, including the case of testing against independence in which Q X Y Z = P X P Y P Z . However, even in this simpler case of testing against independence, which is our main concern in this paper, the authors in [11] only established a weak converse.
In this paper, we strengthen the result by Salehkalaibar, Wigger and Wang in [11] by deriving a strong converse for the case of testing against independence. Our proof follows by combining two recently proposed strong converse techniques by Liu et al. in [12] and by Tyagi and Watanabe in [13]. In [12], the authors proposed a framework to prove strong converse theorems based on functional inequalities and reverse hypercontractivity of Markov semigroups. In particular, they applied their framework to derive strong converse theorems for a collection of problems including the hypothesis testing with communication constraints problem in [1]. In [13], the authors proposed another framework for strong converse proofs, where they used a change-of-measure technique and replaced hard Markov constraints with soft information costs. They also leveraged variational formulas for various information-theoretic quantities; these formulas were introduced by Oohama in [14,15,16].

Notation

Random variables and their realizations are in upper (e.g., X) and lower case (e.g., x) respectively. All sets are denoted in calligraphic font (e.g., X ). We use X c to denote the complement of X . Let X n : = ( X 1 , , X n ) be a random vector of length n and x n its realization. Given any x n , we use P ^ x n to denote its type (empirical distribution). All logarithms are base e. We use R + and N to denote the set of non-negative real numbers and natural numbers respectively. Given any positive integer a N , we use [ a ] to denote { 1 , , a } . We use 1 { · } to denote the indicator function and use standard asymptotic notation such as O ( · ) . The set of all probability distributions on a finite set X is denoted as P ( X ) . Given any two random variables ( X , Y ) and any realization of x, we use P Y | x ( · ) to denote the conditional distribution P Y | X ( · | x ) . Given a distribution P P ( X ) and a function f : X R , we use P ( f ) to denote E P [ f ( X ) ] . For information-theoretic quantities, we follow [17]. In particular, when the joint distribution of ( X , Y ) is P X Y P ( X × Y ) , we use I P X Y ( X ; Y ) and I P ( X ; Y ) interchangeably. Throughout the paper, for ease of notation, we drop the subscript for distributions when there is no confusion. For example, when the joint distribution of ( X , Y , Z ) is P X Y Z , we use I P ( X ; Y | Z ) and I P X Y Z ( X ; Y | Z ) interchangeably. For any ( p , q ) [ 0 , 1 ] 2 , let D b ( p q ) denote the binary divergence function, i.e., D b ( p q ) = p log ( p / q ) + ( 1 p ) log ( ( 1 p ) / ( 1 q ) ) .

2. Problem Formulation and Existing Results

2.1. Problem Formulation

Fix a joint distribution P X Y Z P ( X × Y × Z ) satisfying the Markov chain X Y Z , i.e.,
P X Y Z ( x , y , z ) = P X ( x ) P Y | X ( y | x ) P Z | Y ( z | y ) .
Let P X , P Y and P Z be induced marginal distributions of P X Y Z . As shown in Figure 1, we consider a two-hop hypothesis testing problem with three terminals. The first terminal, which we term the transmitter, observes a source sequence X n and sends a compression index M 1 to the second terminal, which we term the relay. Given M 1 and side information Y n , the relay sends another compression index M 2 to the third terminal, which we term the receiver. The main task in this problem is to construct hypothesis tests at both the relay and the receiver to distinguish between
H 0 : ( X n , Y n , Z n ) P X Y Z n = P X P Y | X n P Z | Y n ,
H 1 : ( X n , Y n , Z n ) P X n P Y n P Z n .
For subsequent analyses, we formally define a code for hypothesis testing over a two-hop network as follows.
Definition 1.
An ( n , N 1 , N 2 ) -code for hypothesis testing over a two-hop network consists of
  • Two encoders:
    f 1 : X n M 1 : = { 1 , , N 1 } ,
    f 2 : M 1 × Y n M 2 : = { 1 , , N 2 } , and
  • Two decoders
    g 1 : M 1 × Y n { H 0 , H 1 } ,
    g 2 : M 2 × Z n { H 0 , H 1 } .
Given an ( n , N 1 , N 2 ) -code with encoding and decoding functions ( f 1 , f 2 , g 1 , g 2 ) , we define acceptance regions for the null hypothesis H 0 at the relay and the receiver as
A Y , n : = { ( m 1 , y n ) : g 1 ( m 1 , y n ) = H 0 } ,
A Z , n : = { ( m 2 , z n ) : g 2 ( m 2 , z n ) = H 0 }
respectively. We also define conditional distributions
P M 1 | X n ( m 1 | x n ) : = 1 { f 1 ( x 1 n ) = m 1 } ,
P M 2 | Y n M 1 ( m 2 | y n , m 1 ) : = 1 { f 2 ( m 1 , y n ) = m 2 } .
Thus, for a ( n , N 1 , N 2 ) -code characterized by ( f 1 , f 2 , g 1 , g 2 ) , the joint distribution of random variables ( X n , Y n , Z n , M 1 , M 2 ) under the null hypothesis H 0 is given by
P X n Y n Z n M 1 M 2 ( x n , y n , z n , m 1 , m 2 ) = P X Y Z n ( x n , y n , z n ) P M 1 | X n ( m 1 | x n ) P M 2 | Y n M 1 ( m 2 | y n , m 1 ) ,
and under the alternative hypothesis H 1 is given by
P ¯ X n Y n Z n M 1 M 2 ( x n , y n , z n , m 1 , m 2 ) = P X n ( x n ) P Y n ( y n ) P Z n ( z n ) P M 1 | X n ( m 1 | x n ) P M 2 | Y n M 1 ( m 2 | y n , m 1 ) .
Now, let P Y n M 1 and P Z n M 2 be marginal distributions induced by P X n Y n Z n M 1 M 2 and let P ¯ Y n M 1 and P ¯ Z n M 2 be marginal distributions induced by P ¯ X n Y n Z n M 1 M 2 . Then, we can define the type-I and type-II error probabilities at the relay as
β 1 : = P M 1 Y n ( A Y , n c ) ,
β 2 : = P ¯ M 1 Y n ( A Y , n )
respectively and at the receiver as
η 1 : = P M 2 Z n ( A Z , n c ) ,
η 2 : = P ¯ M 2 Z n ( A Z , n )
respectively. Clearly, β 1 , β 2 , η 1 , and η 2 are functions of n but we suppress these dependencies for brevity.
Given above definitions, the achievable rate-exponent region for the hypothesis testing problem in a two-hop network is defined as follows.
Definition 2.
Given any ( ε 1 , ε 2 ) ( 0 , 1 ) 2 , a tuple ( R 1 , R 2 , E 1 , E 2 ) is said to be ( ε 1 , ε 2 ) -achievable if there exists a sequence of ( n , N 1 , N 2 ) -codes such that
lim sup n 1 n log N i R i , i { 1 , 2 } ,
lim sup n β 1 ε 1 ,
lim sup n η 1 ε 2 ,
lim inf n 1 n log β 2 E 1 ,
lim inf n 1 n log η 2 E 2 .
The closure of the set of all ( ε 1 , ε 2 ) -achievable rate-exponent tuples is called the ( ε 1 , ε 2 ) -rate-exponent region and is denoted as R ( ε 1 , ε 2 ) . Furthermore, define the rate-exponent region as
R : = R ( 0 , 0 ) .

2.2. Existing Results

In the following, we recall the exact characterization of R given by Salehkalaibar, Wigger and Wang ([11] (Corollary 1)). For this purpose, define the following set of joint distributions
Q : = { Q X Y Z U V P ( X × Y × Z × U × V ) : Q X Y Z = P X Y Z , U X Y , V Y Z } .
Given Q X Y Z U V Q , define the following set
R ( Q X Y Z U V ) : = { ( R 1 , R 2 , E 1 , E 2 ) : R 1 I Q ( U ; X ) , R 2 I Q ( V ; Y ) , E 1 I Q ( U ; Y ) , E 2 I Q ( U ; Y ) + I Q ( V ; Z ) }
Finally, let
R * : = Q X Y Z U V Q R ( Q X Y Z U V ) .
Theorem 1.
The rate-exponent region R for the hypothesis testing over a two-hop network problem satisfies
R = R * .
In the following, inspired by Oohama’s variational characterization of rate regions for multiuser information theory [14,15,16], we provide an alternative characterization of R * . For this purpose, given any ( b , c , d ) R + 3 and any Q X Y Z U V Q , let
R b , c , d ( Q X Y Z U V ) : = I Q ( U ; Y ) + b I Q ( U ; X ) c ( I Q ( U ; Y ) + I Q ( V ; Z ) ) + d I Q ( V ; Y ) .
be a linear combination of the mutual information terms in (25). Furthermore, define
R b , c , d : = min Q X Y Z U V Q R b , c , d ( Q X Y Z U V ) .
An alternative characterization of R * is given by
R * = ( b , c , d ) R + 3 ( R 1 , R 2 , E 1 , E 2 ) : E 1 + b R 1 c E 2 + d R 2 R b , c , d .

3. Strong Converse Theorem

3.1. The Case ε 1 + ε 2 < 1

Theorem 2.
Given any ( ε 1 , ε 2 ) ( 0 , 1 ) 2 such that ε 1 + ε 2 < 1 and any ( b , c , d ) R + 3 , for any ( n , N 1 , N 2 ) -code such that β 1 ε 1 , η 1 ε 2 , we have
log β 2 + b log N 1 + c log η 2 + d log N 2 n R b , c , d + Θ ( n 3 / 4 log n ) .
The proof of Theorem 2 is given in Section 4. Several remarks are in order.
First, using the alternative expression of the rate-exponent region in (30), we conclude that for any ( ε 1 , ε 2 ) ( 0 , 1 ) 2 such that ε 1 + ε 2 < 1 , we have R ( ε 1 , ε 2 ) = R * . This result significantly strengthens the weak converse result in ([11] (Corollary 1)) in which it was shown that R ( 0 , 0 ) = R * .
Second, it appears difficult to establish the strong converse result in Theorem 2 using existing classical techniques including image-size characterizations (a consequence of the blowing-up lemma) [4,6] and the perturbation approach [18]. In Section 4, we combine two recently proposed strong converse techniques by Liu, van Handel, and Verdú [12] and by Tyagi and Watanabe [13]. In particular, we use the strong converse technique based on reverse hypercontractivity in [12] to bound the exponent of the type-II error probability at the receiver and the strong converse technique in [13], which leverages an appropriate change-of-measure technique and replaces hard Markov constraints with soft information costs, to analyze the exponent of type-II error probability at the relay. Finally, inspired by the single-letterization steps in ([19] (Lemma C.2)) and [13], we single-letterize the derived multi-letter bounds from the previous steps to obtain the desired result in Theorem 2.
Third, we briefly comment on the apparent necessity of combining the two techniques in [12,13] instead of applying just one of them to obtain Theorem 2. The first step to apply the technique in [13] is to construct a “truncated source distribution” which is supported on a smaller set (often defined in terms of the decoding region) and is not too far away from the true source distribution in terms of the relative entropy. For our problem, the source satisfies the Markov chain X n Y n Z n . If we naïvely apply the techniques in [13], the Markovian property would not hold for the truncated source ( X ˜ n , Y ˜ n , Z ˜ n ) . On the other hand, it appears rather challenging to extend the techniques in [12] to the hypothesis testing over a multi-hop network problem since the techniques therein rely heavily on constructing semi-groups and it is difficult to devise appropriate forms of such semi-groups to be used and analyzed in this multi-hop setting. Therefore, we carefully combine the two techniques in [12,13] to ameliorate the aforementioned problems. In particular, we first use the technique in [13] to construct a truncated source ( X ˜ n , Y ˜ n ) and then let the conditional distribution of Z ˜ n given ( X ˜ n , Y ˜ n ) be given by the true conditional source distribution P Z | Y n to maintain the Markovian property of the source (see (56)). Subsequently, in the analysis of error exponents, we use the technique in [12] to analyze the exponent of type-II error probability at the receiver to circumvent the need to construct new semi-groups.
Finally, we remark that the techniques (or a subset of the techniques) used to prove Theorem 2 can also be used to establish a strong converse result for other multiterminal hypothesis testing against independence problems, e.g., hypothesis testing over the Gray-Wyner network [7], the interactive hypothesis testing problem [9] and the cascaded hypothesis testing problem [10].

3.2. The Case ε 1 + ε 2 > 1

In this subsection, we consider the case where the sum of type-I error probabilities at the relay and the receiver is upper bounded by a quantity strictly greater than one. For ease of presentation of our results, let
Q 2 : = { Q X Y Z U 1 U 2 V Q ( X × Y × Z × U 1 × U 2 × V ) : Q X Y Z = P X Y Z , U 1 X Y , U 2 X Y , V Y Z } .
Given any Q X Y Z U 1 U 2 V Q 2 , define the following set of rate-exponent tuples
R ˜ ( Q X Y Z U 1 U 2 V ) : = { ( R 1 , R 2 , E 1 , E 2 ) : R 1 max { I Q ( U 1 ; X ) , I Q ( U 2 ; X ) } , R 2 I Q ( V ; X ) , E 1 I Q ( U 1 ; Y ) , E 2 I Q ( U 2 ; Y ) + I Q ( V ; Z ) } .
Furthermore, define
R ˜ : = Q X Y Z U 1 U 2 V R ˜ ( Q X Y Z U 1 U 2 V ) .
Given any Q X Y Z U 1 U 2 V Q 2 and ( b 1 , b 2 , c , d ) R + 4 , define the following linear combination of the mutual information terms
R ˜ b 1 , b 2 , c , d ( Q X Y Z U 1 U 2 V ) : = I Q ( U 1 ; Y ) + b 1 I Q ( U 1 ; X ) + b 2 I Q ( U 2 ; X ) c ( I Q ( U 2 ; Y ) + I Q ( V ; Z ) ) + d I Q ( V ; Y ) ,
and let
R ˜ b 1 , b 2 , c , d : = min Q X Y Z U 1 U 2 V R ˜ b 1 , b 2 , c , d ( Q X Y Z U 1 U 2 V ) .
Then, based on [14,15,16], an alternative characterization of R ˜ is given by
R ˜ = ( b 1 , b 2 , c , d ) R + 4 ( R 1 , R 2 , E 1 , E 2 ) : E 1 + b 1 R 1 + b 2 R 1 c E 2 + d R 2 R ˜ b 1 , b 2 , c , d .
Analogously to Theorem 2, we obtain the following result.
Theorem 3.
Given any ( ε 1 , ε 2 ) ( 0 , 1 ) 2 and any ( b 1 , b 2 , c , d ) R + 4 , for any ( n , N 1 , N 2 ) -code such that β 1 ε 1 , η 1 ε 2 , we have
log β 2 + b 1 log N 1 + b 2 log N 1 + c log η 2 + d log N 2 n R ˜ b 1 , b 2 , c , d + Θ ( n 3 / 4 log n ) .
The proof of Theorem 3 is similar to that of Theorem 2 and thus omitted for simplicity.
To prove Theorem 3, we need to analyze two special cases (cf. Figure 2) of our system model separately:
(i)
Firstly, we consider the first hop, which involves the transmitter and the relay only. The first hop itself is a hypothesis testing problem with a communication constraint [1]. Using the techniques either in [13] or [12], we can obtain bounds on a linear combination of the rate of the first encoder and the type-II error exponent of the relay, (i.e., log β 2 + b 1 log N 1 for any b 1 R + ) for any type-I error probability β 1 ( 0 , 1 ) at the relay.
(ii)
Secondly, we study the second special case in which the relay does not make a decision. Using similar steps to the proof of Theorem 2, we can obtain a lower bound on a linear combination of the rate at the transmitter, the rate at the relay and the type-II exponent at the receiver (i.e., b 2 log N 1 + c log η 2 + d log N 2 for any ( b 2 , c , d ) R + 3 ) for any type-I error probability η 1 ( 0 , 1 ) at the receiver.
(iii)
Finally, combining the results obtained in the first two steps, we obtain a lower bound on the linear combination of rates and type-II exponents (as shown in Theorem 3). The proof is completed by using standard single-letterization steps and the variational formula in Equation (37).
Using Theorem 3, we obtain the following proposition.
Proposition 1.
For any ( ε 1 , ε 2 ) ( 0 , 1 ) 2 such that ε 1 + ε 2 > 1 , we have
R ( ε 1 , ε 2 ) = R ˜ .
The converse proof of Proposition 1 follows from Theorem 3 and the alternative characterization of R ˜ in (37). The achievability proof is inspired by ([6] (Theorem 5)) and is provided in Appendix A. The main idea is that we can time-share between two close-to optimal coding schemes, each of which corresponds to one special case of the current problem as mentioned after Theorem 3.
Recall that in the first remark of Theorem 2, we provide an exact characterization of the rate-exponent region for any ( ε 1 , ε 2 ) ( 0 , 1 ) 2 such that ε 1 + ε 2 < 1 . The converse proof follows from Theorem 2 and the achievability part was given in ([20] (Corollary 1)). Combining the first remark of Theorem 2 and Proposition 1, we provide an exact characterization of R ( ε 1 , ε 2 ) for any ( ε 1 , ε 2 ) ( 0 , 1 ) 2 such that ε 1 + ε 2 1 . We remark the case in which ε 1 + ε 2 = 1 was also excluded in the analysis of the successive refinement of hypothesis testing with communication constraints problem studied by Tian and Chen [6]. In fact, our converse result in Theorem 3 holds for any ( ε 1 , ε 2 ) ( 0 , 1 ) 2 including the case ε 1 + ε 2 = 1 . However, the achievability result presented in Appendix A holds only when ε 1 + ε 2 > 1 and thus we are unable to characterize R ( ε 1 , ε 2 ) when ε 1 + ε 2 = 1 . Because of the need to propose an achievability scheme which uses completely different techniques to handle the case in which ε 1 + ε 2 = 1 , which does not dovetail with the main message and contribution of this paper, we omit this case in this paper.

4. Proof of Theorem 2

4.1. Preliminaries

Before presenting the proof of Theorem 2, in this subsection, we briefly review the two strong converse techniques that we judiciously combine in this work, namely the change-of-measure technique by Tyagi and Watanabe [13] and the hypercontractivity technique by Liu et al. [12].
The critical step in the strong converse technique by Tyagi and Watanabe [13] is to construct a truncated source distribution, which is supported over a small set related to the decoding regions. Furthermore, the constructed truncated distribution should satisfy the following conditions:
(i)
The truncated distribution is close to the original source distribution in terms of the KL divergence;
(ii)
Under the truncated distribution, the (type-I) error probability is small.
Subsequent steps proceed similarly as the weak converse analysis of the problem and lead to bounds on the rates and (type-II) exponents. We then single-letterize the obtained bounds (using classical techniques in information theory without the memoryless property, e.g., [21]). Finally, we relate the single-letterized results to the the variational characterization [14,16] of the fundamental limit of the problem, which uses the idea of replacing hard Markov constraints with soft information costs.
The advantage of the Tyagi-Watanabe technique lies in its simplicity and similarity to weak converse analyses. In contrast, the disadvantage of the technique is that the structure of the source distribution (e.g., Markovian) is potentially lost in the constructed truncated distribution. As we have illustrated briefly after Theorem 2, this disadvantage prevents us from solely using the Tyagi-Watanabe technique to prove the strong converse theorem for our setting.
On the other hand, the key technique in the strong converse technique by Liu et al. [12] is the use of ideas from reverse hypercontractivity. In particular, one needs to use the variational formula of the KL divergence ([22] (Chapter 12)) and carefully construct Markov semigroups. The operation of applying a Markov semigroup is similar to a soft version of blowing up of decoding sets [4] for the discrete memoryless case. The advantage of the strong converse technique by Liu et al. lies in its wide applicability (beyond discrete settings) and its versatile performance (beyond showing strong converses it can be used to show that the second order terms scale as O ( n ) ). However, the construction of appropriate Markov semigroups is problem-specific, which limits its applicability to other information-theoretic problems in the sense that one has to construct specific semigroups for each problem. Fortunately, in our setting this construction and combination with Tyagi-Watanabe’s technique, is feasible.

4.2. Summary of Proof Steps

In the rest of this section, we present the proof of strong converse theorem for the hypothesis testing over the two-hop network. The proof follows by combining the techniques in [12,13] and is separated into three main steps. First, we construct a truncated source distribution P X ˜ n Y ˜ n Z ˜ n and show that this truncated distribution is not too different from P X Y Z n in terms of the relative entropy. Subsequently, we analyze the exponents of type-II error probabilities at the relay and the receiver under the constraint that their type-I error probabilities are non-vanishing. Finally, we single-letterize the constraints on rate and error exponents to obtain desired result in Theorem 2.
To begin with, let us fix an ( n , N 1 , N 2 ) -code with functions ( f 1 , f 2 , g 1 , g 2 ) such that the type-I error probabilities are bounded above by ε 1 ( 0 , 1 ) and ε 2 ( 0 , 1 ) respectively, i.e., β 1 ε 1 and η 1 ε 2 . We note from (19) and (20) that β 1 ε 1 + o ( 1 ) and β 2 ε 2 + o ( 1 ) . Since the o ( 1 ) terms are immaterial in the subsequent analyses, they are omitted for brevity.

4.3. Step 1: Construction of a Truncated Distribution

Paralleling the definitions of acceptance regions in (8) and (9), we define the following acceptance regions at the relay and the receiver as
D Y , n = { ( x n , y n ) : g 1 ( f 1 ( x n ) , y n ) = H 0 } ,
D Z , n = { ( x n , y n , z n ) : g 2 ( f 2 ( f 1 ( x n ) , y n ) , z n ) = H 0 } ,
respectively. Note that the only difference between A Y , n and D Y , n lies in whether we consider the compression index m 1 or the original source sequence x n . Recalling the definitions of the type-I error probabilities for the relay denoted by β 1 in (14) and for the receiver denoted by η 1 in (16), and using (40) and (41), we conclude that
P X Y n ( D Y , n ) = 1 β 1 ,
P X Y Z n ( D Z , n ) = 1 η 1 .
For further analysis, given any m 2 M 2 , define a conditional acceptance region at the receiver (conditioned on m 2 ) as
G ( m 2 ) : = { z n : g 2 ( m 2 , z n ) = H 0 } .
For ease of notation, given any ( x n , y n ) X n × Y n , we use G ( x n , y n ) and G ( f 2 ( f 1 ( x n ) , y n ) ) (here f 2 ( f 1 ( x n ) , y n ) plays the role of m 2 in (44)) interchangeably and define the following set
B n : = ( x n , y n ) : P Z | Y n ( G ( x n , y n ) | y n ) 1 ε 1 ε 2 1 + 3 ε 2 ε 1 .
Combining (41), (43) and (44), we obtain
1 ε 2 P X Y Z n ( D Z , n )
= ( x n , y n ) B n P X Y n ( x n , y n ) P Z | Y n ( G ( x n , y n ) | y n ) + ( x n , y n ) B n P X Y n ( x n , y n ) P Z | Y n ( G ( x n , y n ) | y n )
P X Y n ( B n ) + ( 1 P X Y n ( B n ) ) 1 ε 1 ε 2 1 + 3 ε 2 ε 1 .
Thus, we have
P X Y n ( B n ) 3 3 ε 2 + ε 1 4 .
For subsequent analyses, let
μ : = min y : P Y ( y ) > 0 P Y ( y ) 1 ,
θ n : = 3 μ n log 8 | Y | 1 ε 1 ε 2 ,
and define the typical set T n ( P Y ) as
T n ( P Y ) = { y n : | P ^ y n ( y ) P Y ( y ) | θ n P Y ( y ) y Y } .
Using the Chernoff bound, we conclude that when n is sufficiently large,
P Y n ( T n ( P Y ) ) 1 1 ε 1 ε 2 4 .
Now, define the following set
C n : = B n D Y , n ( X n × T n ( P Y ) ) .
Then, combining (42), (49) and (53), we conclude that when n is sufficiently large,
P X Y n ( C n ) 1 P X Y n ( B n c ) P X Y n ( D Y , n c ) P Y n ( T n c ( P Y ) ) 1 ε 1 ε 2 2 .
Let the truncated distribution P X ˜ n Y ˜ n Z ˜ n be defined as
P X ˜ n Y ˜ n Z ˜ n ( x n , y n , z n ) : = P X Y n ( x n , y n ) 1 { ( x n , y n ) C n } P X Y n ( C n ) P Z | Y n ( z n | y n ) .
Note that under our constructed truncated distribution P X ˜ n Y ˜ n Z ˜ n , the Markov chain X ˜ n Y ˜ n Z ˜ n holds.In other words, the Markovian property of the original source distribution P X Y Z n is retained for the truncated distribution P X ˜ n Y ˜ n Z ˜ n , which appears to be necessary to obtain a tight result if one wishes to use weak converse techniques. This is critical for our subsequent analyses.
Using the result in (55), we have that the marginal distribution P X ˜ n satisfies that for any x n X N ,
P X ˜ n ( x n ) = y n , z n P X ˜ n Y ˜ n Z ˜ n ( x n , y n , z n )
P X n ( x n ) P X Y n ( C n ) 2 P X n ( x n ) 1 ε 1 ε 2 .
Analogously to (58), we obtain that
P Y ˜ n ( y n ) 2 P Y n ( y n ) 1 ε 1 ε 2 , y n Y n ,
P Z ˜ n ( z n ) 2 P Z n ( z n ) 1 ε 1 ε 2 , z n Z n .
Finally, note that
D ( P X ˜ n Y ˜ n Z ˜ n P X Y Z n ) = D ( P X ˜ n Y ˜ n P X Y n )
= log 1 P X Y n ( C n )
log 2 1 ε 1 ε 2 .

4.4. Step 2: Analyses of the Error Exponents of Type-II Error Probabilities

4.4.1. Type-II Error Probability β 2 at the Relay

Let M ˜ 1 and M ˜ 2 be the outputs of encoders f 1 and f 2 respectively when the tuple of source sequences ( X ˜ n , Y ˜ n , Z ˜ n ) is distributed according to P X ˜ n Y ˜ n Z ˜ n defined in (56). Thus, recalling the definitions in (10), (11) and (56), we find that the joint distribution of ( X ˜ n , Y ˜ n , Z ˜ n , M ˜ 1 , M ˜ 2 ) is given by
P X ˜ n Y ˜ n Z ˜ n M ˜ 1 M ˜ 2 ( x n , y n , z n , m 1 , m 2 ) = P X ˜ n Y ˜ n Z ˜ n ( x n , y n , z n ) P M 1 | X n ( m 1 | x n ) P M 2 | Y n M 1 ( m 2 | y n , m 1 ) .
Let P M ˜ 1 Y ˜ n be induced by P X ˜ n Y ˜ n Z ˜ n M ˜ 1 M ˜ 2 . Combining (8) and (56), we conclude that
P M ˜ 1 Y ˜ n ( A Y , n ) = x n , y n , z n , m 1 , m 2 : g 1 ( m 1 , y n ) = H 0 P X ˜ n Y ˜ n Z ˜ n M ˜ 1 M ˜ 2 ( x n , y n , z n , m 1 , m 2 )
= x n , y n : g 1 ( f 1 ( x n ) , y n ) = H 0 P X Y n ( x n , y n ) 1 { ( x n , y n ) C n } P X Y n ( C n )
= x n , y n P X Y n ( x n , y n ) 1 { ( x n , y n ) C n } P X Y n ( C n )
= 1 .
where (67) follows from the definition of D Y , n in (40) and the fact that D Y , n C n .
Thus, using the data processing inequality for the relative entropy and the definition of β 2 in (15), we obtain that
D ( P M ˜ 1 Y ˜ n P M 1 P Y n ) D b ( P M ˜ 1 Y ˜ n ( A Y , n ) P M 1 P Y n ( A Y , n ) )
= log P M 1 P Y n ( A Y , n )
= log β 2 .
Furthermore, recalling that M 1 denotes the output of encoder f 1 when ( X n , Y n , Z n ) P X Y Z n and M ˜ 1 denotes the output of encoder f 1 when ( X n , Y n , Z n ) P X ˜ n Y ˜ n Z ˜ n , and using the result in (58), we conclude that
P M ˜ 1 ( m 1 ) = x n , y n , z n : f 1 ( x n ) = m 1 P X ˜ n Y ˜ n Z ˜ n ( x n , y n , z n )
= x n : f 1 ( x n ) = m 1 P X ˜ n ( x n )
x n : f 1 ( x n ) = m 1 2 P X n ( x n ) 1 ε 1 ε 2
2 P M 1 ( m 1 ) 1 ε 1 ε 2 ,
for any m 1 M 1 . Thus, combining (59), (71) and (75), we have
log β 2 D ( P M ˜ 1 Y ˜ n P M 1 P Y n )
= D ( P M ˜ 1 Y ˜ n P M ˜ 1 P Y ˜ n ) + E P M ˜ 1 Y ˜ n log P M ˜ 1 ( M ˜ 1 ) P Y ˜ n ( Y ˜ n ) P M 1 ( M ˜ 1 ) P Y n ( Y ˜ n )
D ( P M ˜ 1 Y ˜ n P M ˜ 1 P Y ˜ n ) + E P M ˜ 1 Y ˜ n log 2 P M 1 ( M ˜ 1 ) 1 ε 1 ε 2 2 P Y n ( Y ˜ n ) 1 ε 1 ε 2 P M 1 ( M ˜ 1 ) P Y n ( Y ˜ n )
= I ( M ˜ ; Y ˜ n ) + 2 log 2 1 ε 1 ε 2 .

4.4.2. Type-II Error Probability η 2 at the Receiver

In this subsection, we analyze the error exponent of the type-II error probability at the receiver. For this purpose, we make use of the method introduced in [12] based on reverse hypercontractivity. We define the following additional notation:
  • Give P Y Z P ( Y × Z ) , define
    α : = max y , z P Z | Y ( z | y ) P Z ( z ) ( 1 , ) .
    In the subsequent analysis, we only consider the case when α > 1 . When α = 1 , choosing t = 1 n instead of the choice in (101), we can obtain a similar upper bound for log η 2 as in (102), where the only difference is that Ψ ( n , ε 1 , ε 2 ) should be replaced by another term scaling in order Θ ( n ) .
  • Given any ( ε 1 , ε 2 ) ( 0 , 1 ) 2 such that ε 1 + ε 2 < 1 , let
    Ψ ( n , ε 1 , ε 2 ) : = 2 n ( α 1 ) log 1 + 3 ε 2 ε 1 1 ε 1 ε 2 .
  • Give any m 2 M 2 and z n Z n , let
    h ( m 2 , z n ) : = 1 { z n G ( m 2 ) } .
  • Two operators in ([12] (Equations (25), (26), (29)))
    Λ α , t = ( exp ( t ) + α ( 1 exp ( t ) ) P Z ) n ,
    T y n , t = i = 1 n ( exp ( t ) + ( 1 exp ( t ) ) P Z | y i ) .
Note that in (84), we use the convenient notation P Z | y ( z ) = P Z | Y ( z | y ) . The two operators in (83) and (84) will be used to lower bound D ( P Z ˜ n M ˜ 2 P Z n P ¯ M 2 ) via a variational formula of the relative entropy (cf. ([12] (Section 4))).
Let P Z ˜ n M ˜ 2 , P Z ˜ n | M ˜ 2 , P Z ˜ n | Y ˜ n be induced by the joint distribution P X ˜ n Y ˜ n Z ˜ n M ˜ 1 M ˜ 2 in (64) and let P ¯ M 2 be induced by the joint distribution P ¯ X n Y n Z n M 1 M 2 in (13). Invoking the variational formula for the relative entropy ([22] (Equation (2.4.67))) and recalling the notation P ( f ) = E P [ f ] , we have
D ( P Z ˜ n M ˜ 2 P Z n P ¯ M 2 ) P Z ˜ n M ˜ 2 log Λ α , t h ( M ˜ 2 , Z ˜ n ) log ( P Z n P ¯ M 2 ) Λ α , t h ( M 2 , Z n ) .
Given any m 2 M 2 , similar to ([12] (Equations (18)–(21))), we obtain
P Z n ( Λ α , t h ( m 2 , Z n ) )
= P Z n ( exp ( t ) + α ( 1 exp ( t ) ) P Z ) n h ( m 2 , Z n )
= exp ( t ) + α ( 1 exp ( t ) ) n P Z n h ( m 2 , Z n )
exp ( ( α 1 ) n t ) P Z n h ( m 2 , Z n ) .
Thus, averaging over m 2 with distribution P ¯ M 2 on both sides of (88), we have
( P Z n P ¯ M 2 ) ( Λ α , t h ( M 2 , Z n ) )
exp ( ( α 1 ) n t ) ( P ¯ M 2 P Z n ) h ( M 2 , Z n )
= exp ( ( α 1 ) n t ) η 2 ,
where (90) follows from the definition of η 2 in (17).
Furthermore, given any m ˜ 2 M 2 , we obtain
P Z ˜ n | m ˜ 2 ( log Λ α , t h ( m ˜ 2 , Z ˜ n ) )
= y ˜ n P Z ˜ n | y ˜ n P Y ˜ n | M ˜ 2 ( y ˜ n | m ˜ 2 ) ( log Λ α , t h ( m ˜ 2 , Z ˜ n ) )
= y ˜ n P Y ˜ n | M ˜ 2 ( y ˜ n | m ˜ 2 ) P Z ˜ n | y ˜ n ( log Λ α , t h ( m ˜ 2 , Z ˜ n ) )
y ˜ n P Y ˜ n | M ˜ 2 ( y ˜ n | m ˜ 2 ) P Z ˜ n | y ˜ n ( log T y n , t h ( m ˜ 2 , Z ˜ n ) )
y ˜ n P Y ˜ n | M ˜ 2 ( y ˜ n | m ˜ 2 ) 1 + 1 t log P Z ˜ n | y ˜ n h ( m ˜ 2 , Z ˜ n )
= 1 + 1 t y ˜ n P Y ˜ n | M ˜ 2 ( y ˜ n | m ˜ 2 ) log P Z ˜ n | y ˜ n ( G ( m ˜ 2 ) ) .
where (94) follows from ([12] (Lemma 4)) and (95) follows similarly to ([12] (Equations (14)–(17))).
Thus, averaging on both sides of (96) over m ˜ 2 with distribution P M ˜ 2 and using the definition of the joint distribution P X ˜ n Y ˜ n Z ˜ n M ˜ 1 M ˜ 2 in (64), we obtain that
P Z ˜ n M ˜ 2 ( log Λ α , t h ( M ˜ 2 , Z ˜ n ) )
1 + 1 t y ˜ n , m ˜ 2 P Y ˜ n M ˜ 2 ( y ˜ n , m ˜ 2 ) log P Z ˜ n | y ˜ n ( G ( m ˜ 2 ) ) = 1 + 1 t x ˜ n , y ˜ n , m ˜ 1 , m ˜ 2 ( P X ˜ n Y ˜ n ( x ˜ n , y ˜ n ) 1 { m ˜ 1 = f 1 ( x ˜ n ) , m ˜ 2 = f 2 ( m ˜ 1 , y ˜ n ) }
× log z ˜ n : g 2 ( z ˜ n , m ˜ 2 ) = H 0 P Z | Y n ( z ˜ n | y ˜ n ) )
= 1 + 1 t x ˜ n , y ˜ n P X Y n ( x ˜ n , y ˜ n ) 1 { ( x ˜ n , y ˜ n ) C n } P X Y n ( C n ) log P Z | Y n ( G ( x ˜ n , y ˜ n ) | y ˜ n )
1 + 1 t log 1 ε 1 ε 2 1 + 3 ε 2 ε 1 ,
where (100) follows from the definitions of B n in (45) and C n in (54).
Therefore, combining (85), (90) and (100) and choosing
t = 1 n ( α 1 ) log 1 + 3 ε 2 ε 1 1 ε 1 ε 2 ,
via simple algebra, we obtain that
log η 2 D ( P Z ˜ n M ˜ 2 P Z n P ¯ M 2 ) + Ψ ( n , ε 1 , ε 2 ) log 1 ε 1 ε 2 1 + 3 ε 2 ε 1 .
In the following, we further upper bound D ( P Z ˜ n M ˜ 2 P Z n P ¯ M 2 ) . For this purpose, define the following distribution
P ¯ M ˜ 2 ( m 2 ) : = y n , m 1 P M ˜ 1 ( m 1 ) P Y ˜ n ( y n ) 1 { m 2 = f 2 ( m 1 , y n ) } .
Combining the results in (59) and (75), and recalling that P ¯ M 2 is induced by joint distribution P ¯ X n Y n Z n M 1 M 2 in (13), for any m 2 M 2 , we have
P ¯ M ˜ 2 ( m 2 ) 2 1 ε 1 ε 2 2 y n , m 1 P M 1 ( m 1 ) P Y n ( y n ) 1 { m 2 = f 2 ( f 1 ( x n ) , y n ) }
= 4 P ¯ M 2 ( m 2 ) ( 1 ε 1 ε 2 ) 2 .
Thus, combining (60) and (105), we have
D ( P Z ˜ n M ˜ 2 P Z n P ¯ M 2 )
= D ( P Z ˜ n M ˜ 2 P Z ˜ n P ¯ M ˜ 2 ) + E P Z ˜ n M ˜ 2 log P Z ˜ n ( Z ˜ n ) P ¯ M ˜ 2 ( M ˜ 2 ) P Z n ( Z ˜ n ) P ¯ M 2 ( M ˜ 2 )
D ( P Z ˜ n M ˜ 2 P Z ˜ n P ¯ M ˜ 2 ) + E P Z ˜ n M ˜ 2 log 2 P Z n ( Z ˜ n ) 1 ε 1 ε 2 4 P ¯ M 2 ( M ˜ 2 ) ( 1 ε 1 ε 2 ) 2 P Z n ( Z ˜ n ) P ¯ M 2 ( M ˜ 2 )
= D ( P Z ˜ n M ˜ 2 P Z ˜ n P ¯ M ˜ 2 ) + 3 log 2 1 ε 1 ε 2 .
Therefore, combining (102) and (108), we have
log η 2 D ( P Z ˜ n M ˜ 2 P Z ˜ n P ¯ M ˜ 2 ) + Ψ ( n , ε 1 , ε 2 ) log 1 ε 1 ε 2 1 + 3 ε 2 ε 1 3 log 1 ε 1 ε 2 2 .

4.5. Step 3: Analyses of Communication Constraints and Single-Letterization Steps

For any ( n , N 1 , N 2 ) -code, since M ˜ i M i for i { 1 , 2 } , we have that
log N 1 H ( M ˜ 1 ) I ( M ˜ 1 ; X ˜ n , Y ˜ n ) ,
log N 2 H ( M ˜ 2 ) I ( M ˜ 2 ; Y ˜ n ) .
Furthermore, from the problem setting (see (64)), we have
I ( M ˜ 1 ; Y ˜ n | X ˜ n ) = 0 ,
For subsequent analyses, given any ( b , c , d , γ ) R + 4 , define
R b , c , d , γ ( n ) : = I ( M ˜ 1 ; Y ˜ n ) + b I ( M ˜ 1 ; X ˜ n , Y ˜ n ) c D ( P Z ˜ n M ˜ 2 P Z ˜ n P ¯ M ˜ 2 ) + d I ( M ˜ 2 ; Y ˜ n ) + γ I ( M ˜ 1 ; Y ˜ n | X ˜ n ) + ( b + d + γ ) D ( P X ˜ n Y ˜ n P X n Y n ) .
Combining the results in (63), (79), (109) to (112), for any γ R + , we obtain
log β 2 + b log N 1 + c log η 2 + d log N 2 + c Ψ ( n , ε 1 , ε 2 ) R b , c , d , γ ( n ) + log 1 ε 1 ε 2 1 + 3 ε 2 ε 1 + ( b + d + γ + 5 ) log 1 ε 1 ε 2 2 .
The proof of Theorem 2 is complete by the two following lemmas which provide a single-letterized lower bound for R b , c , d , γ ( n ) and relate the derived lower bound to R b , c , d . For this purpose, recalling the definition of θ n in (51), we define the following set of joint distributions
Q 1 : = { Q X Y Z U V P ( X × Y × Z × U × V ) : Q Z | Y = P Z | Y , X Y Z , V Y Z , | Q Y ( y ) P Y ( y ) | θ n P Y ( y ) , y Y } .
Given Q X Y Z U V Q 1 , define
Δ b , d , γ ( Q X Y Z U V ) : = ( b + γ ) D ( Q X Y P X Y ) + d D ( Q Y P Y ) + γ I Q ( U ; Y | X ) .
Recall the definition of R b , c , d ( Q X Y Z U V ) in (28). Define
R b , c , d , γ : = min Q X Y Z U V Q 1 R b , c , d ( Q X Y Z U V ) + Δ b , d , γ ( Q X Y Z U V ) .
The following lemma presents a single-letterized lower bound for R b , c , d , γ ( n ) .
Lemma 1.
For any ( b , c , d , γ ) R + 4 ,
R b , c , d , γ ( n ) n R b , c , d , γ .
The proof of Lemma 1 is inspired by ([13] (Prop. 2)) and provided in Appendix B.
Combining the results in (114) and Lemma 1, we obtain the desired result and this completes the proof of Theorem 2.
Lemma 2.
Choosing γ = n , we have
n R b , c , d , γ + log 1 ε 1 ε 2 1 + 3 ε 2 ε 1 + ( b + d + γ + 5 ) log 1 ε 1 ε 2 2 n R b , c , d + Θ ( n 3 / 4 log n ) .
The proof of Lemma 2 is inspired by ([19] (Lemma C.2)) and provided in Appendix C.

5. Discussion and Future Work

We strengthened the result in ([11] (Corollary 1)) by deriving a strong converse theorem for hypothesis testing against independence over a two-hop network with communication constraints (see Figure 1). In our proof, we combined two recently proposed strong converse techniques [12,13]. The apparent necessity of doing so comes from the Markovian requirement in the source distribution (recall (1)) and is reflected in the construction of a truncated distribution in (56) to ensure the Markovian structure of the source sequences is preserved. Subsequently, due to this constraint, the application the strong converse technique by Tyagi and Watanabe in [13] was only amenable in analyzing the type-II error exponent at the relay. On the other hand, to analyze the type-II error exponent at the receiver, we need to carefully adapt the strong converse technique based on reverse hypercontractivity by Liu, van Handel and Verdú in [12]. Furthermore, to complete the proof, we carefully combine the single-letterization techniques in [12,13].
Another important take-home message is the techniques (or a subset of the techniques) used in this paper can be applied to strengthen the results of other multiterminal hypothesis testing against independence problems. If the source distribution has no Markov structure, it is usually the case that one can directly apply the technique by Tyagi and Watanabe [13] to obtain strong converse theorems. Such examples include [7,8,9]. On the other hand, if the source sequences admit Markovian structure, then it appears necessary to combine techniques in [12,13] to obtain strong converse theorems, just as it was done in this paper.
Finally, we discuss some avenues for future research. In this paper, we only derived the strong converse but not a second-order converse result as was done in ([12] (Section 4.4)) for the problem of hypothesis testing against independence with a communication constraint [1]. Thus, in the future, one may refine the proof in the current paper by deriving second-order converse or exact second-order asymptotics. Furthermore, one may also consider deriving strong converse theorems or simplifying existing strong converse proofs for hypothesis testing problems with both communication and privacy constraints such as that in [23] by using the techniques in the current paper. It is also interesting to explore whether current techniques can be applied to obtain strong converse theorems for hypothesis testing with zero-rate compression problems [3].

Author Contributions

Formal analysis, D.C. and L.Z.; Supervision, V.Y.F.T.; Writing—original draft, D.C. and L.Z.; Writing—review & editing, V.Y.F.T.

Funding

D.C. is supported by the China Scholarship Council with No. 201706090064 and the National Natural Science Foundation of China under Grant 61571122. L.Z. was supported by NUS RSB grants (C-261-000-207-532 and C-261-000-005-001).

Acknowledgments

The authors acknowledge Sadaf Salehkalaibar (University of Tehran) for drawing our attention to ([11] (Corollary 1)) and providing helpful comments and suggestions.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Achievability Proof of Proposition 1

Fix any joint distribution Q X Y Z U 1 U 2 V Q 2 . Let ( f 1 , g 1 ) be an encoder-decoder pair with rate R 1 = I Q ( U 1 ; X ) for the hypothesis testing with communication constraint problem [1] (i.e., no receiver in Figure 1) such that the type-II error probability decays exponentially fast at speed no smaller than E 1 = I Q ( U 1 ; Y ) and the type-I error probability is vanishing, i.e., log N 1 n R 1 , β 2 exp ( n E 1 ) and β 1 ε 1 for any ε 1 > 0 . Furthermore, let ( f 1 , f 2 , g 1 , g 2 ) be a tuple of encoders and decoders with rates ( R 1 , R 2 ) = ( I Q ( U 2 ; X ) , I Q ( V ; Y ) ) for the problem in Figure 1 such that the type-II error probability at the receiver decays exponentially fast at speed no smaller E 2 = I Q ( V ; Z ) and type-I error probability at the receiver is vanishing, i.e., log N 1 n R 1 , log N 2 n R 2 , η 2 exp ( n E 2 ) and η 1 ε 2 for any ε 2 > 0 . Such tuples of encoders and decoders exist as proved in [1,11]. Furthermore, let A 1 X n × Y n be the acceptance region associated with ( f 1 , g 1 ) at the relay and let A 2 X n × Y n × Z n be the acceptance region associated with ( f 1 , f 2 , g 1 , g 2 ) at the receiver.
Now, let us partition the source space X n into two disjoint sets X 1 n and X 2 n such that X 1 n X 2 n = X n , P X n ( X 1 n ) > 1 ε 1 and P X n ( X 2 n ) > 1 ε 2 . We construct an ( n , N 1 , N 2 ) -code as follows. Given a source sequence X n , if X n X 1 n , then encoder f 1 is used; and if otherwise, the encoder f 1 is used. Furthermore, an additional bit indicating whether X n X 1 n is also sent to the relay and further forwarded to the receiver by the relay. Given encoded index M 1 , if X n X 1 n , the relay uses decoder g 1 to make the decision; otherwise, if X n X 2 n , the relay declares hypothesis H 1 to be true. Furthermore, in both cases, the relay transmits an index M 2 using encoder f 2 . Given the index M 2 , if X n X 1 n , the receiver declares hypothesis H 1 to be true; otherwise, the receiver uses decoder g 2 to make the decision.
The performance of the constructed ( n , N 1 , N 2 ) -code is as follows. In terms of rates, we have
log N 1 n R 1 + 1 ,
log N 2 n R 2 + 1 .
The type-I error probability at the relay satisfies that
1 β 1 = P X Y n { A 1 ( X 1 n × Y n ) }
P X n { X 1 n } P X Y n { ( A 1 ) c }
1 ε 1 ,
where (A5) follows when n is sufficiently large and thus ε 1 can be made arbitrarily close to zero. Furthermore, the type-II error probability at the relay can be upper bounded as follows
β 2 = P X n P Y n { A 1 ( X 1 n × Y n ) }
P X n P Y n { A 1 }
= β 2
exp ( n E 1 ) .
Similarly, for n sufficiently large, the error probabilities at the receiver can be upper bounded as follows
η 1 = 1 P X Y Z n { A 2 ( X 2 n × Y n × Z n ) }
1 P X n ( X 2 n ) + P X Y Z n ( A 2 ) c
ε 2 ,
and
η 2 = P X n P Y n P Z n { A 2 ( X 2 n × Y n × Z n ) }
P X n P Y n P Z n { A 2 }
exp ( n E 2 ) .
The achievability proof of Proposition 1 is now complete.

Appendix B. Proof of Lemma 1

Recall the definition of distribution P ¯ M ˜ 2 (see (103)). Noting that P M ˜ 2 is the marginal distribution induced by P X ˜ n Y ˜ n Z ˜ n M ˜ 1 M ˜ 2 (see (64)), we have that for any m ˜ 2 M 2
P M ˜ 2 ( m ˜ 2 ) = y n , m 1 P Y ˜ n M ˜ 1 ( y n , m 1 ) 1 { m ˜ 2 = f 2 ( m 1 , y n ) } .
Thus, applying the data processing inequality for the relative entropy, we have that
I ( M ˜ 1 ; Y ˜ n ) = D ( P Y ˜ n M ˜ 1 P Y ˜ n P M ˜ 1 )
D ( P M ˜ 2 P ¯ M ˜ 2 ) .
Using (A18) and following similar steps to the proof of weak converse in ([11] (Equationation (186))), we obtain
D ( P Z ˜ n M ˜ 2 P Z ˜ n P ¯ M ˜ 2 ) = I ( M ˜ 2 ; Z ˜ n ) + D ( P M ˜ 2 P ¯ M ˜ 2 )
I ( M ˜ 2 ; Z ˜ n ) + I ( M ˜ 1 ; Y ˜ n ) .
Using (A20) and the definition of R b , c , d , γ ( n ) in (113), we have the following lower bound for R b , c , d , γ ( n )
R b , c , d , γ ( n ) I ( M ˜ 1 ; Y ˜ n ) + b D ( P X ˜ n Y ˜ n P X n Y n ) + H ( X ˜ n , Y ˜ n ) H ( X ˜ n , Y ˜ n | M ˜ 1 ) c I ( M ˜ 2 ; Z n ) + I ( M ˜ 1 ; Y ˜ n ) + d D ( P X ˜ n Y ˜ n P X n Y n ) + H ( Y ˜ n ) h ( Y ˜ n | M ˜ 2 ) + γ D ( P X ˜ n Y ˜ n P X n Y n ) + H ( Y ˜ n | X ˜ n ) H ( Y ˜ n | X ˜ n , M ˜ 1 ) .
The rest of the proof concerns single-letterizing each term in (A21). For this purpose, for each j [ n ] , we define two auxiliary random variables U j : = ( M ˜ 1 , X ˜ j 1 , Y ˜ j 1 ) and V j : = ( M ˜ 2 , Y ˜ j 1 ) and let J be a random variable which is distributed uniformly over the set [ n ] and is independent of all other random variables.
Using standard single-letterization techniques as in [21], we obtain
I ( M ˜ 1 ; Y ˜ n ) = j [ n ] I ( M ˜ 1 ; Y ˜ j | Y ˜ j 1 )
j [ n ] I ( M ˜ 1 , Y ˜ j 1 ; Y ˜ j )
j [ n ] I ( M ˜ 1 , X ˜ j 1 , Y ˜ j 1 ; Y ˜ j )
= n I ( U J , J ; Y ˜ J ) ,
and
H ( X ˜ n , Y ˜ n | M ˜ 1 ) = n H ( X ˜ J , Y ˜ J | U J , J ) .
Furthermore, analogous to ([13] (Prop. 1)), we obtain that
H ( X ˜ n , Y ˜ n ) + D ( P X ˜ n Y ˜ n P X Y n ) = x n , y n P X ˜ n Y ˜ n ( x n , y n ) log 1 P X Y n ( x n , y n )
= x n , y n P X ˜ n Y ˜ n ( x n , y n ) j [ n ] log 1 P X Y ( x j , y j )
= j [ n ] P X ˜ j Y ˜ j ( x j , y j ) log 1 P X Y ( x j , y j )
= n H ( X ˜ J , Y ˜ J ) + D ( P X ˜ J Y J P X Y ) .
Subsequently, we can single-letterize I ( M ˜ 2 ; Z ˜ n ) as follows:
I ( M ˜ 2 ; Z ˜ n ) = j [ n ] I ( M ˜ 2 ; Z ˜ j | Z ˜ j 1 )
j [ n ] I ( M ˜ 2 , Z ˜ j 1 , Y ˜ j 1 ; Z ˜ j )
= j [ n ] I ( M ˜ 2 , Y ˜ j 1 ; Z ˜ j )
= n I ( V J , J ; Z ˜ J ) ,
where (A33) follows from the Markov chain Z ˜ j 1 M ˜ 2 Y ˜ j 1 Z ˜ j implied by the joint distribution of ( X ˜ n , Y ˜ n , Z ˜ n , M ˜ 1 , M ˜ 2 ) in (64). Furthermore, using similar proof techniques to ([13] (Prop. 1)) and standard single-letterization techniques (e.g., in [4] or [21]), we obtain that
H ( Y ˜ n | X ˜ n ) + D ( P X ˜ n Y ˜ n P X Y n ) n H ( Y ˜ J | X ˜ J ) + D ( P X ˜ J Y ˜ J P X Y ) ,
H ( Y ˜ n ) + D ( P X ˜ n Y ˜ n P X Y n ) n H ( Y ˜ J ) + D ( P Y J P Y ) ,
H ( Y ˜ n | M ˜ 2 ) = n H ( Y ˜ J | V J , J ) ,
H ( Y ˜ n | M ˜ 1 , X ˜ n ) n H ( Y ˜ J | X J , U J , J ) .
Let U : = ( U J , J ) , V : = ( V J , J ) , X : = X ˜ J , Y : = Y ˜ J and Z : = Z ˜ J . Using the joint distribution P X ˜ n Y ˜ n Z ˜ n M ˜ 1 M ˜ 2 in (64), we conclude that the joint distribution of random variables ( X , Y , Z , U , V ) , denoted by Q X Y Z U V , belongs to the set Q 1 defined in (115). The proof of Lemma 1 is complete by combining (A21) to (A38) and noting that I Q ( X , Y ; U ) I Q ( X ; U ) .

Appendix C. Proof of Lemma 2

Given any γ R + , let Q X Y Z U V ( γ ) achieve the minimum in (117). Recall the definition of θ n in (51) and define a new alphabet V ˜ : = V { v * } . We then define a joint distribution P Y V ˜ ( γ ) by specifying the following (conditional) marginal distributions
P V ˜ ( γ ) ( v ) : = 1 1 + θ n Q V ( γ ) ( v ) 1 { v v * } + θ n 1 + θ n 1 { v = v * } ,
P Y | V ˜ ( γ ) ( y | v ) : = Q Y | V ( γ ) ( y | v ) 1 { v v * } + 1 + θ n θ n P Y ( y ) 1 θ n Q Y ( γ ) ( y ) 1 { v = v * } .
Thus, the induced marginal distribution P Y ( γ ) satisfies
P Y ( γ ) ( y ) = v V ˜ P V ˜ ( γ ) ( v ) P Y | V ˜ ( γ ) ( y | v )
= v V 1 1 + θ n Q V ( γ ) ( v ) Q Y | V ( γ ) ( y | v ) + P Y ( y ) 1 1 + θ n Q Y ( γ ) ( y )
= P Y ( y ) .
Furthermore, let P V ˜ | Y ( γ ) be induced by P Y V ˜ ( γ ) and define the following distribution
P X Y Z U V ˜ ( γ ) = P X Y Z Q U | X ( γ ) P V ˜ | Y ( γ ) .
Recall the definition of R b , c , d ( · ) in (28). The following lemma lower bounds the difference between R b , c , d ( Q X Y Z U V ( γ ) ) and R b , c , d ( P X Y Z U V ˜ ( γ ) ) and is critical in the proof of Lemma 2.
Lemma A1.
When γ = n , we have
R b , c , d ( Q X Y Z U V ( γ ) ) R b , c , d ( P X Y Z U V ˜ ( γ ) ) Θ log n n 1 / 4 .
The proof of Lemma A1 is deferred to Appendix D.
Now, using the assumption that Q X Y Z U V ( γ ) is a minimizer for R b , c , d , γ in (117), the fact that Δ b , d , γ ( Q X Y Z U V ( γ ) ) 0 (see (116)) and the result in (A45), we conclude that when γ = n ,
R b , c , d , γ = R b , c , d ( Q X Y Z U V ( γ ) ) + Δ b , d , γ ( Q X Y Z U V ( γ ) )
R b , c , d ( P X Y Z U V ˜ ( γ ) ) + Θ log n n 1 / 4
R b , c , d + Θ log n n 1 / 4 ,
where (A48) follows from the definition of R b , c , d in (29) and the fact that P X Y Z U V ˜ ( γ ) Q (see (24)).
The proof of Lemma 2 is complete by using (A48) and noting that when γ = n ,
log 1 ε 1 ε 2 1 + 3 ε 2 ε 1 + ( b + d + γ + 5 ) log 1 ε 1 ε 2 2 = Θ ( n ) .

Appendix D. Proof of Lemma A1

In subsequent analyses, all distributions indicated by P ( γ ) are induced by P X Y Z U V ˜ ( γ ) . We have
D ( Q X Y U ( γ ) P X Y U ( γ ) ) = D ( Q X Y ( γ ) P X Y ( γ ) ) + I Q ( γ ) ( U ; Y | X ) .
Recalling the definitions of R b , c , d in (29) and R b , c , d , γ in (117), we conclude that for any γ R + ,
R b , c , d , γ R b , c , d b log | X | + d log | Y | = : a .
Using the definition of Δ b , d , γ ( Q X Y Z U V ) in (116) and recalling that Q X Y Z U V ( γ ) is a minimizer for R b , c , d , γ , we have
γ D ( Q X Y U ( γ ) P X Y U ( γ ) ) Δ b , d , γ ( Q X Y Z U V ( γ ) )
= R b , c , d , γ R b , c , d ( Q X Y Z U V ( γ ) )
a + ( c + 1 ) log | Y | + c log | Z | = : a .
We can now upper bound I P γ ( V ˜ ; Y ) as follows:
I P ( γ ) ( V ˜ ; Y ) = D ( P Y | V ˜ ( γ ) P Y ( γ ) | P V ˜ ( γ ) )
= D ( P Y | V ˜ ( γ ) P Y | P V ˜ ( γ ) )
= 1 1 + θ n D ( Q Y | V ( γ ) P Y | Q V ( γ ) ) + θ n 1 + θ n D 1 + θ n θ n P Y 1 θ n Q Y ( γ ) P Y
= 1 1 + θ n D ( Q Y | V ( γ ) Q Y ( γ ) | Q V ( γ ) ) + D ( Q Y ( γ ) P Y ) + θ n 1 + θ n D 1 + θ n θ n P Y 1 θ n Q Y ( γ ) P Y
1 1 + θ n I Q ( γ ) ( V ; Y ) + 1 1 + θ n a γ + θ n 1 + θ n log μ ,
where (A56) follows from (A43), and (A59) follows from the result in (A54), the fact that D ( Q Y ( γ ) P Y ) D ( Q X Y U ( γ ) P X Y U ( γ ) ) and the definition of μ in (50). Thus, when γ = n , recalling the definition of θ n in (51), we have
I Q ( γ ) ( V ; Y ) I P ( γ ) ( V ˜ ; Y ) a γ θ n log μ
= I P ( γ ) ( V ˜ ; Y ) + Θ 1 n .
Similar to (A59), we obtain
I P ( γ ) ( V ˜ ; Z )
= D ( P Z | V ˜ ( γ ) P Z ( γ ) | P V ˜ ( γ ) )
= D ( P Z | V ˜ ( γ ) P Z | P V ˜ ( γ ) )
= 1 1 + θ n D ( Q Z | V ( γ ) P Z | Q V ( γ ) ) + θ n 1 + θ n D 1 + θ n θ n P Z 1 θ n Q Z ( γ ) P Z
= 1 1 + θ n D ( Q Z | V ( γ ) Q Z ( γ ) | Q V ( γ ) ) + D ( Q Z ( γ ) P Z ) + θ n 1 + θ n D 1 + θ n θ n P Z 1 θ n Q Z ( γ ) P Z
1 1 + θ n I Q ( γ ) ( V ; Z ) ,
where (A64) follows since Q ( γ ) Q 1 (see (115)) implies that Q Z | Y ( γ ) = P Z | Y and the Markov chains Z Y X and V Y Z holds and thus using (A39) to (A40), we have
P Z | V ˜ ( γ ) ( z | v ) = y P Z | Y ( z | y ) P V ˜ ( v ) P Y | V ˜ ( y | v ) P V ˜ ( v )
= y Q Z | Y ( γ ) ( z | y ) Q V ( γ ) ( v ) Q Y | V ( γ ) ( y | v ) Q V ( γ ) ( v )
= Q Z | V ( γ ) ( z | v ) ,
and
P Z | V ˜ ( γ ) ( z | v * ) = y P Z | Y ( z | y ) P V ˜ ( v * ) P Y | V ˜ ( y | v * ) P V ˜ ( v * )
= y Q Z | Y ( γ ) ( z | y ) 1 + θ n θ n P Y ( y ) 1 θ n Q Y ( γ ) ( y )
= 1 + θ n θ n P Z ( z ) 1 θ n Q Z ( γ ) ( z ) ,
Therefore, we have
I Q ( γ ) ( V ; Z ) ( 1 + θ n ) I P ( γ ) ( V ˜ ; Z )
I P ( γ ) ( V ˜ ; Z ) + θ n log | Z |
= I P ( γ ) ( V ˜ ; Z ) + Θ 1 n .
Let P Q be the 1 norm between P and Q regarded as vectors. Using Pinsker’s inequality, the result in (105), and the data processing inequality for the relative entropy [17], we obtain
Q U X ( γ ) P U X ( γ ) 2 log 2 · D ( Q U X ( γ ) P U X ( γ ) )
2 log 2 · D ( Q X Y U ( γ ) P X Y U ( γ ) )
2 a log 2 γ .
From the support lemma ([21] (Appendix C)), we conclude that the cardinality of U can be upper bounded by a function depending only on | X | , | Y | and | Z | (these alphabets are all finite). Thus, when γ = n , invoking ([4] (Lemma 2.2.7)), we have
| H ( Q U X ( γ ) ) H ( P U X ( γ ) ) | 2 a log 2 γ log | U | | X | 2 a log 2 γ = Θ log n n 1 / 4 .
Similar to (A79), we have
| I Q ( γ ) ( U ; X ) I P ( γ ) ( U ; X ) | Θ log n n 1 / 4 ,
| I Q ( γ ) ( U ; Y ) I P ( γ ) ( U ; Y ) | Θ log n n 1 / 4 .
Combining (A61), (A75), (A80) and (A81), when γ = n , using the definition of R b , c , d ( · ) in (28), we have
R b , c , d ( Q X Y Z U V ( γ ) )
( c + 1 ) I Q ( γ ) ( U ; Y ) + b I Q ( γ ) ( U ; X ) c I Q ( γ ) ( V ; Z ) + d I Q ( γ ) ( V ; Y )
( c + 1 ) I P ( γ ) ( U ; Y ) + b I P ( γ ) ( U ; X ) c I P ( γ ) ( V ˜ ; Z ) + d I P ( γ ) ( V ˜ ; Y ) + Θ log n n 1 / 4
= R b , c , d ( P X Y Z U V ˜ ( γ ) ) + Θ log n n 1 / 4 .
The proof of Lemma A1 is now complete.

References

  1. Ahlswede, R.; Csiszár, I. Hypothesis testing with communication constraints. IEEE Trans. Inf. Theory 1986, 32, 533–542. [Google Scholar] [CrossRef]
  2. Han, T. Hypothesis testing with multiterminal data compression. IEEE Trans. Inf. Theory 1987, 33, 759–772. [Google Scholar] [CrossRef]
  3. Shalaby, H.M.; Papamarcou, A. Multiterminal detection with zero-rate data compression. IEEE Trans. Inf. Theory 1992, 38, 254–267. [Google Scholar] [CrossRef]
  4. Csiszár, I.; Körner, J. Information Theory: Coding Theorems for Discrete Memoryless Systems; Cambridge University Press: Cambridge, UK, 2011. [Google Scholar]
  5. Han, T.S.; Amari, S. Statistical inference under multiterminal data compression. IEEE Trans. Inf. Theory 1998, 44, 2300–2324. [Google Scholar]
  6. Tian, C.; Chen, J. Successive refinement for hypothesis testing and lossless one-helper problem. IEEE Trans. Inf. Theory 2008, 54, 4666–4681. [Google Scholar] [CrossRef]
  7. Wigger, M.; Timo, R. Testing against independence with multiple decision centers. In Proceedings of the IEEE SPCOM, Bengaluru, India, 12–15 June 2016; pp. 1–5. [Google Scholar]
  8. Zhao, W.; Lai, L. Distributed detection with vector quantizer. IEEE Trans. Signal Inf. Process. Netw. 2016, 2, 105–119. [Google Scholar] [CrossRef]
  9. Xiang, Y.; Kim, Y.H. Interactive hypothesis testing with communication constraints. In Proceedings of the IEEE 50th Annual Allerton on Communication, Control, and Computing, Monticello, IL, USA, 1–5 October 2012; pp. 1065–1072. [Google Scholar]
  10. Zhao, W.; Lai, L. Distributed testing with cascaded encoders. IEEE Trans. Inf. Theory 2018, 64, 7339–7348. [Google Scholar] [CrossRef]
  11. Salehkalaibar, S.; Wigger, M.; Wang, L. Hypothesis Testing Over the Two-Hop Relay Network. IEEE Trans. Inf. Theory 2019, 65, 4411–4433. [Google Scholar] [CrossRef]
  12. Liu, J.; van Handel, R.; Verdú, S. Beyond the blowing-up lemma: Sharp converses via reverse hypercontractivity. In Proceedings of the 2017 IEEE International Symposium on Information Theory (ISIT), Aachen, Germany, 25–30 June 2017; pp. 943–947. [Google Scholar]
  13. Tyagi, H.; Watanabe, S. Strong Converse using Change of Measure. IEEE Trans. Inf. Theory 2020, in press. [Google Scholar]
  14. Oohama, Y. Exponential strong converse for source coding with side information at the decoder. Entropy 2018, 20, 352. [Google Scholar] [CrossRef]
  15. Oohama, Y. Exponent function for one helper source coding problem at rates outside the rate region. In Proceedings of the 2015 IEEE International Symposium on Information Theory (ISIT), Hong Kong, China, 14–19 June 2015; pp. 1575–1579. [Google Scholar]
  16. Oohama, Y. Exponential Strong Converse for One Helper Source Coding Problem. Entropy 2019, 21, 567. [Google Scholar] [CrossRef]
  17. Cover, T.M.; Thomas, J.A. Elements of Information Theory; John Wiley & Sons: New York, NY, USA, 2012. [Google Scholar]
  18. Gu, W.; Effros, M. A strong converse for a collection of network source coding problems. In Proceedings of the 2009 IEEE International Symposium on Information Theory, Seoul, Korea, 28 June–3 July 2009; pp. 2316–2320. [Google Scholar] [CrossRef]
  19. Liu, J.; van Handel, R.; Verdú, S. Beyond the Blowing-Up Lemma: Optimal Second-Order Converses via Reverse Hypercontractivity. Available online: http://web.mit.edu/jingbo/www/preprints/msl-blup.pdf (accessed on 17 April 2019).
  20. Salehkalaibar, S.; Wigger, M.; Wang, L. Hypothesis testing over cascade channels. In Proceedings of the 2017 IEEE Information Theory Workshop (ITW), Kaohsiung, Taiwan, 6–10 November 2017; pp. 369–373. [Google Scholar] [CrossRef]
  21. El Gamal, A.; Kim, Y.H. Network Information Theory; Cambridge University Press: Cambridge, UK, 2011. [Google Scholar]
  22. Raginsky, M.; Sason, I. Concentration of measure inequalities in information theory, communications, and coding. Found. Trends® Commun. Inf. Theory 2013, 10, 1–246. [Google Scholar] [CrossRef]
  23. Gilani, A.; Amor, S.B.; Salehkalaibar, S.; Tan, V.Y.F. Distributed Hypothesis Testing with Privacy Constraints. Entropy 2019, 21, 478. [Google Scholar] [CrossRef] [Green Version]
Figure 1. System model for hypothesis testing over a two-hop network
Figure 1. System model for hypothesis testing over a two-hop network
Entropy 21 01171 g001
Figure 2. Illustration of the proof sketch of Theorem 3.
Figure 2. Illustration of the proof sketch of Theorem 3.
Entropy 21 01171 g002

Share and Cite

MDPI and ACS Style

Cao, D.; Zhou, L.; Tan, V.Y.F. A Strong Converse Theorem for Hypothesis Testing Against Independence over a Two-Hop Network. Entropy 2019, 21, 1171. https://doi.org/10.3390/e21121171

AMA Style

Cao D, Zhou L, Tan VYF. A Strong Converse Theorem for Hypothesis Testing Against Independence over a Two-Hop Network. Entropy. 2019; 21(12):1171. https://doi.org/10.3390/e21121171

Chicago/Turabian Style

Cao, Daming, Lin Zhou, and Vincent Y. F. Tan. 2019. "A Strong Converse Theorem for Hypothesis Testing Against Independence over a Two-Hop Network" Entropy 21, no. 12: 1171. https://doi.org/10.3390/e21121171

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop