Next Article in Journal
A New Simple Method for Estimating Exergy Destruction in Heat Exchangers
Previous Article in Journal
All-Optically Controlled Quantum Memory for Light with a Cavity-Optomechanical System
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Wiretap Channel with Action-Dependent Channel State Information

1
School of Information Science and Technology, Southwest JiaoTong University, Northbound Section Second Ring Road 111, Chengdu, China
2
Computer Science and Engineering Department, Shanghai Jiao Tong University, Dongchuan Road 800, Shanghai, China
3
Institute for Experimental Mathematics, Duisburg-Essen University, Ellernstr.29, Essen, Germany
*
Author to whom correspondence should be addressed.
Entropy 2013, 15(2), 445-473; https://doi.org/10.3390/e15020445
Submission received: 23 November 2012 / Revised: 9 January 2013 / Accepted: 17 January 2013 / Published: 28 January 2013

Abstract

:
In this paper, we investigate the model of wiretap channel with action-dependent channel state information. Given the message to be communicated, the transmitter chooses an action sequence that affects the formation of the channel states, and then generates the channel input sequence based on the state sequence and the message. The main channel and the wiretap channel are two discrete memoryless channels (DMCs), and they are connected with the legitimate receiver and the wiretapper, respectively. Moreover, the transition probability distribution of the main channel depends on the channel state. Measuring wiretapper’s uncertainty about the message by equivocation, inner and outer bounds on the capacity-equivocation region are provided both for the case where the channel inputs are allowed to depend non-causally on the state sequence and the case where they are restricted to causal dependence. Furthermore, the secrecy capacities for both cases are bounded, which provide the best transmission rate with perfect secrecy. The result is further explained via a binary example.

1. Introduction

Communication through state-dependent channels, with states known at the transmitter, was first investigated by Shannon [1] in 1958. In [1], the capacity of the discrete memoryless channel with causal (past and current) channel state information at the encoder was totally determined. After that, in order to solve the problem of coding for a computer memory with defective cells, Kuznetsov and Tsybakov [2] considered a channel in the presence of non-causal channel state information at the transmitter. They provided some coding techniques without determination of the capacity. The capacity was found in 1980 by Gel’fand and Pinsker [3]. Furthermore, Costa [4] investigated a power constrained additive noise channel, where part of the noise is known at the transmitter as side information. This channel is also called dirty paper channel. The assumption in these seminar papers, as well as in the work on communication with state dependent channels that followed, is that the channel states are generated by nature, and can not be affected or controlled by the communication system.
In 2009, Weissman [5] revisited the above problem setting for the case where the transmitter can take actions that affect the formation of the states, see Figure 1. Specifically, Weissman considered a communication system where encoding is in two parts: given the message, an action sequence is created. The actions affect the formation of the channel states, which are accessible to the transmitter when producing the channel input sequence. The capacity of this model is totally determined both for the case where the channel inputs are allowed to depend non-causally on the state sequence and the case where they are restricted to causal dependence. This framework captures various new channel coding scenarios that may arise naturally in recording for magnetic storage devices or coding for computer memories with defects.
Figure 1. Channel with action-dependent states.
Figure 1. Channel with action-dependent states.
Entropy 15 00445 g001
Transmission of confidential messages has been studied in the literature of several classes of channels. Wyner, in his well-known paper on the wiretap channel [6], studied the problem how to transmit the confidential messages to the legitimate receiver via a degraded broadcast channel, while keeping the wiretapper as ignorant of the messages as possible, see Figure 2. Measuring the uncertainty of the wiretapper by equivocation, the capacity-equivocation region was established. Furthermore, the secrecy capacity was also established, which provided the maximum transmission rate with perfect secrecy. After the publication of Wyner’s work, Csisz a ´ r and Körner [7] investigated a more general situation: the broadcast channels with confidential messages (BCC). In this model, a common message and a confidential message were sent through a general broadcast channel. The common message was assumed to be decoded correctly by the legitimate receiver and the wiretapper, while the confidential message was only allowed to be obtained by the legitimate receiver. This model is also a generalization of [8], where no confidentiality condition is imposed. The capacity-equivocation region and the secrecy capacity region of BCC [7] were totally determined, and the results were also a generalization of those in [6]. Based on Wyner’s work, Leung-Yan-Cheong and Hellman studied the Gaussian wiretap channel(GWC) [9], and showed that its secrecy capacity was the difference between the main channel capacity and the overall wiretap channel capacity (the cascade of main channel and wiretap channel).
Figure 2. Wiretap channel.
Figure 2. Wiretap channel.
Entropy 15 00445 g002
Inspired by the above works, Mitrpant et al. [10] studied transmission of confidential messages in the channels with channel state information (CSI). In [10], an inner bound on the capacity-equivocation region was provided for the Gaussian wiretap channel with CSI. Furthermore, Chen et al. [11] investigated the discrete memoryless wiretap channel with noncausal CSI (see Figure 3), and also provided an inner bound on the capacity-equivocation region. Note that the coding scheme of [11] is a combination of those in [3,6] Based on the work of [11], Dai [12] provided an outer bound on the wiretap channel with noncausal CSI, and determined the capacity-equivocation region for the model of wiretap channel with memoryless CSI, where the memoryless means that at the i-th time, the output of the channel encoder depends only on the i-th time CSI.
Figure 3. Wiretap channel with noncausal channel state information.
Figure 3. Wiretap channel with noncausal channel state information.
Entropy 15 00445 g003
In this paper, we study the wiretap channel with action-dependent channel state information, see Figure 4. Concretely, the transmitted message W is firstly encoded as an action sequence A N , and A N is the input of a discrete memoryless channel (DMC). The output of this DMC is the channel state sequence S N . Then, the transmitted message W and the state sequence S N are encoded as X N . The main channel is a DMC with inputs X N and S N , and output Y N . The wiretap channel is also a DMC with input Y N and output Z N . Since the action-dependent state captures various new coding scenarios for channels with a rewrite option that may arise naturally in storage for computer memories with defects or in magnetic recoding, it is natural to ask: how about the security of these channel models in the presence of a wiretapper? Measuring wiretapper’s uncertainty about the transmitted message by equivocation, the inner and outer bounds on the capacity-equivocation region of the model of Figure 4 are provided both for the case where the channel input is allowed to depend non-causally on the state sequence and the case where it is restricted to causal dependence.
Figure 4. Wiretap channel with action-dependent channel state information.
Figure 4. Wiretap channel with action-dependent channel state information.
Entropy 15 00445 g004
In this paper, random variables, sample values and alphabets are denoted by capital letters, lower case letters and calligraphic letters, respectively. A similar convention is applied to the random vectors and their sample values. For example, U N denotes a random N-vector ( U 1 , . . . , U N ) , and u N = ( u 1 , . . . , u N ) is a specific vector value in U N that is the Nth Cartesian power of U . U i N denotes a random N i + 1 -vector ( U i , . . . , U N ) , and u i N = ( u i , . . . , u N ) is a specific vector value in U i N . Let p V ( v ) denote the probability mass function P r { V = v } . Throughout the paper, the logarithmic function is to the base 2.
The remainder of this paper is organized as follows. In Section 2, we present the basic definitions and the main result on the capacity-equivocation region of wiretap channel with action-dependent channel state information. In Section 3, we provide a binary example of the model of Figure 4. Final conclusions are presented in Section 4.

2. Notations, Definitions and the Main Results

In this section, the model of Figure 4 is considered into two parts. The model of Figure 4 with noncausal channel state information is described in Subsection 2.1, and the causal case is described in Subsection 2.2, see the followings.

2.1. The Model of Figure 4 with Noncausal Channel State Information

In this subsection, a description of the wiretap channel with noncausal action-dependent channel state information is given by Definition 1 to Definition 6. The inner and outer bounds on the capacity-equivocation region C n composed of all achievable ( R , R e ) pairs are given in Theorem 1 and Theorem 2, respectively, where the achievable ( R , R e ) pair is defined in Definition 6.
Definition 1
(Action encoder) The message W take values in W , and it is uniformly distributed over its range. The action encoder is a deterministic mapping:
f 1 N : W A N
The input of the action encoder is W, while the output is A N .
The channel state sequence S N is generated by a DMC with input A N and output S N . The transition probability distribution is given by
p S N | A N ( s N | a N ) = i = 1 N p S i | A i ( s i | a i )
Note that the components of the state sequence S N may not be i.i.d. random variables, and this is due to the fact that A N is not i.i.d. generated.
The transmission rate of the message is log W N .
Definition 2
(Channel encoder) The inputs of the channel encoder are W and S N , while the output is X N . The channel encoder f 2 N is a matrix of conditional probabilities f 2 N ( x N | w , s N ) , where x N X N , w W , s N S N , x N f 2 N ( x N | w , s N ) = 1 , and f 2 N ( x N | w , s N ) is the probability that the message w and the channel state sequence s N are encoded as the channel input x N .
Since the channel encoder knows the state sequence s N in a noncausal manner, at the i-th time ( 1 i N ), the channel encoder f 2 , i N is a matrix of conditional probabilities f 2 , i N ( x i | w , s N ) , where x i X , w W , s N S N , x i f 2 , i N ( x i | w , s N ) = 1 , and f 2 , i N ( x i | w , s N ) is the probability that the message w and the channel state sequence s N are encoded as the i-th time channel input x i .
The transmission rate of the message is log W N .
Definition 3
(Main channel) The main channel is a DMC with finite input alphabet X × S , finite output alphabet Y , and transition probability Q M ( y | x , s ) , where x X , s S , y Y . Q M ( y N | x N , s N ) = n = 1 N Q M ( y n | x n , s n ) . The inputs of the main channel are X N and S N , while the output is Y N .
Definition 4
(Wiretap channel) The wiretap channel is also a DMC with finite input alphabet Y , finite output alphabet Z , and transition probability Q W ( z | y ) , where y Y , z Z . The input and output of the wiretap channel are Y N and Z N , respectively. The equivocation to the wiretapper is defined as
Δ = H ( W | Z N ) N
The cascade of the main channel and the wiretap channel is another DMC with transition probability
Q M W ( z | x , s ) = y Y Q W ( z | y ) Q M ( y | x , s )
Note that, ( X N , S N ) Y N Z N and W A N S N are two Markov chains in the model of Figure 4.
Definition 5
(Decoder) The decoder for the legitimate receiver is a mapping f D 1 : Y N W , with input Y N and output W ^ . Let P e be the error probability of the receiver , and it is defined as P r { W W ^ } .
Definition 6
(Achievable ( R , R e ) pair in the model of Figure 4) A pair ( R , R e ) (where R , R e > 0 ) is called achievable if, for any ϵ > 0 (where ϵ is an arbitrary small positive real number and ϵ 0 ), there exists channel encoders-decoders ( N , P e ) such that
lim N log W N = R , lim N Δ R e P e ϵ
The capacity-equivocation region R n is a set composed of all achievable ( R , R e ) pairs. Inner and outer bounds on R n are respectively provided in the following Theorem 1 and Theorem 2. Theorem 1 and Theorem 2 are respectively proved in Section A and Section B.
Theorem 1
(Inner bound) A single-letter characterization of the region R n i is as follows,
R ( n i ) = { ( R , R e ) : 0 R e R R I ( U ; Y ) I ( U ; S | A ) R e I ( U ; Y ) I ( U ; Z ) R e H ( A | Z ) }
where p U A S X Y Z ( u , a , s , x , y , z ) = p Z | Y ( z | y ) p Y | X , S ( y | x , s ) p U A X S ( u , a , x , s ) , which implies that ( A , U ) ( X , S ) Y Z .
The region R ( n i ) satisfies R ( n i ) R ( n ) .
Remark 1
There are some notes on Theorem 1, see the following.
  • The formula R e H ( A | Z ) in Theorem 1 implies that the wiretapper obtains the information about the message not only from the codeword transmitted in the channels, but also from the action sequence a N . If the wiretapper knows a N , he knows the corresponding message.
  • The region R ( n i ) is convex, and the proof is directly obtained by introducing a time sharing random variable into Theorem 1, and therefore, we omit the proof here.
  • The range of the random variable U satisfies
    U X A S + 2
    The proof is in Section C.
  • Without the equivocation parameter, the capacity of the main channel is given by
    C M = max p X | U , S ( x | u , s ) p U | A , S ( u | a , s ) p A ( a ) ( I ( U ; Y ) I ( U ; S | A ) )
    The formula (6) is proved by Weissman [5], and it is omitted here.
  • Secrecy capacity
    The points in R ( n i ) for which R e = R are of considerable interest, which imply the perfect secrecy H ( W ) = H ( W | Z N ) . Clearly, we can easily bound the secrecy capacity C s n of the model of Figure 4 with noncausal channel state information by
    C s n max p U A X S ( u , a , x , s ) min { I ( U ; Y ) I ( U ; Z ) , I ( U ; Y ) I ( U ; S | A ) , H ( A | Z ) }
    Proof 1 (Proof of (7)) Substituting R e = R into the region R ( n i ) in Theorem 1, we have
    R I ( U ; Y ) I ( U ; Z )
    R I ( U ; Y ) I ( U ; S | A )
    R H ( A | Z )
    Note that the pair ( R = max min { I ( U ; Y ) I ( U ; Z ) , I ( U ; Y ) I ( U ; S | A ) , H ( A | Z ) } , R e = R ) is achievable, and therefore, the secrecy capacity C s ( n ) max min { I ( U ; Y ) I ( U ; Z ) , I ( U ; Y ) I ( U ; S | A ) , H ( A | Z ) } . Thus the proof is completed.
Theorem 2
(Outer bound) A single-letter characterization of the region R n o is as follows,
R ( n o ) = { ( R , R e ) : 0 R e R R I ( U ; Y ) I ( U ; S | A ) R e I ( U ; Y ) I ( K ; Z | V ) }
where p U K V A S X Y Z ( u , k , v , a , s , x , y , z ) = p Z | Y ( z | y ) p Y | X , S ( y | x , s ) p X | U , S ( x | u , s ) p V | K ( v | k ) p K | U ( k | u ) p U , A , S ( u , a , s ) , which implies that ( A , U , K , V ) ( X , S ) Y Z and V K U Y Z are two Markov chains.
The region R ( n o ) satisfies R ( n ) R ( n o ) .
Remark 2
There are some notes on Theorem 2, see the following.
  • The region R ( n o ) is convex, and the proof is similar to that of Theorem 1. Therefore, we omit the proof here.
  • The ranges of the random variables U, V and K satisfy
    U X A S + 1
    V X A S
    K X 2 A 2 S 2
    The proof is in Section D.
  • Observing the formula R e I ( U ; Y ) I ( K ; Z | V ) in Theorem 2, we have
    I ( U ; Y ) I ( K ; Z | V ) = ( a ) I ( U ; Y ) H ( Z | V ) + H ( Z | K ) I ( U ; Y ) H ( Z ) + H ( Z | K ) I ( U ; Y ) H ( Z ) + H ( Z | K , U ) = ( b ) I ( U ; Y ) H ( Z ) + H ( Z | U ) = I ( U ; Y ) I ( U ; Z )
    where (a) is from the fact that V K Z , and (b) is from the Markov chain K U Y Z . Then it is easy to see that R ( n i ) R ( n o ) .
  • The secrecy capacity C s n of the model of Figure 4 with noncausal channel state information is upper bounded by
    C s n max p X , U , K , V , A , S ( x , u , k , v , a , s ) min { I ( U ; Y ) I ( K ; Z | V ) , I ( U ; Y ) I ( U ; S | A ) }
    The upper bound is easily obtained by substituting R e = R into the region R ( n o ) in Theorem 2, and therefore, we omit the proof here.

2.2. The Model of Figure 4 with Causal Channel State Information

The model of Figure 4 with causal channel state information is similar to the model with noncausal channel state information in Subsection 2.1, except that the state sequence S N in Definition 1 is known to the channel encoder in a causal manner, i.e., at the i-th time ( 1 i N ), the output of the encoder x i = f 2 , i ( w , s i ) , where s i = ( s 1 , s 2 , . . . , s i ) and f 2 , i is the probability that the message w and the state s i are encoded as the channel input x i at time i. Define
f N ( x N | w , s N ) = i = 1 N f i ( x i | w , s i )
Inner and outer bounds on the capacity-equivocation region R c for the model of Figure 4 with causal channel state information are respectively provided in the following Theorem 3 and Theorem 4, and they are proved in Section E and Section F.
Theorem 3
(Inner bound) A single-letter characterization of the region R c i is as follows,
R ( c i ) = { ( R , R e ) : 0 R e R R I ( U ; Y ) R e I ( U ; Y ) I ( U ; Z ) R e H ( A | Z ) }
where p U A S X Y Z ( u , a , s , x , y , z ) = p Z | Y ( z | y ) p Y | X , S ( y | x , s ) p X | U , S ( x | u , s ) p S | A ( s | a ) p U A ( u , a ) , which implies that ( A , U ) ( X , S ) Y Z and U A S .
The region R ( c i ) satisfies R ( c i ) R ( c ) .
Remark 3
There are some notes on Theorem 3, see the following.
  • The region R ( c i ) is convex.
  • The range of the random variable U satisfies
    U X A S + 1
    The proof is similar to that in Theorem 1, and it is omitted here.
  • Without the equivocation parameter, the capacity of the main channel is given by
    C M * = max p X | U , S ( x | u , s ) p U , A ( u , a ) I ( U ; Y )
    The formula (14) is proved by Weissman [5], and it is omitted here.
  • Secrecy capacity
    The points in R ( c i ) for which R e = R are of considerable interest, which imply the perfect secrecy H ( W ) = H ( W | Z N ) . Clearly, we can easily bound the secrecy capacity C s ( c ) of the model of Figure 4 with causal channel state information by
    C s ( c ) max p X | U , S ( x | u , s ) p U A ( u , a ) min { I ( U ; Y ) I ( U ; Z ) , H ( A | Z ) }
    Proof 2 (Proof of (15)) Substituting R e = R into the region R ( c i ) in Theorem 3, we have
    R I ( U ; Y ) I ( U ; Z )
    R I ( U ; Y )
    R H ( A | Z )
    Note that the pair ( R = max min { I ( U ; Y ) I ( U ; Z ) , H ( A | Z ) } , R e = R ) is achievable, and therefore, the secrecy capacity C s ( c ) max min { I ( U ; Y ) I ( U ; Z ) , H ( A | Z ) } . Thus the proof is completed.
Theorem 4
(Outer bound) A single-letter characterization of the region R c o is as follows,
R ( c o ) = { ( R , R e ) : 0 R e R R I ( U ; Y ) R e I ( U ; Y ) I ( K ; Z | V ) }
where p U K V A S X Y Z ( u , k , v , a , s , x , y , z ) = p Z | Y ( z | y ) p Y | X , S ( y | x , s ) p X | U , S ( x | u , s ) p V | K ( v | k ) p K | U ( k | u ) p U , A , S ( u , a , s ) , which implies that ( A , U , K , V ) ( X , S ) Y Z and V K U Y Z are two Markov chains.
The region R ( c o ) satisfies R ( c ) R ( c o ) .
Remark 4 There are some notes on Theorem 4, see the following.
  • The region R ( c o ) is convex.
  • The ranges of the random variables U, V and K satisfy
    U X A S
    V X A S
    K X 2 A 2 S 2
    The proof is similar to that in Section D, and it is omitted here.
  • The secrecy capacity C s ( c ) of the model of Figure 4 with causal channel state information is upper bounded by
    C s ( c ) max p X | U , S ( x | u , s ) p U , K , V , A , S ( u , k , v , a , s ) I ( U ; Y ) I ( K ; Z | V )
    The upper bound is easily obtained by substituting R e = R into the region R ( c o ) in Theorem 4, and therefore, we omit the proof here.

3. A Binary Example for the Model of Figure 4 with Causal Channel State Information

In this section, we calculate the bound on secrecy capacity of a special case of the model of Figure 4 with causal channel state information.
Suppose that the channel state information S N is available at the channel encoder in a casual manner. Meanwhile, the random variables X, Y and Z take values in { 0 , 1 } , and the transition probability of the main channel is defined as follows:
When s = 0 ,
p Y | X , S ( y | x , s = 0 ) = 1 p , if y = x , p , otherwise .
When s = 1 ,
p Y | X , S ( y | x , s = 1 ) = p , if y = x , 1 p , otherwise .
The wiretap channel is a BSC (binary symmetric channel) with crossover probability q, i.e.,
p Z | Y ( z | y ) = 1 q , if y = x , q , otherwise .
The channel for generating the state sequence S N is a BSC with crossover probability r, i.e.,
p S | A ( s | a ) = 1 r , if y = x , r , otherwise .
From Remark 3 and Remark 4 we know that the secrecy capacity for the causal case is bounded by
max min { I ( U ; Y ) I ( U ; Z ) , H ( A | Z ) } C s c max ( I ( U ; Y ) I ( K ; Z | V ) ) ( a ) max I ( U ; Y ) .
Note that in (a), “=” is achieved if V = K . Moreover, max I ( U ; Y ) , max H ( A | Z ) and max ( I ( U ; Y ) I ( U ; Z ) ) are achieved if A is a function of U and X is a function of U and S, and this is similar to the argument in [5]. Define a = g ( u ) and x = f ( u , s ) , then (24) can be written as
max f , g , p U ( u ) min { I ( U ; Y ) I ( U ; Z ) , H ( A | Z ) } C s c max f , g , p U ( u ) I ( U ; Y )
and this is because the joint probability distribution p A U S X Y Z ( a , u , s , x , y , z ) can be calculated by
p A U S X Y Z ( a , u , s , x , y , z ) = p Z | Y ( z | y ) p Y | X , S ( y | x , s ) 1 x = f ( u , s ) p S | A ( s | a ) 1 a = g ( u ) p U ( u )
Now it remains to calculate the characters max f , g , p U ( u ) ( I ( U ; Y ) I ( U ; Z ) ) , max f , g , p U ( u ) H ( A | Z ) and max f , g , p U ( u ) I ( U ; Y ) , see the remaining of this section.
Let U take values in { 0 , 1 } . The probability of U is defined as follows. p U ( 0 ) = α and p U ( 1 ) = 1 α .
In addition, there are 16 kinds of f and 4 kinds of g. Define
f ( 1 ) ( u , s ) : 00 0 , 01 0 , 10 0 , 11 0 . f ( 2 ) ( u , s ) : 00 0 , 01 0 , 10 0 , 11 1 .
f ( 3 ) ( u , s ) : 00 0 , 01 0 , 10 1 , 11 0 . f ( 4 ) ( u , s ) : 00 0 , 01 0 , 10 1 , 11 1 .
f ( 5 ) ( u , s ) : 00 0 , 01 1 , 10 0 , 11 0 . f ( 6 ) ( u , s ) : 00 0 , 01 1 , 10 0 , 11 1 .
f ( 7 ) ( u , s ) : 00 0 , 01 1 , 10 1 , 11 0 . f ( 8 ) ( u , s ) : 00 0 , 01 1 , 10 1 , 11 1 .
f ( 9 ) ( u , s ) : 00 1 , 01 0 , 10 0 , 11 0 . f ( 10 ) ( u , s ) : 00 1 , 01 0 , 10 0 , 11 1 .
f ( 11 ) ( u , s ) : 00 1 , 01 0 , 10 1 , 11 0 . f ( 12 ) ( u , s ) : 00 1 , 01 0 , 10 1 , 11 1 .
f ( 13 ) ( u , s ) : 00 1 , 01 1 , 10 0 , 11 0 . f ( 14 ) ( u , s ) : 00 1 , 01 1 , 10 0 , 11 1 .
f ( 15 ) ( u , s ) : 00 1 , 01 1 , 10 1 , 11 0 . f ( 16 ) ( u , s ) : 00 1 , 01 1 , 10 1 , 11 1 .
g ( 1 ) ( u ) : 0 0 , 1 0 . g ( 2 ) ( u ) : 0 0 , 1 1 .
g ( 3 ) ( u ) : 0 1 , 1 0 . g ( 4 ) ( u ) : 0 1 , 1 1 .
The character I ( U ; Y ) depends on the joint probability mass functions p U Y ( u , y ) , and we have
p U Y ( u , y ) = x , s , a p U Y X S A ( u , y , x , s , a ) = x , s , a p Y | X S ( y | x , s ) p X | U , S ( x | u , s ) p U ( u ) p A | U ( a | u ) p S | A ( s | a )
The character I ( U ; Z ) depends on the joint probability mass functions p U Z ( u , z ) , and we have
p U Z ( u , z ) = y p U Y Z ( u , y , z ) = y p Z | Y ( z | y ) p U , Y ( u , y )
By choosing the above f, g and α, we find that
max f , g , p U ( u ) ( I ( U ; Y ) I ( U ; Z ) ) = max { h ( p q ) h ( p ) , h ( q ( r p ) ) h ( p r ) 2 r ( 1 2 r 1 ) ( h ( p q ) h ( p ) ) }
where p q = p + q 2 p q . Moreover, h ( p q ) h ( p ) is achieved when f = f ( 7 ) , g = g ( 2 ) and α = 1 2 , and h ( q ( r p ) ) h ( p r ) 2 r ( 1 2 r 1 ) ( h ( p q ) h ( p ) ) is achieved when f = f ( 2 ) , g = g ( 2 ) and α = 1 2 .
Moreover,
max f , g , p U ( u ) H ( A | Z ) = h ( p q )
where p q = p + q 2 p q . Moreover, h ( p q ) is achieved when f = f ( 7 ) , g = g ( 2 ) and α = 1 2 .
In addition,
max f , g , p U ( u ) I ( U ; Y ) = 1 h ( p )
and “=” is achieved if f = f ( 7 ) , g = g ( 2 ) and α = 1 2 .
It is easy to see that max f , g , p U ( u ) H ( A | Z ) = h ( p q ) max f , g , p U ( u ) ( I ( U ; Y ) I ( U ; Z ) ) and therefore, the secrecy capacity for the causal case is bounded by
max { h ( p q ) h ( p ) , h ( q ( r p ) ) h ( p r ) 2 r ( 1 2 r 1 ) ( h ( p q ) h ( p ) ) } C s c 1 h ( p )
The following Figure 5 gives lower and upper bounds on the secrecy capacity of the model of Figure 4 with causal channel state information. It is easy to see that when q = 0 . 5 , the lower bound meets with the upper bound, i.e., the secrecy capacity C s c satisfies C s c = 1 h ( p ) . This is because when q = 0 . 5 , zero leakage is always satisfied and the problem reduces to the problem of coding for channel with causal states. Moreover, when r is fixed, the bounds on secrecy capacity are getting better while p is decreasing.
Figure 5. When r=0.2, lower and upper bounds on the secrecy capacity of the model of Figure 4 with causal channel state information.
Figure 5. When r=0.2, lower and upper bounds on the secrecy capacity of the model of Figure 4 with causal channel state information.
Entropy 15 00445 g005

4. Conclusions

In this paper, we study the model of the wiretap channel with action-dependent channel state information. Inner and outer bounds on the capacity-equivocation region are provided both for the case where the channel inputs are allowed to depend non-causally on the state sequence and the case where they are restricted to causal dependence. Furthermore, the secrecy capacities for both cases are bounded, which provide the best transmission rate with perfect secrecy. The result is further explained via a binary’example.

Acknowledgement

The authors would like to thank N. Cai for his help to improve this paper. This work was supported by a sub-project in National Basic Research Program of China under Grant 2012CB316100 on Broadband Mobile Communications at High Speeds, the National Natural Science Foundation of China under Grant 61271222, and the Research Fund for the Doctoral Program of Higher Education of China (No. 20100073110016).

References

  1. Shannon, C.E. Channels with side information at the transmitter. IBM J. Res. Dev. 1958, 2, 289–293. [Google Scholar] [CrossRef]
  2. Kuznetsov, N.V.; Tsybakov, B.S. Coding in memories with defective cells. Probl. Peredachi Informatsii 1974, 10, 52–60. [Google Scholar]
  3. Gel’fand, S.I.; Pinsker, M.S. Coding for channel with random parameters. Problems. Control Inf. Theory 1980, 9, 19–31. [Google Scholar]
  4. Costa, M.H.M. Writing on dirty paper. IEEE Trans. Inf. Theory 1983, 29, 439–441. [Google Scholar] [CrossRef]
  5. Weissman, T. Capacity of channels with action-dependent states. IEEE Trans. Inf. Theory 2010, 56, 5396–5411. [Google Scholar] [CrossRef]
  6. Wyner, A.D. The wire-tap channel. Bell Syst. Tech. J. 1975, 54, 1355–1387. [Google Scholar] [CrossRef]
  7. Csisz´r, I.; Körner, J. Broadcast channels with confidential messages. IEEE Trans. Inf. Theory 1978, 24, 339–348. [Google Scholar]
  8. Körner, J.; Marton, K. General broadcast channels with degraded message sets. IEEE Trans. Inf. Theory 1977, 23, 60–64. [Google Scholar] [CrossRef]
  9. Leung-Yan-Cheong, S.K.; Hellman, M.E. The Gaussian wire-tap channel. IEEE Trans. Inf. Theory 1978, 24, 451–456. [Google Scholar] [CrossRef]
  10. Mitrpant, C.; Han Vinck, A.J.; Luo, Y. An achievable region for the gaussian wiretap channel with side information. IEEE Trans. Inf. Theory 2006, 52, 2181–2190. [Google Scholar] [CrossRef]
  11. Chen, Y.; Han Vinck, A.J. Wiretap channel with side information. IEEE Trans. Inf. Theory 2008, 54, 395–402. [Google Scholar] [CrossRef]
  12. Dai, B.; Luo, Y. Some new results on wiretap channel with side information. Entropy 2012, 14, 1671–1702. [Google Scholar] [CrossRef]
  13. Csisz´r, I.; Körner, J. Information Theory: Coding Theorems for Discrete Memoryless Systems; Academic: London, UK, 1981; pp. 123–124. [Google Scholar]

A. Proof of Theorem 1

In this section, we will show that any pair ( R , R e ) R n i is achievable. Gel’fand-Pinsker’s binning and Wyner’s random binning technique are used in the construction of the code-books.
Now the remainder of this section is organized as follows. The code construction is in Subsection A.1. The proof of achievability is given in Subsection A.2.

A.1. Code Construction

Since R e I ( U ; Y ) I ( U ; Z ) , R e H ( A | Z ) and R e R I ( U ; Y ) I ( U ; S | A ) , it is sufficient to show that the pair ( R , R e = min { I ( U ; Y ) max ( I ( U ; Z ) , I ( U ; S | A ) ) ) , H ( A | Z ) } is achievable, and note that this implies that R R e = min { I ( U ; Y ) max ( I ( U ; Z ) , I ( U ; S | A ) ) ) , H ( A | Z ) } .
The construction of the code and the proof of achievability are considered into two cases:
  • (Case 1) If H ( A | Z ) I ( U ; Y ) max ( I ( U ; Z ) , I ( U ; S | A ) ) ) , double binning technique [11] is used in the construction of the code-book.
  • (Case 2) If H ( A | Z ) I ( U ; Y ) max ( I ( U ; Z ) , I ( U ; S | A ) ) ) , Gel’fand-Pinsker’s binning technique [3] is used in the construction of the code-book.
  • (Code construction for Case 1)
    Given a pair ( R , R e ) , choose a joint probability mass function p U , A , S , X , Y , Z ( u , a , s , x , y , z ) such that
    0 R e R
    R I ( U ; Y ) I ( U ; S | A )
    R e = I ( U ; Y ) max ( I ( U ; Z ) , I ( U ; S | A ) ) )
    The message set W satisfies the following condition:
    lim N 1 N log W = R = I ( U ; Y ) I ( U ; S | A ) γ
    where γ is a fixed positive real numbers and
    0 γ ( a ) max ( I ( U ; Z ) , I ( U ; S | A ) ) ) I ( U ; S | A )
    Note that (a) is from R R e = I ( U ; Y ) max ( I ( U ; Z ) , I ( U ; S | A ) ) ) and (A1). Let W = { 1 , 2 , . . . , 2 N R } .
    Code-book generation:
    (Construction of A N )
    Generate 2 N R i.i.d. sequences a N , according to the probability mass function p A ( a ) . Index each sequence by i { 1 , 2 , . . . , 2 N R } . For a given message w ( w W ), choose a corresponding a N ( w ) as the output of the action encoder.
    (Construction of U N )
    For the transmitted action sequence a N ( w ) , generate 2 N ( I ( U ; Y ) ϵ 2 , N ) ( ϵ 2 , N 0 as N ) i.i.d. sequences u N , according to the probability mass function p U | A ( u i | a i ( w ) ) . Distribute these sequences at random into 2 N R = 2 N ( I ( U ; Y ) I ( U ; S | A ) γ ) bins such that each bin contains 2 N ( I ( U ; S | A ) + γ ϵ 2 , N ) sequences. Index each bin by i { 1 , 2 , . . . , 2 N R } . Then place the 2 N ( I ( U ; S | A ) + γ ϵ 2 , N ) sequences in every bin randomly into 2 N ( max ( I ( U ; S | A ) , I ( U ; Z ) ) I ( U ; Z ) + ϵ 3 , N ) ( ϵ 3 , N 0 as N ) subbins such that every subbin contains
    2 N ( I ( U ; S | A ) + γ ϵ 2 , N max ( I ( U ; S | A ) , I ( U ; Z ) ) + I ( U ; Z ) ϵ 3 , N ) sequences. Let J be the random variable to represent the index of the subbin. Index each subbin by
    j { 1 , 2 , . . . , 2 N ( max ( I ( U ; S | A ) , I ( U ; Z ) ) I ( U ; Z ) + ϵ 3 , N ) } , i.e.,
    log J = N ( max ( I ( U ; S | A ) , I ( U ; Z ) ) I ( U ; Z ) + ϵ 3 , N ) .
    Here note that the number of the sequences in every subbin is upper bounded as follows.
    I ( U ; S | A ) + γ ϵ 2 , N max ( I ( U ; S | A ) , I ( U ; Z ) ) + I ( U ; Z ) ϵ 3 , N ( a ) I ( U ; Z ) ϵ 2 , N ϵ 3 , N
    where (a) is from (A2). This implies that
    lim N H ( U N | W , J , Z N ) = 0 .
    Note that (A5) can be proved by using Fano’s inequality and (A4).
    Let s N be the state sequence generated in response to the action sequence a N ( w ) . For a given message w ( w W ) and channel state s N , try to find a sequence u N ( w , i * ) in bin w such that ( u N ( w , i * ) , a N ( w ) , s N ) T U A S N ( ϵ 2 ) . If multiple such sequences in bin w exist, choose the one with the smallest index in the bin. If no such sequence exists, declare an encoding error.
    Figure A1 shows the construction of U N for case 1, see the following.
    Figure A1. Code-book construction for U N in Theorem 1 for case 1.
    Figure A1. Code-book construction for U N in Theorem 1 for case 1.
    Entropy 15 00445 g006
    (Construction of X N ) The x N is generated according to a new discrete memoryless channel (DMC) with inputs u N , s N , and output x N . The transition probability of this new DMC is p X | U , S ( x | u , s ) , which is obtained from the joint probability mass function p U , A , S , X , Y , Z ( u , a , s , x , y , z ) . The probability
    p X N | U N , S N ( x N | u N , s N ) is calculated as follows.
    p X N | U N , S N ( x N | u N , s N ) = i = 1 N p X | U , S ( x i | u i , s i )
    Decoding:
    Given a vector y N Y N , try to find a sequence u N ( w ^ , i ^ ) such that ( u N ( w ^ , i ^ ) , a N ( w ^ ) , y N ) T U A Y N ( ϵ 3 ) . If there exist sequences with the same w ^ , put out the corresponding w ^ . Otherwise, i.e., if no such sequence exists or multiple sequences have different message indices, declare a decoding error.
  • (Code construction for Case 2)
    Given a pair ( R , R e ) , choose a joint probability mass function p U , A , S , X , Y , Z ( u , a , s , x , y , z ) such that
    0 R e R
    R I ( U ; Y ) I ( U ; S | A )
    R e = H ( A | Z )
    The message set W satisfies the following condition:
    lim N 1 N log W = R = I ( U ; Y ) I ( U ; S | A ) γ 1 ,
    where γ 1 is a fixed positive real numbers and
    0 γ 1 ( b ) I ( U ; Y ) I ( U ; S | A ) H ( A | Z ) .
    Note that (b) is from R R e = H ( A | Z ) and (A7). Let W = { 1 , 2 , . . . , 2 N R } .
    Code-book generation:
    (Construction of A N )
    Generate 2 N R i.i.d. sequences a N , according to the probability mass function p A ( a ) . Index each sequence by i { 1 , 2 , . . . , 2 N R } . For a given message w ( w W ), choose a corresponding a N ( w ) as the output of the action encoder.
    (Construction of U N )
    For the transmitted action sequence a N ( w ) , generate 2 N ( I ( U ; Y ) ϵ 2 , N ) ( ϵ 2 , N 0 as N ) i.i.d. sequences u N , according to the probability mass function p U | A ( u i | a i ( w ) ) . Distribute these sequences at random into 2 N R = 2 N ( I ( U ; Y ) I ( U ; S | A ) γ 1 ) bins such that each bin contains 2 N ( I ( U ; S | A ) + γ 1 ϵ 2 , N ) sequences. Index each bin by i { 1 , 2 , . . . , 2 N R } .
    Let s N be the state sequence generated in response to the action sequence a N ( w ) . For a given message w ( w W ) and channel state s N , try to find a sequence u N ( w , i * ) in bin w such that ( u N ( w , i * ) , a N ( w ) , s N ) T U A S N ( ϵ 2 ) . If multiple such sequences in bin w exist, choose the one with the smallest index in the bin. If no such sequence exists, declare an encoding error.
    Figure A2 shows the construction of U N for case 2, see the following.
    Figure A2. Code-book construction for U N in Theorem 1 for case 2.
    Figure A2. Code-book construction for U N in Theorem 1 for case 2.
    Entropy 15 00445 g007
    (Construction of X N ) The x N is generated the same as that for the case 1, and it is omitted here.
    Decoding:
    Given a vector y N Y N , try to find a sequence u N ( w ^ , i ^ ) such that ( u N ( w ^ , i ^ ) , a N ( w ^ ) , y N ) T U A Y N ( ϵ 3 ) . If there exist sequences with the same w ^ , put out the corresponding w ^ . Otherwise, i.e., if no such sequence exists or multiple sequences have different message indices, declare a decoding error.

A.2. Proof of Achievability

By using the above definitions, it is easy to verify that lim N log W N = R .
Then, for the two cases, note that the above encoding and decoding schemes are similar to the one used in [5]. Hence, by similar arguments as in [5], it is easy to show that P e ϵ for both cases, and the proof is omitted here. It remains to show that lim N Δ R e for the two cases, see the following.
Proof of lim N Δ R e for case 1
lim N Δ = lim N 1 N H ( W | Z N ) = lim N 1 N ( H ( W , Z N ) H ( Z N ) ) = lim N 1 N ( H ( W , Z N , U N , J ) H ( J , U N | Z N , W ) H ( Z N ) ) = ( a ) lim N 1 N ( H ( Z N | U N ) + H ( U N , J , W ) H ( J , U N | Z N , W ) H ( Z N ) ) = ( b ) lim N 1 N ( H ( Z N | U N ) + H ( U N ) H ( J , U N | Z N , W ) H ( Z N ) ) = lim N 1 N ( H ( U N ) H ( J , U N | Z N , W ) I ( Z N ; U N ) ) = lim N 1 N ( H ( U N ) H ( J | Z N , W ) H ( U N | Z N , W , J ) I ( Z N ; U N ) ) ( c ) lim N 1 N ( H ( U N ) log J H ( U N | Z N , W , J ) I ( Z N ; U N ) ) lim N 1 N ( H ( U N ) H ( U N | Y N ) log J H ( U N | Z N , W , J ) I ( Z N ; U N ) ) = lim N 1 N ( I ( Y N ; U N ) log J H ( U N | Z N , W , J ) I ( Z N ; U N ) ) = ( d ) lim N 1 N ( N I ( Y ; U ) log J H ( U N | Z N , W , J ) N I ( Z ; U ) ) = ( e ) lim N 1 N ( N I ( Y ; U ) N max ( I ( U ; S | A ) , I ( U ; Z ) ) + N I ( U ; Z ) N ϵ 3 , N N I ( Z ; U ) ) = ( f ) I ( Y ; U ) max ( I ( U ; S | A ) , I ( U ; Z ) ) = R e
where (a) is from ( W , J ) U N Z N , (b) is from H ( J , W | U N ) = 0 , (c) is from H ( J | Z N , W ) H ( J ) log J , (d) is from that S N , U N and X N are i.i.d. generated random vectors, and the channels are discrete memoryless, (e) is from (A3) and (A5), and (f) is from ϵ 3 , N 0 as N .
Thus, lim N Δ R e for case 1 is proved.
Proof of lim N Δ R e for case 2
lim N Δ = lim N 1 N H ( W | Z N ) = ( 1 ) lim N 1 N H ( A N | Z N ) = ( 2 ) lim N 1 N ( N H ( A | Z ) ) = H ( A | Z ) = R e
where (1) is from A N is a function of W, and (2) is from A N and X N are i.i.d. generated random vectors, and the channels are discrete memoryless.
Thus, lim N Δ R e for case 2 is proved.
The proof of Theorem 1 is completed.

B. Proof of Theorem 2

In this section, we prove Theorem 2: all the achievable ( R , R e ) pairs are contained in the set R ( n o ) . Suppose ( R , R e ) is achievable, i.e., for any given ϵ > 0 , there exists a channel encoder-decoder ( N , Δ , P e ) such that
lim N log W N = R , lim N Δ R e , P e ϵ
Then we will show the existence of random variables ( A , U , K , V ) ( X , S ) Y Z such that
0 R e R
R I ( U ; Y ) I ( U ; S | A )
R e I ( U ; Y ) I ( K ; Z | V )
Since W is uniformly distributed over W , we have H ( W ) = log W . The formulas (A12) and (A13) are proved by Lemma 1, see the following.
Lemma 1
The random vectors Y N , Z N and the random variables W, V, U, K, A, Y, Z of Theorem 2, satisfy:
1 N H ( W ) I ( U ; Y ) I ( U ; S | A ) + 1 N δ ( P e )
1 N H ( W | Z N ) I ( U ; Y ) I ( K ; Z | V ) + 1 N δ ( P e )
where δ ( P e ) = h ( P e ) + P e log ( | W | 1 ) . Note that h ( P e ) = P e log P e ( 1 P e ) log ( 1 P e )
Substituting H ( W ) = log W and (5) into (A14) and (A14), and using the fact that ϵ 0 , the formulas (A12) and (A13) are obtained. The formula (A11) is from
R e lim N Δ = lim N 1 N H ( W | Z N ) lim N 1 N H ( W ) = R
It remains to prove Lemma 1, see the following.
Proof 3 (Proof of Lemma 1)
The formula (A14) follows from (A16), (A18) and (A28). The formula (A15) is from (A16), (A17), (A18), (A22), (A28) and (A29).
<Part i> We begin with the left parts of the inequalities (A14) and (A15), see the following.
Since W Y N Z N is a Markov chain, for the message W, we have
1 N H ( W ) = 1 N H ( W | Y N ) + 1 N I ( Y N ; W ) ( a ) 1 N δ ( P e ) + 1 N I ( Y N ; W )
For the equivocation to the wiretapper, we have
1 N H ( W | Z N ) = 1 N ( H ( W ) I ( W ; Z N ) ) = 1 N ( H ( W ) + H ( W | Y N ) H ( W | Y N ) I ( W ; Z N ) ) = 1 N ( I ( W ; Y N ) + H ( W | Y N ) I ( W ; Z N ) ) ( b ) 1 N ( I ( W ; Y N ) I ( W ; Z N ) + δ ( P e ) )
Note that (a) and (b) follow from Fano’s inequality.
<Part ii> By using chain rule, the character I ( Y N ; W ) in formulas (A16) and (A17) can be bounded as follows,
1 N I ( Y N ; W ) = 1 N i = 1 N I ( Y i ; W | Y i 1 ) = ( 1 ) 1 N i = 1 N ( I ( Y i ; W | Y i 1 ) I ( S i ; W | S i + 1 N , A N ) ) = 1 N i = 1 N ( I ( Y i ; W , S i + 1 N , A N | Y i 1 ) I ( Y i ; S i + 1 N , A N | W , Y i 1 ) I ( S i ; W , Y i 1 | S i + 1 N , A N ) + I ( S i ; Y i 1 | W , S i + 1 N , A N ) ) = ( 2 ) 1 N i = 1 N ( I ( Y i ; W , S i + 1 N , A N | Y i 1 ) I ( S i ; W , Y i 1 | S i + 1 N , A N ) ) = 1 N i = 1 N ( H ( Y i | Y i 1 ) H ( Y i | Y i 1 , W , S i + 1 N , A N ) H ( S i | S i + 1 N , A N ) + H ( S i | S i + 1 N , A N , W , Y i 1 ) ) ( 3 ) 1 N i = 1 N ( H ( Y i ) H ( Y i | Y i 1 , W , S i + 1 N , A N ) H ( S i | A i ) + H ( S i | S i + 1 N , A N , W , Y i 1 ) )
where formula (1) follows from that W A N S N , formula (2) follows from that
i = 1 N I ( Y i ; S i + 1 N , A N | W , Y i 1 ) = i = 1 N I ( S i ; Y i 1 | W , S i + 1 N , A N )
and formula (3) follows from that S i A i ( S i + 1 N , A i 1 , A i + 1 N ) .
Proof 4 (Proof of (A19))
The left part of (A19) can be rewritten as
i = 1 N I ( Y i ; S i + 1 N , A N | W , Y i 1 ) = ( 1 ) i = 1 N I ( Y i ; S i + 1 N , A N | W , Y i 1 , A N ) = i = 1 N I ( Y i ; S i + 1 N | W , Y i 1 , A N ) = i = 1 N j = i + 1 N I ( Y i ; S j | A N , Y i 1 , W , S j + 1 N ) = j = 1 N i = j + 1 N I ( Y j ; S i | A N , Y j 1 , S i + 1 N , W ) = i = 1 N j = 1 i 1 I ( Y j ; S i | A N , Y j 1 , S i + 1 N , W )
where (1) is from the fact that A N is a deterministic function of W.
The right part of (A19) can be rewritten as
i = 1 N I ( S i ; Y i 1 | W , S i + 1 N , A N ) = i = 1 N j = 1 i 1 I ( Y j ; S i | A N , W , Y j 1 , S i + 1 N )
The formula (A19) is proved by (A20) and (A21). The proof is completed.
<Part iii> Similar to (A18), the character I ( W ; Z N ) in formula (A17) can be rewritten as follows,
1 N I ( Z N ; W ) = 1 N i = 1 N I ( Z i ; W | Z i 1 ) = ( a ) 1 N i = 1 N ( I ( Z i ; W | Z i 1 ) I ( S i ; W | S i + 1 N , A N ) ) = 1 N i = 1 N ( I ( Z i ; W , S i + 1 N , A N | Z i 1 ) I ( Z i ; S i + 1 N , A N | W , Z i 1 ) I ( S i ; W , Z i 1 | S i + 1 N , A N ) + I ( S i ; Z i 1 | W , S i + 1 N , A N ) ) = ( b ) 1 N i = 1 N ( I ( Z i ; W , S i + 1 N , A N | Z i 1 ) I ( S i ; W , Z i 1 | S i + 1 N , A N ) ) = 1 N i = 1 N ( H ( Z i | Z i 1 ) H ( Z i | Z i 1 , W , S i + 1 N , A N ) H ( S i | S i + 1 N , A N ) + H ( S i | S i + 1 N , A N , W , Z i 1 ) ) 1 N i = 1 N ( H ( Z i | Z i 1 ) H ( Z i | Z i 1 , W , S i + 1 N , A N ) H ( S i | A i ) + H ( S i | S i + 1 N , A N , W , Z i 1 , Y i 1 ) ) = ( c ) 1 N i = 1 N ( H ( Z i | Z i 1 ) H ( Z i | Z i 1 , W , S i + 1 N , A N ) H ( S i | A i ) + H ( S i | S i + 1 N , A N , W , Y i 1 ) )
where formula (a) follows from that W A N S N , formula (b) follows from that
i = 1 N I ( Z i ; S i + 1 N , A N | W , Z i 1 ) = i = 1 N I ( S i ; Z i 1 | W , S i + 1 N , A N )
and formula (c) follows from that Z i 1 ( S i + 1 N , A N , W , Y i 1 ) S i . Note that the proof of (A23) is similar to the proof of (A21), and therefore, it is omitted here.
<Part iv> (single letter) To complete the proof, we introduce a random variable J, which is independent of W, A N , X N , S N , Y N and Z N . Furthermore, J is uniformly distributed over { 1 , 2 , . . . , N } . Define
U = ( W , Y J 1 , S J + 1 N , A N , J )
K = ( W , Z J 1 , S J + 1 N , A N , J )
V = ( Z J 1 , J )
X = X J , Y = Y J , Z = Z J , S = S J , A = ( A J , J )
<Part v> Then (A18) can be rewritten as
1 N I ( W ; Y N ) 1 N i = 1 N ( H ( Y i ) H ( Y i | Y i 1 , W , S i + 1 N , A N ) H ( S i | A i ) + H ( S i | S i + 1 N , A N , W , Y i 1 ) ) = 1 N i = 1 N ( H ( Y i | J = i ) H ( Y i | Y i 1 , W , S i + 1 N , A N , J = i ) H ( S i | A i , J = i ) + H ( S i | S i + 1 N , A N , W , Y i 1 , A i , J = i ) ) = H ( Y J | J ) H ( Y J | Y J 1 , W , S J + 1 N , A N , J ) H ( S J | A J , J ) + H ( S J | S J + 1 N , A N , W , Y J 1 , A J , J ) H ( Y J ) H ( Y J | Y J 1 , W , S J + 1 N , A N , J ) H ( S J | A J , J ) + H ( S J | S J + 1 N , A N , W , Y J 1 , A J , J ) = H ( Y ) H ( Y | U ) H ( S | A ) + H ( S | U , A ) = I ( U ; Y ) I ( U ; S | A )
Analogously, (A22) is rewritten as follows,
1 N I ( Z N ; W ) 1 N i = 1 N ( H ( Z i | Z i 1 ) H ( Z i | Z i 1 , W , S i + 1 N , A N ) H ( S i | A i ) + H ( S i | S i + 1 N , A N , W , Y i 1 ) ) = 1 N i = 1 N ( H ( Z i | Z i 1 , J = i ) H ( Z i | Z i 1 , W , S i + 1 N , A N , J = i ) H ( S i | A i , J = i ) + H ( S i | S i + 1 N , A N , W , Y i 1 , A i , J = i ) ) = H ( Z J | Z J 1 , J ) H ( Z J | Z J 1 , W , S J + 1 N , A N , J ) H ( S J | A J , J ) + H ( S J | S J + 1 N , A N , W , Y J 1 , A J , J ) = H ( Z | V ) H ( Z | K , V ) H ( S | A ) + H ( S | U , A ) = I ( Z ; K | V ) I ( U ; S | A )
Substituting (A28), (A29) into (A16) and (A17), Lemma 1 is proved.
In addition, by using the definitions of U, K, V, Y and Z (see (A24), (A25), (A26) and (A27), note that V is a part of K), and observing that Z J 1 ( Y J 1 , W , S J + 1 N , A N , J ) Y J Z J is a Markov chain, it is easy to check that the Markov chain V K U Y Z holds.
The proof of Theorem 2 is completed.

C. Size Constraint of The Random Variables in Theorem 1

By using the support lemma (see [13], p.310), it suffices to show that the random variable U can be replaced by new one, preserving the Markovity ( U , A ) ( X , S ) Y Z and the mutual information I ( U ; Z ) , I ( U ; Y ) , I ( U ; S | A ) , and furthermore, the range of the new U satisfies: U X S A + 2 . The proof is in the reminder of this section.
Let
p ¯ = p X S A ( x , s , a )
Define the following continuous scalar functions of p ¯ :
f X S A ( p ¯ ) = p X S A ( x , s , a ) , f Y ( p ¯ ) = H ( Y ) , f Z ( p ¯ ) = H ( Z ) , f S | A ( p ¯ ) = H ( S | A )
Since there are X S A 1 functions of f X S A ( p ¯ ) , the total number of the continuous scalar functions of p ¯ is X S A +2.
Let p ¯ X S A | U = P r { X = x , S = s , A = a | U = u } . With these distributions p ¯ X S A | U = P r { X = x , S = s , A = a | U = u } , we have
p X S A ( x , s , a ) = u U p ( U = u ) f X S A ( p ¯ X S A | U )
I ( U ; Z ) = f Z ( p ¯ ) u U p ( U = u ) f Z ( p ¯ X S A | U )
I ( U ; S | A ) = f S | A ( p ¯ ) u U p ( U = u ) f S | A ( p ¯ X S A | U )
H ( Y | U ) = u U p ( U = u ) f Y ( p ¯ X S A | U )
According to the support lemma ([13], p.310), the random variable U can be replaced by new ones such that the new U takes at most X S A + 2 different values and the expressions (A31), (A32), (A33) and (A34) are preserved.

D. Size Constraint of The Random Variables in Theorem 2

By using the support lemma (see [13], p.310), it suffices to show that the random variables U, V and K can be replaced by new ones, preserving the Markovities ( U , K , A , V ) ( X , S ) Y Z , V K Y Z and the mutual information I ( U ; Y ) , I ( K ; Z | V ) , I ( U ; S | A ) , and furthermore, the ranges of the new U, V and K satisfy: U X S A + 1 , V X S A , K X 2 S 2 A 2 . The proof is in the reminder of this section.
  • (Proof of U X S A + 1 )
    Let
    p ¯ = p X S A ( x , s , a )
    Define the following continuous scalar functions of p ¯ :
    f X S A ( p ¯ ) = p X S A ( x , s , a ) , f Y ( p ¯ ) = H ( Y ) , f S | A ( p ¯ ) = H ( S | A )
    Since there are X S A 1 functions of f X S A ( p ¯ ) , the total number of the continuous scalar functions of p ¯ is X S A +1.
    Let p ¯ X S A | U = P r { X = x , S = s , A = a | U = u } . With these distributions p ¯ X S A | U = P r { X = x , S = s , A = a | U = u } , we have
    p X S A ( x , s , a ) = u U p ( U = u ) f X S A ( p ¯ X S A | U )
    I ( U ; S | A ) = f S | A ( p ¯ ) u U p ( U = u ) f S | A ( p ¯ X S A | U )
    H ( Y | U ) = u U p ( U = u ) f Y ( p ¯ X S A | U )
    According to the support lemma ([13], p.310), the random variable U can be replaced by new ones such that the new U takes at most X S A + 1 different values and the expressions (A36), (A37) and (A38) are preserved.
  • (Proof of V X S A )
    Let
    p ¯ = p X S A ( x , s , a )
    Define the following continuous scalar functions of p ¯ :
    f X S A ( p ¯ ) = p X S A ( x , s , a ) , f Z ( p ¯ ) = H ( Z )
    Since there are X S A 1 functions of f X S A ( p ¯ ) , the total number of the continuous scalar functions of p ¯ is X S A .
    Let p ¯ X S A | V = P r { X = x , S = s , A = a | V = v } . With these distributions p ¯ X S A | V = P r { X = x , S = s , A = a | V = v } , we have
    p X S A ( x , s , a ) = v V p ( V = v ) f X S A ( p ¯ X S A | V )
    H ( Z | V ) = v V p ( V = v ) f Z ( p ¯ X S A | V )
    According to the support lemma ([13], p.310), the random variable V can be replaced by new ones such that the new V takes at most X S A different values and the expressions (A40) and (A41) are preserved.
  • (Proof of K X 2 S 2 A 2 )
    Once the alphabet of V is fixed, we apply similar arguments to bound the alphabet of K, see the following. Define X S A continuous scalar functions of p ¯ X S A :
    f X S A ( p ¯ X S A ) = p X S A ( x , s , a ) , f Z ( p ¯ X S A ) = H ( Z )
    where of the functions f X S A ( p ¯ X S A ) , only X S A 1 are to be considered.
    For every fixed v, let p ¯ X S A | K , V = P r { X = x , S = s , A = a | K = k , V = v } . With these distributions p ¯ X S A | K , V , we have
    P r { X = x , S = s , A = a | V = v } = k K P r { K = k | V = v } f X S A ( p ¯ X S A | K , V )
    I ( K ; Z | V = v ) = H ( Z | V = v ) k K f Z ( p ¯ X S A | K , V ) P r { K = k | V = v }
    By the support lemma ([13], p.310), for every fixed v, the size of the alphabet of the random variable K can not be larger than X S A , and therefore, K X 2 S 2 A 2 is proved.

E. Proof of Theorem 3

In this section, we will show that any pair ( R , R e ) R c i is achievable. Wyner’s random binning technique is used in the construction of the code-book.
Now the remainder of this section is organized as follows. The code construction is in Subsection E.1. The proof of achievability is given in Subsection E.2.

E.1. Code Construction

Since R e I ( U ; Y ) I ( U ; Z ) , R e H ( A | Z ) and R e R I ( U ; Y ) , it is sufficient to show that the pair ( R , R e = min { I ( U ; Y ) I ( U ; Z ) , H ( A | Z ) } is achievable, and note that this implies that R R e = min { I ( U ; Y ) I ( U ; Z ) , H ( A | Z ) } .
The construction of the code and the proof of achievability are considered into two cases:
  • (Case 1) If H ( A | Z ) I ( U ; Y ) I ( U ; Z ) , Wyner’s random binning technique [6] is used in the construction of the code-book.
  • (Case 2) If H ( A | Z ) I ( U ; Y ) I ( U ; Z ) , Shannon’s strategy [1] is used in the construction of the code-book.
  • (Code construction for case 1)
    Given a pair ( R , R e ) , choose a joint probability mass function p U , A , S , X , Y , Z ( u , a , s , x , y , z ) such that
    0 R e R
    R I ( U ; Y )
    R e = I ( U ; Y ) I ( U ; Z )
    The message set W satisfies the following condition:
    lim N 1 N log W = R = I ( U ; Y ) γ
    where γ is a fixed positive real numbers and
    0 γ ( a ) I ( U ; Z )
    Note that (a) is from R R e = I ( U ; Y ) I ( U ; Z ) and (A44). Let W = { 1 , 2 , . . . , 2 N R } .
    Code-book generation:
    (Construction of A N )
    Generate 2 N R i.i.d. sequences a N , according to the probability mass function p A ( a ) . Index each sequence by i { 1 , 2 , . . . , 2 N R } . For a given message w ( w W ), choose a corresponding a N ( w ) as the output of the action encoder.
    (Construction of U N )
    For the transmitted action sequence a N ( w ) , generate 2 N ( I ( U ; Y ) ϵ 2 , N ) ( ϵ 2 , N 0 as N ) i.i.d. sequences u N , according to the probability mass function p U | A ( u i | a i ( w ) ) . Distribute these sequences at random into 2 N R = 2 N ( I ( U ; Y ) γ ) bins such that each bin contains 2 N ( γ ϵ 2 , N ) sequences. Index each bin by i { 1 , 2 , . . . , 2 N R } .
    Here note that the number of the sequences in every bin is upper bounded as follows.
    γ ϵ 2 , N ( a ) I ( U ; Z ) ϵ 2 , N
    where (a) is from (A45). This implies that
    lim N H ( U N | W , Z N ) = 0
    Note that (A47) can be proved by using Fano’s inequality and (A46).
    For a given message w ( w W ), randomly choose a sequence u N ( w , i * ) in bin w as the realization of U N .
    Let s N be the state sequence generated in response to the action sequence a N ( w ) .
    (Construction of X N ) The x N is generated according to a new discrete memoryless channel (DMC) with inputs u N , s N , and output x N . The transition probability of this new DMC is p X | U , S ( x | u , s ) , which is obtained from the joint probability mass function p U , A , S , X , Y , Z ( u , a , s , x , y , z ) . The probability
    p X N | U N , S N ( x N | u N , s N ) is calculated as follows.
    p X N | U N , S N ( x N | u N , s N ) = i = 1 N p X | U , S ( x i | u i , s i )
    Decoding:
    Given a vector y N Y N , try to find a sequence u N ( w ^ , i ^ ) such that ( u N ( w ^ , i ^ ) , a N ( w ^ ) , y N ) T U A Y N ( ϵ 3 ) . If there exist sequences with the same w ^ , put out the corresponding w ^ . Otherwise, i.e., if no such sequence exists or multiple sequences have different message indices, declare a decoding error.
  • (Code construction for case 2)
    Given a pair ( R , R e ) , choose a joint probability mass function p U , A , S , X , Y , Z ( u , a , s , x , y , z ) such that
    0 R e R
    R I ( U ; Y )
    R e = H ( A | Z )
    The message set W satisfies the following condition:
    lim N 1 N log W = R = I ( U ; Y ) γ 1
    where γ 1 is a fixed positive real numbers and
    0 γ 1 ( b ) I ( U ; Y ) H ( A | Z )
    Note that (b) is from R R e = H ( A | Z ) and (A49). Let W = { 1 , 2 , . . . , 2 N R } .
    Code-book generation:
    (Construction of A N )
    Generate 2 N R i.i.d. sequences a N , according to the probability mass function p A ( a ) . Index each sequence by i { 1 , 2 , . . . , 2 N R } . For a given message w ( w W ), choose a corresponding a N ( w ) as the output of the action encoder.
    (Construction of U N )
    For the transmitted action sequence a N ( w ) , generate 2 N R i.i.d. sequences u N , according to the probability mass function p U | A ( u i | a i ( w ) ) . Index each u N by i { 1 , 2 , . . . , 2 N R } .
    For a given message w ( w W ), choose a sequence u N ( w ) as the realization of U N .
    Let s N be the state sequence generated in response to the action sequence a N ( w ) .
    (Construction of X N ) The x N is generated the same as that for the case 1, and it is omitted here.
    Decoding:
    Given a vector y N Y N , try to find a sequence u N ( w ^ ) such that ( u N ( w ^ ) , a N ( w ^ ) , y N ) T U A Y N ( ϵ 3 ) . If there exist sequences with the same w ^ , put out the corresponding w ^ . Otherwise, i.e., if no such sequence exists or multiple sequences have different message indices, declare a decoding error.

E.2. Proof of Achievability

By using the above definitions, it is easy to verify that lim N log W N = R .
Then, for the two cases, note that the above encoding and decoding schemes are similar to the one used in [5]. Hence, by similar arguments as in [5], it is easy to show that P e ϵ for both cases, and the proof is omitted here. It remains to show that lim N Δ R e for the two cases, see the following.
Proof of lim N Δ R e for case 1
lim N Δ = lim N 1 N H ( W | Z N ) = lim N 1 N ( H ( W , Z N ) H ( Z N ) ) = lim N 1 N ( H ( W , Z N , U N ) H ( U N | Z N , W ) H ( Z N ) ) = ( a ) lim N 1 N ( H ( Z N | U N ) + H ( U N , W ) H ( U N | Z N , W ) H ( Z N ) ) = ( b ) lim N 1 N ( H ( Z N | U N ) + H ( U N ) H ( U N | Z N , W ) H ( Z N ) ) = lim N 1 N ( H ( U N ) H ( U N | Z N , W ) I ( Z N ; U N ) ) lim N 1 N ( H ( U N ) H ( U N | Y N ) H ( U N | Z N , W ) I ( Z N ; U N ) ) = lim N 1 N ( I ( U N ; Y N ) H ( U N | Z N , W ) I ( Z N ; U N ) ) = ( c ) lim N 1 N ( N I ( U ; Y ) N I ( U ; Z ) ) = I ( U ; Y ) I ( U ; Z ) = R e ,
where (a) is from W U N Z N , (b) is from H ( W | U N ) = 0 , (c) is from that S N , U N and X N are i.i.d. generated random vectors, the channels are discrete memoryless and (A47).
Thus, lim N Δ R e for case 1 is proved.
Proof of lim N Δ R e for case 2
lim N Δ = lim N 1 N H ( W | Z N ) = ( 1 ) lim N 1 N H ( A N | Z N ) = ( 2 ) lim N 1 N ( N H ( A | Z ) ) = H ( A | Z ) = R e ,
where (1) is from A N is a function of W, and (2) is from A N and X N are i.i.d. generated random vectors, and the channels are discrete memoryless.
Thus, lim N Δ R e for case 2 is proved.
The proof of Theorem 3 is completed.

F. Proof of Theorem 4

In this section, we prove Theorem 4: all the achievable ( R , R e ) pairs are contained in the set R ( c o ) . Suppose ( R , R e ) is achievable, i.e., for any given ϵ > 0 , there exists a channel encoder-decoder ( N , Δ , P e ) such that
lim N log W N = R , lim N Δ R e , P e ϵ .
Then we will show the existence of random variables ( A , U , K , V ) ( X , S ) Y Z such that
0 R e R
R I ( U ; Y )
R e I ( U ; Y ) I ( K ; Z | V )
Since W is uniformly distributed over W , we have H ( W ) = log W . The formulas (A54) and (A55) are proved by Lemma 2, see the following.
Lemma 2
The random vectors Y N , Z N and the random variables W, V, U, K, A, Y, Z of Theorem 4, satisfy:
1 N H ( W ) I ( U ; Y ) + 1 N δ ( P e ) ,
1 N H ( W | Z N ) I ( U ; Y ) I ( K ; Z | V ) + 1 N δ ( P e ) ,
where δ ( P e ) = h ( P e ) + P e log ( | W | 1 ) . Note that h ( P e ) = P e log P e ( 1 P e ) log ( 1 P e ) .
Substituting H ( W ) = log W and (5) into (A56) and (A57), and using the fact that ϵ 0 , the formulas (A54) and (A55) are obtained. The formula (A53) is from
R e lim N Δ = lim N 1 N H ( W | Z N ) lim N 1 N H ( W ) = R .
It remains to prove Lemma 2, see the following.
Proof 5 (Proof of Lemma 2)
The formula (A56) follows from (A58), (A60) and (A66). The formula (A57) is from (A58), (A59), (A60), (A61), (A66) and (A67).
<Part i> We begin with the left parts of the inequalities (A56) and (A57), see the following.
Since W Y N Z N is a Markov chain, for the message W, we have
1 N H ( W ) = 1 N H ( W | Y N ) + 1 N I ( Y N ; W ) ( a ) 1 N δ ( P e ) + 1 N I ( Y N ; W ) .
For the equivocation to the wiretapper, we have
1 N H ( W | Z N ) = 1 N ( H ( W ) I ( W ; Z N ) ) = 1 N ( H ( W ) + H ( W | Y N ) H ( W | Y N ) I ( W ; Z N ) ) = 1 N ( I ( W ; Y N ) + H ( W | Y N ) I ( W ; Z N ) ) ( b ) 1 N ( I ( W ; Y N ) I ( W ; Z N ) + δ ( P e ) ) .
Note that (a) and (b) follow from Fano’s inequality.
<Part ii> By using chain rule, the character I ( Y N ; W ) in formulas (A58) and (A59) can be bounded as follows,
1 N I ( Y N ; W ) = 1 N i = 1 N I ( Y i ; W | Y i 1 ) 1 N i = 1 N I ( Y i ; W , Y i 1 ) 1 N i = 1 N I ( Y i ; W , Y i 1 , S i 1 ) .
<Part iii> Similar to (A60), the character I ( W ; Z N ) in formula (A59) can be rewritten as follows,
1 N I ( Z N ; W ) = 1 N i = 1 N I ( Z i ; W | Z i 1 ) = 1 N i = 1 N ( H ( Z i | Z i 1 ) H ( Z i | Z i 1 , W ) ) .
<Part iv> (single letter) To complete the proof, we introduce a random variable J, which is independent of W, A N , X N , S N , Y N and Z N . Furthermore, J is uniformly distributed over { 1 , 2 , . . . , N } . Define
U = ( W , Y J 1 , S J 1 , J )
K = ( W , Z J 1 , J )
V = ( Z J 1 , J )
X = X J , Y = Y J , Z = Z J , S = S J , A = A J
<Part v> Then (A60) can be rewritten as
1 N I ( W ; Y N ) 1 N i = 1 N I ( Y i ; W , Y i 1 , S i 1 ) = 1 N i = 1 N I ( Y i ; W , Y i 1 , S i 1 | J = i ) = I ( Y J ; W , Y J 1 , S J 1 | J ) I ( Y J ; W , Y J 1 , S J 1 , J ) = I ( U ; Y )
Analogously, (A61) is rewritten as follows,
1 N I ( Z N ; W ) = 1 N i = 1 N ( H ( Z i | Z i 1 ) H ( Z i | Z i 1 , W ) ) = 1 N i = 1 N ( H ( Z i | Z i 1 , J = i ) H ( Z i | Z i 1 , W , J = i ) ) = H ( Z J | Z J 1 , J ) H ( Z J | Z J 1 , W , J ) = H ( Z | V ) H ( Z | K , V ) = I ( Z ; K | V )
Substituting (A66), (A67) into (A58) and (A59), Lemma 2 is proved.
In addition, by using the definitions of U, K, V, Y and Z (see (A62), (A63), (A64) and (A65), note that V is a part of K), and observing that Z J 1 ( Y J 1 , W , S J 1 , J ) Y J Z J and ( W , Y J 1 , S J 1 , J ) A J S J are two Markov chains, it is easy to check that the Markov chains V K U Y Z and U A S hold.
The proof of Theorem 4 is completed.

Share and Cite

MDPI and ACS Style

Dai, B.; Vinck, A.J.H.; Luo, Y.; Tang, X. Wiretap Channel with Action-Dependent Channel State Information. Entropy 2013, 15, 445-473. https://doi.org/10.3390/e15020445

AMA Style

Dai B, Vinck AJH, Luo Y, Tang X. Wiretap Channel with Action-Dependent Channel State Information. Entropy. 2013; 15(2):445-473. https://doi.org/10.3390/e15020445

Chicago/Turabian Style

Dai, Bin, A. J. Han Vinck, Yuan Luo, and Xiaohu Tang. 2013. "Wiretap Channel with Action-Dependent Channel State Information" Entropy 15, no. 2: 445-473. https://doi.org/10.3390/e15020445

Article Metrics

Back to TopTop