Directed Information and Channel with Feedback Prapun Suksompong December 10, 2006 In this article, we consider the problem of sending message via discrete channel with noiseless feedback as shown below. Message W

Xk

Encoder

(

p yk y0k −1 , x1k

)

Yk

Decoder



Yk −1 At time k, the encoder produce X k from W, the previous encoder outputs X 1k −1 , and the fed-back channel outputs Y0k −1 . The encoder can be a deterministic function; that

is X k = f k ( X 1k −1 , Y0k −1 ) . We also allow random encoder; that is X k may be governed

(

)

by the conditional distribution p xk x1k −1 , y0k −1 . This X k is put into channel. The channel output Yk is generated according to the conditional distribution

(

)

p yk x1k , y0k −1 which not only depends on the current channel input X k but may also depend on past channel inputs X 1k −1 and outputs Y0k −1 . We modify the classical Shannon’s information measures (entropy, mutual information, and their conditional versions) so that they explicitly incorporate feedback. In particular, we shall focus on I ( X N → Y N ) , a notation introduced by Massey [1990] to capture the directed information flowing from the length N sequence of random variables X N to the length N sequence of random variables X N . In fact, the idea of given direction to information has already been thought of by Marko [Marko 1973] whose paper define a quantity called directed transinformation. The direct information in [Massey 1990] refines this directed transinformation.

Equivalent System Models In this section, we show three general models of the system we are interested in. We then prove that they are all equivalent in the sense that we can convert one into another. Hence, in some sense, it is sufficient to analyze only one of them. As usual, k is the time index. Xk and Yk represent the channel input and output at time k respectively. Sk denotes the channel state at time k. In model 2, we use Vk instead of Yk to represent the channel output. This notational difference is used so that the proof can be done more smoothly. Model 3 is introduced as an intermediate model to bridge model 1 and model 2. The definitions and proof are almost the same as those in [Chen, Suksompong, and Berger 2004]. What follows is conditioned on the initial state S1 = s1. • Model 1:

1

Channel

Xk

Encoder

(

k 1

k 1

P1 Yk X , S

W

Yk

)

Sk



S k +1 Yk −1

State

(

P2 Sk +1 X 1k , S1k



Decoder

)

Model 2:

Channel Q Vk X 1k ,V1k −1

Xk

Encoder

(

W

Vk

)

Decoder



Vk −1 •

Model 3: Encoder

Xk

Channel

(

k 1

k −1 1

P Yk , Sk +1 X , Y

W

k 1

,S

)

(Yk , Sk +1 )

Decoder



(Yk −1 , Sk ) •

Theorem: Model 1, 2, and 3 are equivalent. Proof Let Mi be the set of systems that can be represented by Model i. Then,

(

)

(1) M1 ⊂ M3 because we can let P Yk , S k +1 X 1k , Y1k −1 , S1k =

(

) (

)

P1 Yk X 1k , S1k P2 Sk +1 X 1k , S1k . (2) M3 ⊂ M2 because we can let Vk = (Yk , S k +1 ) . Then, V1k −1 = (Y1k −1 , S1k ) .

(

)

(

)

Set Q Vk X 1k ,V1k −1 = P Yk , S k +1 X 1k , Y1k −1 , S1k . (3) M2 ⊂ M1 because we can set S k = Vk −1 ,

(

)

(

)

P2 Sk +1 X 1k , S1k = Q Vk X 1k ,V1k −1 , and Yk = c a.s. (Equivalently, ∃c

(

)

∀ ( x1k , s1k ) P1 Yk = c X 1k = x1k , S1k = s1k = 1 .) Henceforth, we shall focus on Model 2.

Discrete Channel In this section, we consider the channel part of the system. Special cases of which are also defined.

2

( (

))

N



Discrete channel: p yn x1n , y1n −1



• [Tatikonda 2000] calls this nonanticipative channel. Definition: [Cover Thomas 1991] A channel is memoryless if p yn x1n , y1n −1 = p ( yn xn ) .

(

n =1

.

)

Simple equivalent statements are

≡ Given xn , ( x1n −1 , y1n −1 ) and yn are independent;





yn − xn − ( x1n −1 , y1n −1 ) forms a Markov chain;



H Yn X 1n , Y1n −1 = H (Yn X n ) ;



∀i I Y j ; ( X 1j −1 , Y1 j −1 ) X j = 0 .

(

)

(

)

Definition [Massey 1990]: The channel is used without feedback if p xn x1n −1 , y1n −1 = p xn x1n −1 .

(

) (

)

Equivalent statements are

(

)

(

)



H X n X 1n −1 , Y1n −1 = H X n X 1n −1 ;



I X n ; Y1n −1 X 1n −1 = 0 ;



Y1n −1 − X 1n −1 − X n forms a Markov chain;

(

)

≡ [Ash 1965] for ∀N > n , p ( yn x1N , y1n −1 ) = p ( yn x1n , y1n −1 ) ;

(Note that this gives exactly the same condition as the definition above when N = 2.) ≡

H (Yn X 1N , Y1n −1 ) = H (Yn X 1n , Y1n −1 ) ;



∀i I Yi ; X iN+1 X 1i , Y1i −1 = 0 ;



∀i Yi − ( X 1i , Y1i −1 ) − X iN+1 forms a Markov chain.



I X i ;Y j X 1i −1 , Y1 j −1 = 0 for any j < i .

(

)

(

)

We shall show later that these are also equivalent to:





I ( X 1N ;Y1N ) = I ( X 1N → Y1N )



I ( 0 * Y1N −1 → X 1N ) = 0 .

Interpretation: The choice of the next channel input digit, given all previous input digits, is not further related to the previous channel output digits. Joint distribution: p ( x1N , y1N ) = ∏ p ( xn , yn x1n −1 , y1n −1 ) = ∏ p ( xn x1n −1 , y1n −1 ) p ( yn x1n , y1n −1 ) . •

N

N

n =1

n =1

Discrete memoryless channel: p ( x1N , y1N ) = ∏ p ( xn x1n −1 , y1n −1 ) p ( yn xn ) . N

n =1

3



Discrete channel used without feedback:

(

)

(

p ( x1N , y1N ) = ∏ p xn x1n −1 p ( yn x1n , y1n −1 ) = ∏ p xn x1n −1 N

n =1

N

n =1

)∏ p ( y N

n =1

n

x1n , y1n −1 )

= p ( x1N ) ∏ p ( yn x1n , y1n −1 ) N

n =1

or equivalently,

(

)

p y1N x1N = ∏ p ( yn x1n , y1n −1 ) •

N

n =1

Used [Ash], p ( yn x1N , y1n −1 ) = p ( yn x1n , y1n −1 ) , we have p ( x1N , y1N ) = p ( x1N ) ∏ p ( yn x1N , y1n −1 ) = p ( x1N ) ∏ p ( yn x1n , y1n −1 ) ;



N

N

n =1

n =1

hence equivalent to Massey’s. Discrete memoryless channel used without feedback:

(

)

p ( x1N , y1N ) = p ( x1n ) ∏ p ( yn xn ) , or equivalently p y1N x1N = ∏ p ( yn xn ) . N

n =1

This is DMC used without feedback. In this case,



H ( X 1N , Y1N ) = H ( X 1N ) + ∑ H (Yi X i ) . N

i =1



(

)

N

H Y1N X 1N = ∑ H (Yi X i ) i =1

4

N

n =1

As an introduction to directed information, we present here one way to look at it. First we shall decompose mutual information into several conditional mutual information using chain rules. We associated each part with an arrow going between Xi and Yj in a diagram representing the general channel in figure 1. The above specialized channels are then just the general channel with some of the arrows missing. Directed information takes only some of these parts and hence it is a part of mutual information. This idea was presented to Prof. T. Berger and L. Tong during the author’s Q exam in June, 2005. This representation of directed information trivializes several of its properties. • The definitions defined above can be represented using the diagrams. Y1 X1 Yn −1 X n −1

Yn

W



Xn Yn +1 X n +1

Figure 1 Discrete Channel

Figure 1 shows the discrete channel in its full generality. We can, in fact, decompose it into three parts which are shown in black, blue, and red. Consider, first, the part in black as shown in figure 2. X1

Y1

X n −1

Yn −1

Xn

Yn

X n +1

Yn +1

Figure 2 Discrete Memoryless Channel without Feedback

Figure 2 represents the discrete memoryless channel without feedback. We can add memory to channel by adding the blue part, as shown in figure 3. This allows yk to depends on ( x1k , y1k −1 ) , not just xk .

5

X1

Y1

X n −1

Yn −1

Xn

Yn

X n +1

Yn +1

Figure 3 Discrete Channel without Feedback

The last part, with red color in figure 1, represents the feedback. Adding it to figure 3 leads us back the general discrete channel in figure 1.



The mutual information I ( X 1N ;Y1N ) delivered via the discrete channel described above can also be partitioned into three parts. First, by applying the chain rule for mutual information twice, we have

(

)

(

)

I ( X 1N ; Y1N ) = ∑ I X i ; Y1N X 1i −1 = ∑∑ I X i ; Y j X 1i −1 , Y1 j −1 . N

i =1

N

N

i =1 j =1

Now, we separate the terms inside the sum above into three groups by asking whether i is equal to, less than, or grater than j. This gives 1) i = j :

∑ I (Y ; X N

i

i =1

i

)

Y1i −1 , X 1i −1 . This quantity will be defined as I ( X 1N ↔ Y1N ) .

We want to say that it relates to the “direct” paths in black above. 2) i < j :

N

j −1

(

)

N

(

)

∑∑ I X i ;Y j X1i −1,Y1 j −1 = ∑ I X 1j −1;Y j Y1 j −1 . This quantity will be j =1 i =1

j =1

defined as I ( 0 * X 1N −1 → Y1N ) . We want to say that it relates to the channel memory paths in blue above. 3)

j < i:

i −1

∑∑ I ( X ;Y N

i =1 j =1

i

j

)

N

(

)

X 1i −1 , Y1 j −1 = ∑ I X i ; Y1i −1 X 1i −1 . This quantity will be i =1

defined as I ( 0 * Y1N −1 → X 1N ) . We want to say that it captures the feedback paths in red above. The directed information only adds up the terms that has i ≤ j (all arrows that point from X to Y), that is I ( X 1N → Y1N ) = I ( X 1N ↔ Y1N ) + I ( 0 * X 1N −1 → Y1N )

= I ( X 1N ; Y1N ) − I ( 0 * Y1N −1 → X 1N ) For the familiar DMC without feedback, we will show that I ( X 1N → Y1N ) = I ( X 1N ; Y1N ) ≤ ∑ I (Yi ; X i ) ≤ nC . N

i =1

6

Directed Information and its properties •



[Kramer 1998] Assume that the sequences X 1N and Y1N are “synchronized”, i.e., that the nth terms in the sequences occur “at the same time”, and that the nth terms occur “before” the (n+1)st terms.

[Massey 1990] The directed information I ( X 1N → Y1N ) from a sequence X 1N to

(

)

a sequence Y1N is defined by I ( X 1N → Y1N ) = ∑ I X 1i ;Yi Y1i −1 . N

i =1



Remarks:



(

)

I ( X 1N ;Y1N ) = ∑ I X 1N ;Yi Y1i −1 by chain rule. We add the information that N

i =1

Yi tells about X 1N which Y1i −1 haven’t already told.



(

)

For directed information, we use I X 1n ; Yn Y1n −1 : To eliminate feedback information, we ignore the information that Yi may be giving about the future X iN+1 .



(

N

(

) (

(



N

)

⎤ ⎥ ⎥ ⎦

)

(

I (Y1N → X 1N ) = ∑ I Y1i ; X i X 1i −1 i =1



)

⎡ p X 1i , Yi Y1i −1 I ( X → Y1 ) = ∑ E ⎢log ⎢ p X 1i Y1i −1 p Yi Y1i −1 i =1 ⎣ N ⎡ i i −1 ⎤ ⎢ ∏ p Yi X 1 , Y1 ⎥ ⎥ = E ⎢log i =1 p (Y1N ) ⎢ ⎥ ⎢⎣ ⎥⎦ N

N 1

)

I ( X 1N → Y1N ) ≤ I ( X 1N ; Y1N ) with equality iff channel is used without feedback

(

)

( ) ( ≥ H (Y Y ) − H (Y = I ( X ;Y Y )

) )

Proof. I X 1N ;Yi Y1i −1 = H Yi Y1i −1 − H Yi X 1N , Y1i −1 i

i −1 1

i 1

i

i

X 1i , Y1i −1

i −1 1

Hence,

(

)

(

)

I ( X 1N ;Y1N ) = ∑ I X 1N ;Yi Y1i −1 ≥ ∑ I X 1n ;Yn Y1n −1 = I ( X 1N → Y1N ) . N

i =1

(

N

n =1

)

(

)

Equality occurs iff ∀i H Yi X 1N , Y1i −1 = H Yi X 1n , Y1i −1 . This is Ash’s condition which is equivalent to the “used without feedback” condition. Alternative Proof. Use I ( X 1 , X 2 ;Y Z ) = I ( X 1;Y ) + I ( X 2 ; Y X 1 , Z ) . Then,

(

) (

) (

)

I X 1N ; Yi Y1i −1 − I X 1i ; Yi Y1i −1 = I X iN+1;Yi X 1i , Y1i −1 ≥ 0

7

This proof give us the next property.



(

)

I ( X 1N ;Y1N ) − I ( X 1N → Y1N ) = ∑ I X iN+1; Yi X 1i , Y1i −1 ; in fact, the last term is 0; so N

i =1

N −1

(

)

I ( X 1N ;Y1N ) − I ( X 1N → Y1N ) = ∑ I X iN+1;Yi X 1i , Y1i −1 . i =1

So, I ( X 1N → Y1N ) ≤ I ( X 1N ; Y1N ) with equality iff ∀i Yi − ( X 1i , Y1i −1 ) − X iN+1 forms a Markov chain.



(

(

) (

If X iN+1 − X 1i − Y1i is a Markov chain i.e. p X iN+1 X 1i , Y1i = p X iN+1 X 1i I ( X 1N ;Y1N ) = I ( X 1N → Y1N ) .





)) , then

The Markov chain condition above says that the future X’s are not influenced by the past and current Y’s when conditioned on the past and current X’s. In the context of channels this state that if there is no feedback, then the two different mutual information measures are equal. [Tatikonda 2000, p. 81] Pearl [1988] call this the “weak union” property of conditional independence.

(

) (

)

(

) (

)

Proof. p X iN+1 X 1i , Y1i = p X iN+1 X 1i ⇒ p X iN+1 X 1i , Y1i = p X iN+1 X 1i , Y1i −1 ⇒ Yi − ( X 1i , Y1i −1 ) − X iN+1 Markov.

(Recall that p ( Z V ,U1 ,U 2 ) = p ( Z V ) ⇒ p ( Z V ,U1 ,U 2 ) = p ( Z V ,U1 ) = p ( Z V ,U 2 ) = p ( Z V ) .)



For DMC, I ( X 1N → Y1N ) ≤ ∑ I ( X i ;Yi ) with equality iff Yi are independent. N

(

i =1

i 1

i −1 1

Proof. I X ;Yi Y

) = H (Y Y ) − H (Y X , Y ) = H (Y Y ) − H (Y X ) ; memoryless i

i −1 1

i

i 1

i

i −1 1

i

i

i −1 1

≤ H (Yi ) − H (Yi X i ) = I ( X i ; Yi )

(

)

Recall that H (Y1N ) = ∑ H Yi Y1i −1 ≤ ∑ H (Yi ) with equality iff Yi are N

i =1

N

i =1

independent. •

Definition: Denote the sequence ( 0, Y1 ,…, YN −1 ) by DY1N [Kramer 1998], or

0 * Y1N −1 [Massey 1990]. • •

The letter D represent delay by one time step (with discard of the last component).

(

)

(

I ( 0 * Y1N −1 → X 1N ) = ∑ I 0 * Y1i −1; X i X 1i −1 = ∑ I Y1i −1; X i X 1i −1 N

i =1 N

(

= ∑ I Y1i −1; X i X 1i −1 i=2

)

8

N

i =1

)



(

N

i =1



)

Define I ( X 1N ↔ Y1N ) = I (Y1N ↔ X 1N ) = ∑ I Yi ; X i Y1i −1 , X 1i −1 . I ( X 1N → Y1N ) = I ( 0 * X 1N −1 → Y1N ) + I ( X 1N ↔ Y1N ) I (Y1N → X 1N ) = I ( 0 * Y1N −1 → X 1N ) + I (Y1N ↔ X 1N )

(

) (

) (

Proof. I Y1i ; X i X 1i −1 − I Y1i −1; X i X 1i −1 = I Yi ; X i Y1i −1 , X 1i −1

)

I (Y1N → X 1N ) − I ( 0 * Y1N −1 → X 1N ) N

(

)

(

N

= ∑ I Y1i ; X i X 1i −1 − ∑ I Y1i −1; X i X 1i −1 i =1 N

(

i =1

= ∑ I Yi ; X i Y1i −1 , X 1i −1 i =1



)

)

Conservation Law:

I ( X 1N ;Y1N ) = I ( X 1N → Y1N ) + I ( 0 * Y1N −1 → X 1N ) = I (Y1N → X 1N ) + I ( 0 * X 1N −1 → Y1N )



N

Equivalently,

i=2



)

i =1

(

)

Interpretation: By putting a 0 in the front, • We shift the delay position from the feedback to the channel. Before, X i and Yi are treated as if they happened during the same time step. •



(

N −1

∑ I Y1i −1; X i X1i −1 = ∑ I X iN+1;Yi X 1i ,Y1i −1 .

Now, the formula doesn’t include the “middle” part which is I ( X 1N ↔ Y1N ) .

I ( X 1N ;Y1N ) = I ( X 1N → Y1N ) + I (Y1N → X 1N ) − I ( X 1N ↔ Y1N ) •

In general, I ( X 1N ;Y1N ) ≠ I ( X 1N → Y1N ) + I (Y1N → X 1N ) . •

This is obvious when N = 1 because

I ( X 1N ;Y1N ) = I ( X 1N → Y1N ) = I (Y1N → X 1N ) .

9



Example: (I-measure) N = 2 I ( X 1 , X 2 ;Y1 , Y2 ) X1

I ( X 1 , X 2 → Y1 , Y2 ) X1

I ( 0 * Y1 → X 1 , X 2 ) X1

I (Y2 ; X 1 , X 2 Y1 )

I (Y2 ; X 1 , X 2 Y1 ) Y2

Y2

I (Y1; X 1 , X 2 )

Y1

Y2

I (Y1; X 1 )

Y1

I (Y1; X 2 X 1 )

Y1

X2

X2

X2

I(X ↔Y

I (Y1 , Y2 → X 1 , X 2 ) X1

2 1

2 1

)

I ( 0 * X 1 → Y12 ) X1

X1 I ( X 2 ;Y2 X 1 , Y1 )

I ( X 1 , Y2 Y1 )

I ( X 2 ;Y1 , Y2 X 1 )

Y2

I (Y1; X 1 )

Y1

I ( X 1;Y1 )

Y1

X2



Y2

Y2

Y1

X2

X2

DMC: I (Y2 ; ( X 1 , Y1 ) X 2 ) = 0 which equivalent to I (Y2 ; X 1 X 2 , Y1 ) = 0 , I (Y2 ;Y1 X 1 , X 2 ) = 0 , and I ( X 1; Y1; Y2 X 2 ) = 0 . DMC: I (Y2 ; ( X 1 , Y1 ) X 2 ) = 0

No feedback: I ( X 2 ; Y1 X 1 ) = 0 X1

X1

0 Y2

Y1

0

Y2

0

Y1

0

X2

X2

We can clearly see now that when there is no feedback, I ( X 12 ; Y12 ) = I ( X 12 → Y12 ) . •

We summarize the definitions and properties involving the directed information below: •

N

i =1



(

I ( X 1N ;Y1N ) = ∑ I X 1N ;Yi Y1i −1

)

(

I ( X 1N → Y1N ) = ∑ I X 1i ;Yi Y1i −1 = H (Y1N ) − ∑ H Yi X 1i , Y1i −1 N

i =1



(

) N

i =1

(

)

N −1

(

)

I ( 0 * Y1N −1 → X 1N ) = ∑ I Y1i −1; X i X 1i −1 = ∑ I X iN+1; Yi X 1i , Y1i −1 N

i =2

(

i =1

)

= I ( X 2 ;Y1 X 1 ) + I X 3 ;Y12 X 12 +

10

)

(

I ( 0 * X 1N −1 → Y1N ) = ∑ I Yi ; X 1i −1 Y1i −1 N

i =1

)

(

)

= I ( X 1;Y2 Y1 ) + I X 12 ;Y3 Y12 +

(

I ( X 1N ↔ Y1N ) = I (Y1N ↔ X 1N ) = ∑ I Yi ; X i Y1i −1 , X 1i −1 N



i =1

)

I ( X 1N → Y1N ) = I ( 0 * X 1N −1 → Y1N ) + I ( X 1N ↔ Y1N )



I (Y1N → X 1N ) = I ( 0 * Y1N −1 → X 1N ) + I (Y1N ↔ X 1N ) I ( X 1N ;Y1N ) = I ( X 1N → Y1N ) + I (Y1N → X 1N ) − I ( X 1N ↔ Y1N )



= I ( 0 * X 1N −1 → Y1N ) + I ( 0 * Y1N −1 → X 1N ) + I ( X 1N ↔ Y1N ) •

{

}

[Tatikonda 2000] If the process p ( x1N , y1N )

∞ N =1

is information stable [p. 89],

1 I ( X 1N → Y1N ) exists and we can work directly with I ( X 1N → Y1N ) N (instead of liminf in probability as defined in [Tatikonda 2000 p. 89] and [Verdú 1994]. Definition: [Kramer’s] causal conditioning • We see above that the channel outputs are given by

then lim

N →∞



(

)

p y1N x1N = ∏ p ( yn x1n , y1n −1 ) . This lead us to define

(

N

n =1

)

(

)

H Y1N X 1N = − E ⎡ p Y1N X 1N ⎤ = ∑ H (Yn X 1n , Y1n −1 ) . ⎣ ⎦ n =1

(



N

(

)

)

Again, H Y1N X 1N differs from the conditional entropy H Y1N X 1N only in that X 1n replaces X 1N .





The term “causal” reflect the conditioning on past and present values of the sequence X 1N only.



It differs from “free information” [Marko 1973] only in that X n is included in the conditioning.

(

)

y1N = ∏ p ( xn x1n −1 , y1n −1 ) ;

p x1N

(

N

n =1

)

(

Y1N = − E ⎡ log p X 1N ⎣

H X 1N

)

Y1N ⎤ = ∑ H ( X n X 1n −1 , Y1n −1 ) . ⎦ n =1 N

Note the asymmetry in the definitions above. •

(

p ( x1N , y1N ) = p x1N

) ( ) Y ) + H (Y

y1N p y1N x1N .

(

H ( X 1N , Y1N ) = H X 1N •

p(x , y N 1

N 1

) = p(x

N 1

N

N

1

1

)

X 1N .

) p ( y ) = p ( y )∏ p ( x N

N 1

y

N 1

N 1

11

n =1

n

x1n −1 , y1n −1 )

(

)

p x1N y1N = ∏ p ( xn x1n −1 , y1n )



(



(

N

n =1

)

(

)

H X 1N Y1N = − E ⎡log p X 1N Y1N ⎤ = ∑ H ( X n X 1n −1 , Y1n ) . ⎣ ⎦ n =1

)

(

H X 1N Y1N ≤ H X 1N Y1N

N

)

Proof. Conditioning only reduces entropy. •

(

)

(

)

H X 1N Y1N − H X 1N Y1N = I ( X 1N ↔ Y1N )

(

)

(

)

(

Proof. H X 1N Y1N − H X 1N Y1N = ∑ H ( X n X 1n −1 , Y1n −1 ) − H ( X n X 1n −1 , Y1n ) N

n =1 N

((

= ∑ I X n ;Yn X 1n −1 , Y1n −1 n =1

))

= I ( X 1N ↔ Y1N )



(

)

I ( X 1N → Y1N ) = H (Y1N ) − H Y1N X 1N .

(

)

(

Proof. I ( X 1N → Y1N ) = ∑ I X 1i ;Yi Y1i −1 = H (Y1N ) − ∑ H Yi X 1i , Y1i −1 N

i =1

N

(

= H (Y1N ) − H Y1N X 1N

i =1

)

)

Alternative Proof. More directly, N ⎡ i i −1 ⎢ ∏ p Yi X 1 , Y1 I ( X 1N → Y1N ) = E ⎢log i =1 p (Y1N ) ⎢ ⎢⎣

(



(

) ⎤⎥ = E ⎡⎢log

⎡ p Y1N X 1N I ( X → Y1 ) = E ⎢log ⎢ p (Y1N ) ⎣ N 1

N

(

⎥ ⎦

(

⎢ ⎣

= D p ( x1N , y1N ) p x1N



) ⎤⎥

(

p ( X 1N , Y1N )

p X 1N Y1N

)

y1N p ( y1N )

(

(

⎡ p Y1N X 1N ⎥ = E ⎢log ⎢ p (Y1N ) ⎥ ⎣ ⎥⎦

)

)

)

⎤ ⎥ p (Y1N ) ⎥ ⎦

(

I ( 0 * Y1N −1 → X 1N ) = ∑ I Y1i −1 ; X i X 1i −1 = H ( X 1N ) − H X 1N N

i =1

(

Y1N

)

Proof. I ( 0 * Y1N −1 → X 1N ) = ∑ I Y1i −1; X i X 1i −1 . N

(

i =1

)

(

)

(

)

I Y1i −1; X i X 1i −1 = H X i X 1i −1 − H X i Y1i −1 , X 1i −1 . •

(

I ( X 1N → Y1N ) = H X 1N

)

(

Y1N − H X 1N Y1N

(

)

Proof. I ( X 1N → Y1N ) = H (Y1N ) − H Y1N X 1N

(

) ( )

= H (Y1N ) − H ( X 1N , Y1N ) − H X 1N

(

= H X 1N

)

(

Y1N − H X 1N Y1N

12

Y1N

))

)

) ⎤⎥ . ⎥ ⎦

)



(

)

(

)

(

)

Definition: I X 1N → Y1N Z1N = H Y1N Z1N − H Y1N X 1N , Z1N . •

(

)

N

(

I X 1N → Y1N Z1N = ∑ I Yn ; X 1n Y1n −1 , Z1n n =1

(

)

)

Proof. I X 1N → Y1N Z1N = ∑ H (Yn Y1n −1 , Z1n ) − H (Yn Y1n −1 , X 1n , Z1n ) N

n =1

= ∑ H (Yn Y1n −1 , Z1n ) − H (Yn Y1n −1 , X 1n , Z1n ) N

n =1

(

N

= ∑ I Yn ; X 1n Y1n −1 , Z1n n =1



(

)

)

Let Z1N = 0 * X 1N −1 , then I X 1N → Y1N 0 * X 1N −1 = I ( X 1N ↔ Y1N ) .

(

)

N

(

)

(

)

Proof. I X 1N → Y1N 0 * X 1N −1 = ∑ I Yn ; X 1n Y1n −1 , X 1n −1 n =1 N

= ∑ I Yn ; X n Y1n −1 , X 1n −1 n =1

• •

(

)

Similarly, by symmetry, we have I Y1N → X 1N 0 * Y1N −1 = I ( X 1N ↔ Y1N ) .

We conclude this section with diagrams. All the diagrams below are the same diagram. First, we have the familiar diagram:

(

H X 1N Y1N

)

I ( X 1N ;Y1N )

(

H Y1N X 1N H(X

N 1

)

)

H (Y1N )

Next, the mutual information I ( X 1N ; Y1N ) is partitioned into three subsets: I ( X 1N ;Y1N ) = I ( 0 * X 1N −1 → Y1N ) + I ( 0 * Y1N −1 → X 1N ) + I ( X 1N ↔ Y1N ) . I ( 0 * Y1N −1 → X 1N )

(

H X 1N Y1N

) (

H Y1N X 1N H(X

N 1

)

)

I ( X 1N ↔ Y1N )

H (Y1N ) I ( 0 * X 1N −1 → Y1N )

I ( X 1N → Y1N ) and I (Y1N → X 1N ) are parts of I ( X 1N ; Y1N ) :

13

I (Y1N → X 1N )

I ( X 1N → Y1N )

( = H(X

)

(

Finally, H ( X 1N , Y1N ) = H X 1N Y1N + H Y1N X 1N N 1

)

(

(

H X

Y1

N

)

Y1N + H Y1N X 1N + I ( X 1N ↔ Y1N )

H Y1N X 1N

N 1

(

)

)

(

H Y1N X 1N

)

I ( X 1N ↔ Y1N )

)

(

)

)

Yk

H X 1N Y1N

Converse Channel Coding Theorem Message W

Xk

Encoder

(

p yk y0k −1 , x1k

Decoder



Yk −1 •

(

) (

)

Assume that the system is causal; that is p yi x1i , y1i −1 , w = p yi x1i , y1i −1 .

The idea is that the source output sequence should be thought of a specified prior to the process of sending sequences over channels and the channel should be aware of such sequences only via its past inputs and outputs and its current input. • •

So, p ( w, x , y N 1

N 1

) = p ( w) ∏ p ( x N

n =1

n

x1n −1 , y1n −1 , w ) p ( yn x1n , y1n −1 ) .

I (W ; Y0N ) ≤ I ( X 1N → Y1N Y0 ) which is ≤ I ( X 1N ; Y1N Y0 ) .

( (

)

(

Proof. I (W ; Y0N ) = H ( Y0N ) − H ( Y0N W ) = ∑ H Yk Y0k −1 − H Yk Y0k −1 ,W N

k =1

)) .

Because X k is produced by (Y0k −1 ,W ) , we have X 1k is produced by f (Y0k −1 ,W ) , and therefore

(

)

(

H Yk Y0k −1 ,W = H Yk Y0k −1 , X 1k ,W

)

(a)

where (a) comes from the causality assumption. A more general X k gives

14

(

)

= H Yk Y0k −1 , X 1k ,

(

)

(

)

(a)

(

)

H Yk Y0k −1 ,W ≥ H Yk Y0k −1 , X 1k ,W = H Yk Y0k −1 , X 1k . In any case,

( (

)

(

I (W ; Y0N ) ≤ ∑ H Yk Y0k −1 − H Yk Y0k −1 , X 1k N

k =1



) ) = ∑ I (Y ; X N

k

k =1

(

k 1

)

Y0k −1 .

)

Remark: Let’s reconsider the equality of H Yk Y0k −1 ,W and

(

)

H Yk Y0k −1 , X 1k ,W using functional dependence graphs as in [Kramer 1998]. Suppose k = 2, then the relevant graph is shown below:

W

X1

Y1

X2

Y2

We then remove the arrows coming out of (W , Y1 ) . The resulted graph shows that (W , Y1 ) does not d-separate Y2 from X 12 . Since all secondary random variables have incoming branches, we also know that (W , Y1 ) does not fdseparate Y2 from X 12 .

W

X1

Y1

X2

Y2

(

)

Now, instead of generating X n by conditional distribution p xn x1n −1 , y1n −1 , we use deterministic encoder; that is X n = f n ( X 1n −1 , Y1n −1 ) . Then, after

removing the arrows coming out of (W , Y1 ) , the secondary random variables

X 1 and X 2 have no incoming branches.

W

X1

Y1

X2

Y2

Hence, we can further delete the arrows coming out of X 1 and X 2 .

W

X1

Y1

X2

Y2

We conclude that (W , Y1 ) fd-separate Y2 from X 12 in the case of deterministic •

encoder. Feedback does not increase the capacity of DMC.

(

)

N

Proof. For DMC, we know that I X 1N → Y1N Y0 ≤ ∑ I ( X i ; Yi Y0 ) . Hence,

15

i =1

(

)

N

I X 1N → Y1N Y0 ≤ ∑ I ( X i ; Yi ) ≤ NCDMC w/o feedback . i =1

• •

This first result in the information theory of feedback channel is due to Shannon [1948]. In fact, if we replace W by U L and replace Wˆ by V L . 1

U1L

Xk

Encoder

1

(

p yk y0k −1 , x1k

)

Yk

Decoder

V1L

Yk −1 Then, •

For any discrete channel, we have I (U1L ; Y0N ) ≤ I ( X 1N → Y1N Y0 ) .



Furthermore, if the channel is memoryless, we have

(

)

N

I X 1N → Y1N Y0 ≤ ∑ I ( X i ;Yi Y0 ) . i =1

References • • • •

• •



C. Shannon, “The zero error capacity of a noisy channel,” 1956. Ash, Robert B., Information theory. Corrected reprint of the 1965 original. Dover Publications, Inc., New York, 1990. xii+339 pp. J. L. Massey, “Causality, Feedback and Directed Information,” pp. 303-305 in Proc. 1990 Int. Symp. on Info. Th. & its Appls., Hawaii, USA, Nov. 27-30, 1990. H. Marko, “The Bidirectional Communication Theory A Generalization of Information Theory”, IEEE Trans. Commun., vol. COM-21, pp. 1345-1351, Dec. 1973. S. Tatikonda, “Control Under Communication Constraints.” MIT Ph.D. thesis, August 2000. G. Kramer, Directed Information for Channels with Feedback, ETH Series in Inform. Proc., vol. 11. Konstanz: Hartung--Gorre, 1998. (PhD Thesis) • G. Kramer, “Capacity Results for the Discrete Memoryless Network,” IEEE Trans. Inform. Th., vol. IT-49, pp. 4-21, Jan. 2003. J. Chen, P. Suksompong and T. Berger, "Communication through a Finite-State Machine with Markov Property", CISS 2004

16

Directed Information and Channel with Feedback

Directed Information and Channel with Feedback. Prapun Suksompong. December 10, 2006. In this article, we consider the problem of sending message via ...

167KB Sizes 0 Downloads 153 Views

Recommend Documents

AutoFDO: Automatic Feedback-Directed ... - Research at Google
about the code's runtime behavior to guide optimization, yielding improvements .... 10% faster than binaries optimized without AutoFDO. 3. Profiling System.

Efficiency in a Directed Search Model with Information Frictions and ...
Mar 31, 2014 - We show that the directed search equilibrium is not constrained efficient in a dy- namic setting .... complement them with the publicly available information. Thus, the ...... correspondence T1,τ on the domain Iτ as. T1,τ (x)=(qτ .

Lightweight Feedback-Directed Cross-Module ... - Research at Google
Apr 28, 2010 - IPO needs FDO is supported by the winning 12% performance (p) increase in Table 1, ..... 1.4. 403.gcc. 143. 216. 82. 13. 2.5. 365. 65. 18. 3.6. 524. 54. 26. 4.4. 429.mcf. 11. 1. 10. 2. 1.1 ..... IEEE Computer Society. ISBN 0-7695-.

Advanced information feedback in intelligent traffic systems
Advanced information feedback in intelligent traffic systems. Wen-Xu Wang,* Bing-Hong Wang, Wen-Chen Zheng, Chuan-Yang Yin, and Tao Zhou. Department ...

Money with partially directed search
20Einzig (1966) shows that independent monetary systems have existed on a ... S. sales by interactive direct marketing (in person or by phone) are only 8% of ...

Money with partially directed search
5As a modern illustration, energy companies know that a lot of oil is produced in the Middle East, although they did not pay ..... the only alternative is autarky.

System with user directed enrichment and import/export control
Jan 12, 2010 - Software Engineering and Computer Science, Blekinge Institute of ... ment of Computer Science and Business Administration, Roneby,.

Money with Partially Directed Search
Email: [email protected]. 1 .... act on that additional information. ..... 24 U.S. sales by interactive direct marketing (in person or by phone) are only 8% of ...

Price and Match-Value Advertising with Directed ...
Jul 5, 2015 - Second, we how that with price advertising equilibrium prices and profits decrease in search costs. With higher search costs consumers are ...

Equilibrium Directed Search with Multiple Applications! - CiteSeerX
Jan 30, 2006 - labor market in which unemployed workers make multiple job applications. Specifically, we consider a matching process in which job seekers, ...

Estimating the directed information to infer causal ... - Springer Link
... 15 December 2009 / Revised: 13 May 2010 / Accepted: 21 May 2010 / Published online: 26 June 2010 .... to infer the directed information between two point.

Output feedback control for systems with constraints and ... - CiteSeerX
(S3) S is constrained controlled invariant. Our goal is to obtain conditions under which there exists an output feedback controller which achieves constrained ...

Conferencing and Feedback - ciclt.net
Feb 13, 2013 - structured process for facilitating ... give actionable information. ... only factual information about how one is progressing in relation to a goal.

Conferencing and Feedback - ciclt.net
Feb 13, 2013 - Conferencing and feedback are required components of the. Teacher Keys Effectiveness System (TKES) and the Leader Keys. Effectiveness ...

Minted doubles performance with Channel Intelligence and ...
Performance doubled with CI optimization. Showcasing their strength ... us bring our design advantage into the search channel in a way that wasn't possible before. PLAs let ... Shopping Engine Services maximize return on ad spend, and CI's ...

Head suspension with resonance feedback transducer
Jun 9, 2006 - 360/104. 5,335,126 A * 8/1994 Tokyuama et al. ........ .. 360/104. 18 .... 1 is an isometric view of a head suspension assembly including a strain ...

/ CHANNEL(
May 1, 2009 - endpoints in a data network via a monitoring unit and a control unit. .... Nr. 21, (Oct. 1989), Cited on EP Search RepOIt for EP application.

Force feedback computer input and output device with coordinated ...
Nov 16, 2004 - Conference on Human Factors in Computing Systems, pp. 21425. Rubine, D.H. (Dec. 1991). .... terms of computational speed, memory, storage and display. However, the interface between users and the computer ...... bus 922 using, for exam

Head suspension with resonance feedback transducer
Jun 9, 2006 - Head Suspension in Hard Disk Drive; IEEE Transactions on. Magnetics; Sep. 1986 ...... UNITED STATES PATENT AND TRADEMARK OFFICE.

/ CHANNEL(
May 1, 2009 - Huang, G M., et al., “A New Had Algorithm for Optimal Routing of. Hierarchically ... Kieser, H “Software fur eine digitale Vermittlungsanlage mit modularem ... egies between Lost-call-cleared and Reservation Traf?c”, Proceed.

directed matching and monetary exchange1
including, e.g., agent 1 abandoning 2, leading to [f1g; f2g; f3; 4g]. We also ... to check bilateral deviations, including agents 1 and 3 abandoning their current.

Minimax Robust A Priori Information Aware Channel ... - IEEE Xplore
but also in the estimate of interference plus noise covariance matrix. An important class of linear equalizers, is the a priori infor- mation aware equalizers, where ...