.


:




:

































 

 

 

 


Find the joint entropy of coin tossing and dice tossing




Indicate a DMC(Discrete Memoryless Channel)

a)

Define a discrete memoryless channel

a)

Find a DMC (Discrete memoryless channel)

a)

Decode precisely by the (7,4) Hamming code an output vector

a) 1110000

7. Let a signal is defined on interval [2,8]. Its quantization step is equal to 0.2. Compute the number of quantization levels.

a) 30

8. Let a signal is defined on interval [-3,1]. Its quantization step is equal to 0.3. Compute the number of quantization levels.

a) 30

9. Let X be a discrete random variable with a probability mass function p(x)=(0.25,0.25,0.25,0.125,0.125)

a) 2.25

Let p(x,y) be given by

X\Y    
  1/3 1/3
    1/3

Find H(X|Y).

a) 2/3

Let p(x,y) be given by

X\Y      
  1/8 1/8 1/8
  1/8   1/8
  1/8 1/8 1/8

Find H(X,Y)

a) 3

Consider the capacity of discrete memoryless channel given by

P(X|Y) A B C D E F
A 1/6 1/6 1/6 1/6 1/6 1/6
B 1/6 1/6 1/6 1/6 1/6 1/6
C 1/6 1/6 1/6 1/6 1/6 1/6
D 1/6 1/6 1/6 1/6 1/6 1/6
E 1/6 1/6 1/6 1/6 1/6 1/6
F 1/6 1/6 1/6 1/6 1/6 1/6

a) 0

Take into account the capacity of a DMC

P(X|Y) A B C D E F
A 1/2 1/4 1/4      
B   1/2 1/4 1/4    
C     1/2 1/4 1/4  
D       1/2 1/4 1/4
E ¼       1/2 1/4
F ¼ 1/4       1/2

a) Log 3-1/2

Channel capacity is defined by

a) max I(X,Y)

Define the formula of channel capacity

a) max {H(X)-H(X|Y)}

Decode 0111110 by the (7,4) Hamming code.

a) 0111100

Calculate precisely the mean length of optional code for probabilities (0.3,0.2,0.2,0.1,0.1,0.05,0.05)

a) 2.6

18. Find H(X) for a discrete random variable with a probability mass function p(x)=(0.5,0.25,0.125,0.125)

a) 1.75

Let an experiment with dice tossing be given. Find its entropy.

a) log 2

20. On interval [-3,1] is defined a signal . The quantization step is equal to 0.1. Then the number of quantization level is

a) 90

For probability distribution (1,0,0,0) find its entropy.

a) 0

22. Let (X,Y) have next joint distribution . Find I(X,Y).

a) 0

23. For probability distribution (1/3, 1/3,1/3,0,0) find its entropy.

a) 1.585

24. Let be given the entropy H(X) of input of the channel. P(X|Y)= .

Find the conditional entropy H(X|Y).

a) 0

25. Let be given the entropy H(X) of input of the channel. P(Y|X)= .

Find the mutual information I(X,Y).

a) H(X)

26. Let be known the entropy H(X) of input of the channel P(Y|X)= .

Calculate H(Y).

a) H(X)

27. Let be given the entropy H(X) of input of the channel. P(Y|X)= .

Find the conditional entropy H(Y|X).

a) 0

A channel is given with (7,4) Hamming coding. Decode an output of a channel 1100000.

a) 1110000

Decode an output vector 0000001 of a channel with the (7,4) Hamming code

a) 0000000

The entropy of coin tossing is equal to

{0.5 0.5

0.5 0.5} C) 1

Let be given the entropy H(X) of input of the noisy channel

P(Y|X) =.Find H(Y|X).

a) 0

Which of matrices define a discrete memoryless channel

a)

Point a matrix of a discrete memoryless channel

a)

b)

 

6. Find H(X) for a discrete random variable with a probability mass function p(x)=(0.5,0.25,0.125,0.125)

b) 1.75

 

Let p(x,y) be given by

X\Y      
    1/3 1/3
      1/3
       

Find H(Y) - H(Y|X).

b) Log 3

8.

9.

10.

11.

12.

13.

14.

15.

Compute the capacity of a binary channel

P(X|Y) A B
A    
B    

 

a) 1

17. Determine the channel capacity for a) 1

P(X|Y) A B C D
A 1/2 1/2    
B 1/2     1/2
C     1/2 1/2
D   1/2 1/2  

 

a) 1

Evaluate the channel capacity of

P(X|Y) A B C D
A 1/2 1/4 1/4  
B 1/4 1/4   1/2
C 1/4   1/2 1/4
D   1/2 1/4 1/4

a)1/2

Estimate the capacity of

P(X|Y) A B C D
A 1/4 1/4 1/4 1/4
B 1/4 1/4 1/4 1/4
C 1/4 1/4 1/4 1/4
D 1/4 1/4 1/4 1/4

a) 0

Calculate precisely the mean length of optional code for probabilities (0.3,0.2,0.2,0.1,0.1,0.05,0.05)

b) 2.6

21. Let (X,Y) have next joint distribution . Find H(X|Y).

b) 1

1. The entropy of a discrete random variable with probability distribution ( , , , ) is equal to

A)

 

Find the joint entropy of coin tossing and dice tossing.

A) 1+log6





:


: 2016-07-29; !; : 258 |


:

:

, .
==> ...

1796 - | 1618 -


© 2015-2024 lektsii.org - -

: 0.019 .