ASSIGNMENT No.1 DIGITAL COMMUNICATION – II ETIT – 312 Q1: Prove the following a) I(X;Y) = H(X) + H(Y) – H(X,Y) b) I(X;Y) = H(X) – H(X/Y) Q2: Explain the concept of amount of information. What is Information Rate? How it is measured? Q3: Channel Transition Matrix is given by [0.6 0.4; 0.1 0.9]. Draw the channel diagrams and determine the probabilities. Also find the mutual information I (X;Y) for the channel. Q4: An analog signal is bandlimited to B Hz sampled at Nyquist Rate and the samples are quantized into four levels. The quantization levels are assumed to be independent and occur with equal likelyhood. Find information rate of the source. Q5: Prove the mutual information is always a non-negative number. Q6: Why probability theory is taken into consideration for studying information theory. Q7: State and Define Shannon’s Source Coding Theorem. Q8. With the differential entropy of a random variable X defined by H(X) = -∫∞-∞ fx(x) log2 fx(x) dx Find the probability density function for which entropy H(X) is maximum. Q9. Find the differential Entropy H(X) of uniformaly distributed random variable X with pdf fx(x) = {1/a for 0 ≤ x ≤ a; 0 otherwise for (i) a =1 (ii) a = 2 (iii) a = ½ Q10. Write short note on (i) Markov Sources. (ii) Differential Entropy