The mathematics behind wireless communication Øyvind Ryan June 2008 CMA 2007

advertisement
CMA 2007
The mathematics behind wireless communication
Øyvind Ryan
June 2008
Øyvind Ryan
The mathematics behind wireless communication
CMA 2007
The mathematics behind wireless communication
Questions and setting
In wireless communication, information is sent through what is
called a channel.
The channel is subject to noise, so that there will be some loss
of information.
How should we send information so that there is as little
information loss as possible?
How should we dene the capacity of a channel?
Can we nd an expression for the capacity from the
characteristics of the channel?
Øyvind Ryan
The mathematics behind wireless communication
CMA 2007
The mathematics behind wireless communication
What is information?
Assume that the random variable X takes values in the
X = {α1 , α2 , ...}.
pi = Pr (X = αi ).
alphabet
Set
How can we dene a measure H for how much
choice/uncertainty/information is associated with each outcome?
Shannon [1] proposed the following requirements for H :
1
2
H should be continous in the pi .
1
If all the pi are equal (pi = ), then H should be an increasing
n
function of n (with equally likely events there is more
uncertainty when there are more possible events).
3
If a choice can be broken down into successive choices, the
original H should be the weighted sum of the individual values
{α1 , α2 , α3 } can rst be split into
{α1 , {α2 , α3 }}, followed by an alternative
{α2 , α3 }.
of H : A choice between
choice between
choice between
Øyvind Ryan
a
The mathematics behind wireless communication
CMA 2007
The mathematics behind wireless communication
Entropy
Denition
The entropy of X is dened by
H (X )
= H (p1 , p2 , ...) = −
X
i
pi log2 pi
The entropy is measured in bits.
Shannon showed that an information measure which satises the
requirements of the previous foil, necessarily has this form!
If p1
= 12 , p2 = 31 , p3 =
foil can be veried as
H
1
2
where the weight
p2
+ p3 =
1
2.
1
1
3
6
, ,
1
6 , the weighting described on the previous
=H
1
2
,
1
2
+
1
2
H
2
3
,
1
3
,
1
2 appearing on the right side is computed as
Øyvind Ryan
The mathematics behind wireless communication
CMA 2007
The mathematics behind wireless communication
Shannon's source coding theorem
We would like to represent data generated by the random variable
X in a shorter way (i.e. compress). Shannon's source coding
theorem addresses the limits of such compression:
Theorem
Assume that we have independent outcomes of the random variable
X (= x1 x2 x3 · · · ) The average number of bits per symbol for any
lossless compression strategy is always greater than or equal to the
entropy H (X ).
The entropy H is therefore a lower limit for achievable
compression.
The theoretical limit given by the entropy is also achievable.
In a previous talk, I focused on methods for achieving the limit
given by the entropy (Human coding, arithmetic coding).
Øyvind Ryan
The mathematics behind wireless communication
CMA 2007
The mathematics behind wireless communication
Sketch of Shannon's proof
(n)
There exists a subset A
such that
(n)
The size of A
is
of all length-n sequences
≈ 2nH (X )
(x1 , x2 , ..., xn )
(which can be small when
compared to the number of all sequences).
(n)
Pr (A
) > 1 − .
(n)
A is called the typical set, and consists of all (x1 , x2 , ..., xn ) with
−
1
n
n)
log2 (p (x1 , x2 , ..., x
(=empirical entropy)
close enough to the actual entropy H (X ). Shannon proved the
source coding theorem by
1
assigning codes with a (smaller) xed length to ALL elements
in the typical set,
2
assigning codes with another (longer) xed length to ALL
elements outside the typical set,
3
letting n
→ ∞,
and
→ 0.
Øyvind Ryan
The mathematics behind wireless communication
CMA 2007
The mathematics behind wireless communication
What is a communication channel?
That A communicates with B means that the physical acts of
A induce a desired physical state in B .
This transfer of information is subject to noise and the
imperfections of the physical signaling process itself.
The communication is succesful if the receiver B and the
transmitter A agree on what was sent.
Denition
(X , p (y |x ), Y), consists of two nite
and Y (the output alphabet), and a
p (y |x ) that expresses the probability of
A discrete channel, denoted by
sets
X
(the input alphabet)
probability transition matrix
observing the output symbol y given that we send the symbol x .
The channel is said to be memoryless if the probability distribution
of the output depends only on the input at that time, and is
conditionally independent of previous channel inputs and outputs.
Øyvind Ryan
The mathematics behind wireless communication
CMA 2007
The mathematics behind wireless communication
A general scheme for communication
/
W
W
Xn
Encoder
/ Channel
p (y |x )
∈ {1, 2, ..., M }
Yn
/
Decoder
/Ŵ
is the message we seek to transfer via the
channel
n : {1, 2, ..., M } → X n , taking values
n
n
n
of size M (X (1), X (2), ..., X (M )).
Y n : Y n → {1, 2, ..., M }. This is a
The encoder is a map X
in a codebook from
Xn
The decoder is a map
deterministic rule that assigns a guess to each possible
received vector.
Ŵ
∈ {1, 2, ..., M }
is the message retrieved by the decoder.
n is the block length. It says how many times the channel is
used for each transmission.
M is the number of possible messages. A message can thus be
represented with
dlog2 (M )e
Øyvind Ryan
bits.
The mathematics behind wireless communication
CMA 2007
The mathematics behind wireless communication
The encoder/decoder pair is called a
(M , n)-code
(i.e. codes
where there are M possible messages, n uses of the channel
per transmission).
When the encoder maps the input to codewords in the data
transmission process, it actually adds redundancy in a
controlled fashion to combat errors in the channel. This is in
contrast to data compression, where one goes the opposite
way, i.e. removing redundancy in the data to form the most
compressed form possible.
The basic question is how one can construct an
encoder/decoder pair, such that there is a high probability that
the received message Ŵ equals the transmitted message W ?
Øyvind Ryan
The mathematics behind wireless communication
CMA 2007
The mathematics behind wireless communication
Denition
Let λW be the probability that the received message Ŵ
is dierent
from the sent message W . This is called the conditional probability
of error given that W was sent. We also dene the maximal
probability of error as
λ(n) = maxW ∈{1,2,...,M } λW .
Denition
The rate of an
(M .n)-code
is dened as R
bits per transmission.
=
log2 (M )
n
, measured in
Denition
A rate R is said to be achievable if there for each n exists a
(d2nR e, n)-code,
n λ(n) = 0
such that lim →
(i.e. the maximal
probability of error goes to 0).
Denition
The (operational) capacity of a channel is the supremum of all
achievable rates.
Øyvind Ryan
The mathematics behind wireless communication
CMA 2007
The mathematics behind wireless communication
Shannon's channel coding theorem
Expresses the capacity in terms of the probability distribution of the
channel, irrespective of the use of encoders/decoders.
Theorem
The capacity of a discrete memoryless channel is given by
C
= max I (X ; Y ),
q(x )
where X /Y is the random input/output to the channel, with X
having distribution q (x ) on
X.
Here I (X ; Y ) is the mutual information between the random
variables X and Y , dened by
I (X ; Y )
=
X
x ,y
p (x , y ) log2
p (x , y )
p (x )p (y )
,
(1)
where p (x , y ) is the joint p.d.f. of X and Y .
Øyvind Ryan
The mathematics behind wireless communication
CMA 2007
The mathematics behind wireless communication
Sketch of proof I
We generalize the denition of the typical set (from the proof of
the source coding theorem) to the following:
The jointly typical set consists of all jointly typical sequences
((x n ), (y n )) = ((x1 , x2 , ..., xn ), (y1 , y2 , ..., yn )),
dened as those
sequences where
1
the empirical entropy of
actual entropy H (X ),
2
the empirical entropy of
actual entropy H (Y ),
3
(x1 , x2 , ..., xn )
is close enough to the
(y1 , y2 , ..., yn )
is close enough to the
the joint empirical entropy (−
Qn
1
n log2 ( i =1 p (xi , yi )))
((x1 , x2 , ..., xn ), (y1 , y2 , ..., yn )) is close enough to the
joint entropy H (X , Y ) dened by
XX
H (X , Y ) = −
p (x , y ) log2 p (x , y ),
of
actual
x ∈X y ∈Y
where p (x , y ) is the joint distribution of X and Y .
Øyvind Ryan
The mathematics behind wireless communication
CMA 2007
The mathematics behind wireless communication
Sketch of proof II
(n )
The jointly typical set is, just as the typical set, denoted A
. It
has the following property similar to the corresponding properties
for the typical set:
1
(n)
The size of A
is approximately
≈ 2nH (X ,Y )
(which is small
when compared to the number of all sequences).
2
(n)
Pr (A
)→1
as n
→ ∞.
Øyvind Ryan
The mathematics behind wireless communication
CMA 2007
The mathematics behind wireless communication
Sketch of proof III
The channel coding theorem can be proved in the following way for
a given rate R
1
< C:
Construct a randomly (dictated by some xed distribution of
nR from X n . Dene
{1, ..., 2nR } into this set.
the input) generated codebook of length 2
the encoder as any mapping from
2
Dene the decoder in the following way
if the output (y1 , y2 , ..., yn ) of the channel is jointly typical
with a unique (x10 , ...xn0 ), dene (x10 , ...xn0 ) as the output of the
decoder
Otherwise, the output of the decoder should be some dummy
index, declaring an error.
3
One can show that, with high probability (going to 1 as
n
→ ∞),
(x1 , x2 , ..., xn ) is jointly
(y1 , y2 , ..., yn ). The expression for the
the input to the channel
typical with the output
mutual information enters the picture when computing the
probability that the output is jointly typical with another
sequence, which is
≈ 2−nI (X ;Y ) .
Øyvind Ryan
The mathematics behind wireless communication
CMA 2007
The mathematics behind wireless communication
More general channels I
In general, channels do not use nite alphabet inputs/outputs.
The most important continous alphabet channel is the
Gaussian channel. This is a time-discrete channel with output
Yi at time i given by
Yi
Xi is input, Zi
∼ N (0, N )
= Xi + Zi .
noise (Gaussian, variance N ).
Capacity can be dened in a similar fashion for such channels
The capacity can be innite, unless we restrict the input. The
most common such restriction is a limitation on its variance.
Assume that the variance of the input is less than P . One can
then show that the capacity of the Gaussian channel is
1
2
log2
1
+
P
N
,
and that the capacity is achieved when X
Øyvind Ryan
∼ N (0, P ).
The mathematics behind wireless communication
CMA 2007
The mathematics behind wireless communication
More general channels II
In general, communication systems consist of multiple
transmitters and receivers, talking and interfering with each
other.
Such communication systems are described by a channel
matrix, whose dimensions match the number of transmitters
and receivers. Its entries is a function of the geometry of the
transmitting and receiving antennas.
Capacity can be described in a meaningful way for such
systems also. It turns out that, for a wide class of channels,
the capacity is given by
C
where
=
1
n
log2 det
In + ρ HHH
m
H is the n × m channel matrix, n,m is the number of
receiving/transmitting antennas,
P
N
1
ρ
is signal to noise ratio (like
for the Gaussian channel).
Øyvind Ryan
The mathematics behind wireless communication
CMA 2007
The mathematics behind wireless communication
Active areas of research and open problems
How do we construct codebooks which help us achieve rates
close to the capacity? In other words, how can we nd the
input distribution p (x ) which maximizes I (X ; Y ) (the mutual
information between the input and the output)? Such codes
should also be implementable. Much progress made in recent
years. Convolutional codes, Turbo codes, LDPC (Low-Density
Parity Check) codes.
Error correcting codes: These codes are able to detect where
bit errors have occured in the received data. Hamming codes.
What is the capacity in more general systems? One has to
account for any number of receivers/transmitters, any type of
interference, cooperation and feedback between the sending
and receiving antennas. General case far from being solved.
Øyvind Ryan
The mathematics behind wireless communication
CMA 2007
The mathematics behind wireless communication
Good sources on information theory are the books [2] (which
most of these foils are based on), and [3].
Related courses at UNIK: UNIK4190, UNIK4220, UNIK4230.
Related courses at NTNU: TTT4125, TTT4110.
This talk is available at
http://heim.i.uio.no/∼oyvindry/talks.shtml.
My publications are listed at
http://heim.i.uio.no/∼oyvindry/publications.shtml
Øyvind Ryan
The mathematics behind wireless communication
CMA 2007
The mathematics behind wireless communication
C. E. Shannon, A mathematical theory of communication,
The Bell System Technical Journal, vol. 27, pp.
379423,623656, October 1948.
T. M. Cover and J. A. Thomas, Elements of Information
Theory, second edition.
Wiley, 2006.
D. J. MacKay, Information Theory, Inference, and Learning
Algorithms.
Cambridge University Press, 2003.
Øyvind Ryan
The mathematics behind wireless communication
Download