LIST DECODING FOR NOISY CHANNELS

advertisement
DOCUMENT ROOM,
~T
ROOK 36-412
RESEARCH LABORATORY CF ELECTRONICS
MASSACHUSETTS INS' ! UTE OF TECHVOCLO '
LIST DECODING FOR NOISY CHANNELS
PETER ELIAS
Technical Report 335
September 20, 1957
RESEARCH LABORATORY OF ELECTRONICS
MASSACHUSETTS INSTITUTE OF TECHNOLOGY
CAMBRIDGE, MASSACHUSETTS
Reprinted from the
1957 IRE WESCON CONVENTION
Part 2
PRINTED IN
_
THE U.S.A.
RECORD
The Research Laboratory of Electronics is
an interdepartmental laboratory of the Department of Electrical Engineering and the
Department of Physics.
The research reported in this document
was made possible in part by support extended the Massachusetts Institute of Technology,
Research Laboratory of Electronics, jointly
by the U. S. Army (Signal Corps), the U. S.
Navy (Office of Naval Research), and the
U. S. Air Force (Office of Scientific Research,
Air Research and Development Command),
under Signal Corps Contract DA36-039-sc64637, Department of the Army Task 3-9906-108 and Project 3-99-00-100.
_
LIST DECODING FOR NOISY CHANNELS
PETER ELIAS
Technical Report 335
September 20, 1957
RESEARCH LABORATORY OF ELECTRONICS
MASSACHUSETTS INSTITUTE OF TECHNOLOGY
CAMBRIDGE, MASSACHUSETTS
Reprinted from the
1957 IRE WESCON CONVENTION RECORD
Part 2
PRINTED IN THE U.SA.
-_II--UXI
LIST DECODING FOR NOISY CHANNELS
Peter Elias
Department of Electrical Engineering and Researcn Laboratory -of Electronics
Massachusetts Institute of Technology, Cambridge, Massachusetts
Summary.
Shannon's fundamental coding
q = 1-p that it will be received correctly.
theorem for noisy channels states that such a
Successive transmitted digits are treated by the
channel has a capacity C, and that for any trans-
channel with statistical independence.
mission rate R less than C it is possible for the
The capacity C of this channel is given by
receiver to use a received sequence of n symbols
C = 1
to select one of the 2n R possible transmitted
H(p)
bits/symbol
(1)
where
H(p) = -p log p - q log q
sequences, with an error probability Pe which can
(2)
be made arbitrarily small by increasing n, keeping
is the entropy of the p,q probability distribution.
R and C fixed.
C is plotted in Fig. 1 as a function of p.
Recently upper and lower bounds
Note
have been found for the best obtainable Pe as a
that if p ? q, the output symbols can be rein-
function of C,R and n. This paper investigates
terpreted, an output 1 being read as a 0 and vice
this relationship for a modified decoding pro-
versa,
cedure, in which the receiver lists L messages,
considered.
rather than one, after reception.
out the paper are taken to the base 2.)
In this case
for given C and R, it is possible to choose L
so that only values of p C q will be
(The logarithms in Eq. 2 and through-
C can be given two interpretations.
First,
large enough so that the ratio of upper and lower
if 0's and l's are transmitted with equal proba-
bounds to the error probability is arbitrarily
bility and statistical independence, then one bit
near to 1 for all large n. This implies that for
per symbol is supplied at the
large L, the average of all codes is almost as
average of H(p) bits per symbol of equivocation
good as the best code, and in fact that almost
remains in the output.
all codes are almost as good as the best code.
coding theorem for noisy channels, if the input
Second, by the fundamental
information rate is reduced from 1 bit per symbol
Introduction
The binary symmetric channel (abbreviated
BSC) is illustrated in Fig. 1.
input, and an
to any rate R < C bits per symbol, it is possible
It accepts two
for the receiver to reconstruct the transmitted
input symbols, 0 and 1, and produces the same two
sequence of input symbols with as low an
symbols as outputs.
probability,as is desired, at the cost of a&delay
There is probability p that
rror
an input symbol will be altered in transmission,
of n symbols between transmission and decoding,
from 0 to 1 or from 1'to 0, and probability
where n must be increased to reduce the error
probability.
This work was supported in part by the Army
(Signal Corps), the Air Force (Office of Scientific Research, Air Research and Development
Command), and the Navy (Office of Naval Research).
In block coding, the input information rate
is reduced by using only a restricted number,
94
_
__
___
_
M = 2R, of the 2n possible sequences of length n.
therefore even the exponent of Popt is not known,
This gives an average rate of (l/n)log M - R bits
although it is
per symbol; the sequences being used with equal
rates, it can be shown that the' exponent of Popt
probability.
is
The receiver then lists the single
Finally, for very low
definitely different from that for either P
av
or Pt'.
transmitter sequence which is most probable, on
As can be seen-from this brief description,
the evidence of the received sequence of n noisy
symbols.
bounded.
It is sometimes wrong, with a probability
the behavior of Popt is complicated, and not well-
which depends on the particular set of sequences
defined, even for the channel which has been
which have been selected for transmission.
studied most intensively.
define Popt
We
The present paper
discusses a somewhat more complicated transmission
a function of n, R and C, as the
minimum average probability of error for a block
procedure for use with the BSC, which has the
code: i.e. as the average error probabilityforthe
advantage that the description of its error prob-
code whose transmitter sequences have been chosen
ability behavior is much simpler.
most
block coding with list decoding - leaves a little
udiciously, the average being taken over the
equivocation in the received message.
different transmitter sequences.
receiver provides a list of L messages, L <
for small n, a
number are known (see Slepian (3,4)).
Instead of
a unique selection of one message out of M, the
Unfortunately we do not know the optimum
codes for large values of n:
Our procedure -
However it
M.
As soon as a little residual equivocation is per-
is possible to define a lower bound to Popto
mitted, it turns out that P
has the same exponent
av
which we denote by Pt. which is a fairly simple
as Pt' and thus as Popt' for all transmission rates
function of the code and channel parameters.
and not just those which are near capacity.
It
This
is also possible to define Pa , the average error
is essentially Theorem 1.
probability of all possible codes, which is an
as L increases, Pa
upper bound to Popt'
that Popt itself, and not just its exponent,
The behavior of Popt is
discussed in some detail in references (1,2).
actually approaches Pt. so
becomes well-defined.
As
It also turns out that
This is the content of
the transmission rate approaches the channel
Theorem 2.
capacity, Pav approaches Pt. so that the error
the cause of detection errors in a system using
probability Popt is well defined.
list detection can be interpreted simply: mistakes
For somewhat
As a result of this simplification,
lower rates the two bounds differ, but for fixed
are made when, in a particular block of n symbols,
R and C their ratio is bounded above and below by
the channel gets too noisy.
constants as n increases.
ordinary block coding, on the other hand, errors
Since each is decreas-
In the case of
ing exponentially with n, at least the exponent of
also occur because of unavoidable weak spots in
Popt is well-defined.
the code, which causes the complication in the
For still lower rates, how-
description of Popt for such codes.
ever, Pav and Pt decrease with different exponents
as n increases with C and R fixed.
In this region
95
_I____·______
I_
I
_
__
_
Operation
being ranked in the same manner as the noise
The operation of a block code with list
sequences.
The received sequence is decoded by
decoding is illustrated in Fig. 2. There are M
means of a decoding codebook, as illustrated in
equiprobable messages, represented by the integers
Fig. 3, which provides a list of L = 3 messages
from 1 to M, and the source selects one, say the
for each received sequence.
integer m = 2.
This is coded into a sequence
The decoding codebook is constructed by list-
um = U2 of n binary digits, 011 in the example,
ing, after each received sequence, the L messages
for which n = 3.
whose transmitter sequences are converted into the
The sequence um is transmitted
over the noisy BSC.
the digits:
The channel alters some of
given received sequence by noise sequences of the
its effect is to add a noise sequence
lowest possible rank.
Because of the equal prob-
wr, in this case the sequence 010, to the trans-
ability of the transmitter sequences and the
mitted sequence um to produce the received noisy
statistical independence of transmitter sequence
sequence,
= 001.
2
(The addition is modulo 2,
and noise sequence in the BSC, the L messages
digit by digit: 1+1 = 0+0 = 0, 1+0 = 0+1 = 1.)
which are most probable on the evidence of the
The received sequence is decoded into a list of
received sequence are those which are converted
L messages - L = 3 in the Figure.
into the received sequence by the L most probable
If the list
includes the message which the source selected,
noise sequences.
the system has operated correctly and no error has
will be listed in the .codebook, except that the
occurred.
ranking of the noise sequences will provide an
If not there has been an error.
Since
These are just the ones which
2 is on the list 1,2,3, the transmission illustrated
unambiguous decision when several noise sequences
was successful.
have the same probability.
The behavior of the system is determined by
the choice of u I ,...u
sequences.
m ....
,uM,
This follows because
the probability Prob[wr] of a noise sequence wr
the transmitter
with k l's is just
An encoding codebook, illustrated for
Prob[wr
pkqn-k
qn (p/q)k
(3)
the example-in Fig. 3, assigns M = 4 such se-
which is monotone decreasing in k for p <
quences each of length n = 3, to the four messages
while rank is monotone increasing with k.
m = 1,2,3,4.
The 2 = 8 possible noise sequences
The behavior of the system under all possible
of length n are ranked, by an index r, so that the
sequence with no l's, w
q,
circumstances is illustrated in the table of
= 000, comes first, a
possible events in Fig. 3.
The message sequence
sequence with k l's precedes one with k+l l's, and
U2 and the noise sequence w3 determine a row and a
sequences with the same number of l's
column in the table.
by their size as binary numbers.
are ranked
In the example
The received sequence v2 =001
which is their digit-by-digit sum modulo two is
the noise sequence w3 = 010 is added to the trans-
entered at the intersection.
mitted sequence to produce the received sequence
v2 : the 2n possible received sequences vl...,v2n
book shaws that v2 will be decoded as the list
The decoding code-
1,2,3 which includes the message m = 2 which was
96
I
-
-
-
-
--
-'
actually transmitted.
Therefore the number of circled entries
sequence.
This entry in the table is
therefore circled, as an event which the system
in row m is equal to the total number of times
will handle without error, and so are all such
n
that the integer m appears in all of the 2 lists
of L messages in the decoding codebook.
The probability of error for the system
events.
total of L
is then just the sum of the probabilities of the
2n
n
uncircled events, which are all of the entries in
total of M-2
the last two columns.
of
2 n(l
Thus a
entries are circled, out of the
entries in the table, or an average
R) circles per message.
Clearly the minimum probability of error
Rate of Transmission
would be obtained if all of the circles were as
The average amount of information received
over this kind of system when no error has been
far to the left as possible in the table.
made depends on the probabilities of the different
true for the code in Fig. 3, but will not be possi-
messages on the decoded list.
ble for other values of M, L and n.
To make it constant,
This is
However it
we assume that the messages on the list are
provides a lower bound to error probability in any
shuffled in order before being given to the sink,
case.
so that they are equiprobable to him.
code (perhaps unrealizeable) with such a table.
Then the
We define Pt as the error probability for a
Let r
rate is
be the last column in which any circled
entries appear:
R =
log M - log L
n
(4)
bits/symbol,
rlM 2 L2
n
> (rl-)M, or
and to transmit at rate R requires that
(6)
r1
M = L.2nR
messages be transmitted.
>
2 n(1-R)>
1
(5)
Then Pt is bounded:
In the example of Figs.2
and 3, we have M = 4, L = 3, n = 3, and R = (1/3)
n
n
2
2
log (4/3)
- 0.14 bits/symbol.
Prob[Wr]
Note that the largest M which can be used
> Pt
r=r I
Prob[wr]
(7)
r=rl+l
effectively is M = 2n, since there are only that
many distinct sequences.
Thus L =
2 n(
1- R
large a list as would be of any interest.
The quantity Pt is characteristic of the
) is as
channel, and of its use at rate R in blocks of
This
would correspond to transmitting uncoded infor-
N symbols, but is independent of the code used.
mation-i.e., all possible sequences.
The channel parameters p,q enter in evaluating
Error Probability: Lower Bound
Prob[wr ]
n
In any row of the table in Fig. 3, all 2
ProbEwr] = pk(r)qn-k(r)
(8)
possible received sequences appear, since each of
the
2n
where k(r) is determined by means of the binomial
noise sequences produces a distinct received
coefficients
sequence when added to the given transmitter
()
= n!/j!(n-j)
,
97
_I__I___ILWLP·_III1II__--L-_·^L-LLI
,;
giving equal weights to each.
k(r)-l
k(r)
Pay can be expressed as a sum over r of the
(9)
(/)
(in
j =0
probability of occurrence of the rth noise sequence
wr, times the ensemble average conditional prob-
Using Eq. 8, and introducing k1 as an
abbreviation for k(r1) defined through Eq. (9), we
ability QL(r) of making a decoding erro
have the following bounds for Pt:
noise sequence wr occurs:
2n
L
pk(r)qn-k(r) > Pt
r=r 1
av
pk(r)n-k(r)
when the
r(12)
(10)
r=rl+l
Now we can bound Pav above and below.
Since
it is an average of quantities all of which are
n
(n) pkqn-k
(kI pq
p
greater than Pt. Pt is a lower bound.
>
kk
k=kl
(l+l
n-k
I(11)
Since QL(r)
is a conditional probability and thus less than
1rr~
one, it can be bounded above by one for a portion
of the range of summation.
The first of these is needed for Theorem 2.
Thus, using Eq. 10,
The second, which is simpler to evaluate, suffices
pk(r) n-k(r)((r
for Theorem 1.
Error Probability: Upper Bound
)
2n
k(r) n-k(r)
An upper bound to the best attainable average
(13)
r+l l
error probability is obtained by Shannon's procedure of random coding - i.e. by computing the
r,
average error probability for all possible codes.
pk(r)n-k(r)%(r) +
This turns out to be much easier than computing
More crudely, we define Q
the error probability for any one particular code.
(k):
Since a code is uniquely determined by its set
ul,...,uM of transmitter sequences, there are as
q(k(r)) =
nR
length n, as there are selections of M = L.2
2n
)
Q(r) ,
(14)
j =0
many codes possible, for rate R, list size L and
objects out of
(n)
(
by Eq. 9 and the fact that QL(r) is monotone
increasing in r (See Eq. 16, below).
possibilities, repetitions
Then
permitted, or
(15)
k0
2 nM
2 nL*
k=O
2n R
Finally, an explicit expression can be given
We define P as the average error
av
probability, averaged over all transmitter
for Q(r).
sequences in any one code and over all codes,
occurred in a transmission, the probability of a
in all.
Given that the rth noise sequence has
98
_
_____
The value of L
decoding error is just the probability that L or
0
needed for the theorem is
more of the other M-1 transmitter sequences differ
log(q/p)
L
from the received sequence by error sequences of
rank
0
=
-1 .
log(ql/pl)
(19)
r . This is the probability of L successes
Corollary
in M-1 statistically independent tries, when the
Under the same conditions, the exponent of
average probability of success is r/2 n, for in the
ensemble Gffcodes the different transmitter sequences are si.seically independent.
P
opt
as n increases is
Thus QL(r)
is a binomial sum:
lim l/n)log opt
cs
= nlim
(l/n)log a)
=
QL(r)
L(Ml )
=
(r/2 n) (1-r/2n)
. (16)
lim
(l/n)lo
-gPt
(20)
)
3JL
=
(P1)-
(p)
-
(pl-p)log(q/p)
The Parameter pi . Theorem 1 .
The proof of the corollary given the theorem
It would be desirable to express the error
probability directly as a function of rate, capaci-
is direct.
ty, block size n and list size L.
shows that Pv and Pt have a common exponent.
However it
Taking logarithms and limits in Eq. 18
turns out to be more convenient for Theorem 1 to
Since they bound Popt above and below, it shares
use, instead of rate, the parameter p
the same exponent.
where k
= k(rl) is defined by Eq.9.
= kl/n,
large n the relation is simple:
the exponent of Pt, as found in references (1,2).
Fixing P1
and n determines the rate R through Eq. 6.
The value given in Eq. 20 is
It is illustrated geometrically in Fig. lb.
For
Proof of Theorem 1.
if we define the
Bounding Pav in terms of Pt requires bounding
rate R1 by the equation
the sum in Eq. 15.
R1 = 1 - H(pl1 )
For the last term in the sum we
(17)
use the bound Q(kl) =1
For the earlier terms
as illustrated in Fig. lb, then for fixed Pi,
we use the bound of Eq. A6, derived in Appendix A.
R approaches R1 from above as n increases.
Eq. 15 then becomes
This
is shown in references (1,2).
We can now state
PaPay -%
eL (n P-
j
e
pkqnk
{ (k
()
/ (knq
/
k=o
Theorem 1
(n
Q+
)
Given a BSC with error probability p, and
kI n-k I
p q
(21)
+ Pt
given any P1 , with o p (p1< i, and any
L > Lo(p,p), the ratio of PaV to Pt for list
decoding with list size L is bounded, independent
of n:
1
Now the sum in Eq. 21 can be bounded by a
Pav/Pt
=
A(pp
1 ,L)
(18)
geometric series, which sums:
99
U·
_ I(~U
~ ^ )__I_
---- y ~ C~~^_
I·-tI
.IY-.. ·-ll·_1X----·_
I-~Y-.
the weakness of the bound of Eq. A6 on Q(k) for
kl-l
kqn-k
()
)
p
k near k1 . Actually A(p,pl,L) decreases as L
/
(kllY
L+l
k-
increases.
k=-o
In fact we have
Theorem 2
kll n-kl+l
P
(22)
Given a BSC with capacity C, and a rate R
q
i - (q/p)(pl/ql)L~
9
with 0 < R < C < 1 , and given any e >
0 ,it
is possible to find an L(e) so large that for all
where we have used k1 = npl and n-k
= nql.
I he
n > n(L),
convergence of the series is guaranteed by thee
requirement that L > L as given by Eq. 19,
1
s0
PaV/Pt
=
.
1+
(26)
that the denominator on the right in Eq. 22 ii
s
positive.
Corollary
Substituting Eq. 22 in Eq. 21 and regrou]ping,
For sufficiently large L and n, almost all
codes are almost as good as the best code.
Pa Is
av
n )
lkn-k1
1 +
Given the theorem, the corollary follows by
\k)
(23)
an argument of Shannon's in reference (5). The
L
+qP
difference between Popt and the error probability
+ Pt
Pql
1 - (q/p)(pl/ql)L+
of any code is positive or zero, and the average
value of the difference, by Theorem 2, is < c.
Now, from Eq. 11,
Thus only a fraction 1/T of the codes may have
such a difference
t
=t
kl+l
k
p
+
n-k -l
(l-l/T) have an error probability
q
(24)
= (P/q)
k)
Te , and at least a fraction
QED.
k1 n-k1
P q
(l+Tc)Pop t ,
The result here is stronger than in Shannon's
application because of the nearby lower bound.
Proof of Theorem 2.
Substituting Eq. 24, read in reverse, in
For a constant rate R,p1 = kl/n is not a
Eq. 23 gives
constant, but depends on n . However for sufficiently large n it is arbitrarily close to the
Pav/Pt
(q/p)
1
+
(25)
limiting value given by setting R
L
+
Let p
qPl
Pq
1
-
(q/ppl
/
L+
(q/p)(p 1 /q1 )LJ
be this limit.
R1 in Eq. 17.
Then
+
QED.
lim
Theorem 2.
The proof of Theorem 1 suggests that
A(p,pl,L) increases with L, but this is due to
100
kl*
/
~k
-1)
=
(27)
Choose J so that
(pl/,l)
=
-(l+a)
r n<
1 k
(28)
,a > O .
11
(33)
p/l
Substituting this in Eq. 32, we have
Then breaking the sum in Eq. 13 at
k1-J
L
°
(29)
( k)o
C1
r1
Q(r) < p
k=o
q
L
r=r
0
and using the bound of Eq. A6 with the substi-
(34)
C2
"tution of Eq. 27 below this point gives a term
n
k,
which is bounded by
(n
-(L+l)a
pkqn-kl
Now the denominator under C1 increases
1-(q/p)(p*/q*L+l
/
(30:
arbitrarily with L, and the denominator under C2
LP)(l
ql)
increases arbitrarily with n.
Next, the portion of the sum in Eq. 13 which
is
between r,
bounded.
in Eq. 34 becomes negligible for large n and L
given in Eq. 29, and rl, is
kn-k
Since p q
Thus the summation
compared to the term
is monotone decreasing in
k1 n-k
q
k, we have
n
k
(35)
(31)
So does the expression in Eq. 30 which bounds the
Z:
pk(r)qn-k(r)QL(r)
(q/p)Jpklqn-kl
(r).
remainder of the sum in EQ. 13, because of its
r=r
r=rr
exponential factor.
We now bound QL(r) by Eq. B4, derived in
appendix B.
true for Pt
This gives
r1
o
e-(L/2)(rl
r=r
Thus in the limit the right
Conclusion.
- r )
/r
Two points should be noted in conclusion.
o
First, if
if
1+
in fact of the order of the
13 approaches Pt, QSD.
2
r2
(r)
r=r
which is
expression in Eq. 35.
side of E.
But from Eq. 24, this is not
f-Lr
2
1
2
/2r
dr
L grows exponentially with n - that is,
there is a residual percentage of equivocation,
rather than a
esidual amount - then the ratio
0
, C
1 +
(r 1/2) V&-IL
From Eq. 9, using a geometric series to
Pav /Pt approaches 1 exponentially,
since\Lf
then a negative exponential
Second,
in n .
discussion of Theorem 2 shows that Popt is
bound the sum of binomial coefficients as in
near Pt
references (1,2,6) gives
them are.
is
the
very
but does not say Just how big either of
Pt is of course given by Eq. 9, with r
101
-I~IP~^___^^_111_·
-
___._
Now from Eq.'s 6 and 9,
kl-l
(1-) =
( n1
defined by -Eq. 6, and the exponent is given by
Eq. 20.
But no really tight bounds for Pt or
(A4)
Popt in terms of C, R, and finite n and L have
been given.
And, by term-by-term comparison, for k - k 1-l,
These take a good deal more of
detailed algebra, which will be presented elsewhere.
Acknowledgements.
k -l
k
1- 1
Theorem 1 was discovered independently by
E
Professor J. M. Wozencraft in the course of his
doctoral research.
J-o
(in)
j=o
.e.
=
(k
/ (k 1 1) , (A5 Y
It does not appear in his
thesis (10) since it was not directly relevent to
so from Eq. A2 and the definition in Eq. 14 we
the final topic, but he will publish a note on the
have for k - kl-,
subject shortly.
Shannon introduced the author to
the work of Chernoff and Cramer (7,8), which makes
the bound of Appendix B possible.
Without this
kind of bound the argument for Theorem 2 is even
more tedious.
QL(k) = (eL/
APPENDIX A:
)
/
{()
(kl
}
A BOUND ON Q( r).
(A6)
4
L
n
G
n
The probability of L or more successes in-M-1
tries, with probability r/2n of success on each
try, can be bounded by the product of (r/2n)L , the
probability of success in a part&cular L o
the
APPENDIX B:
M-1 tries, times the number of ways of selecting
ANO1HERO
BOUND ON
(r).
L from M-1 :
A result due to Chernoff (7), which may also
QL(r)
(M1)
(r/2n)L
be derived from Cramer (8), provides bounds on the
(Al)
tails of distributions of sums of random variables
< ( /Lt)(r/ 2n)
Stirling's lower bound to L
in terms of the moment-generating function of the
gives
parent distribution. Applying this to the binomial
distribution, Shannon (9)
(r)
)(/L(rn)
< (L/E'
i
(e /'2nL
L
)(Mr/L2n)L
kb = N%> Npa ,
Pb (a) } N
(A2)
)(r/2n(1-R),
k=kb
making use of Eq. 5. From Eq. 6 we have
QL(r) = (e/
shows that for
)(r/(rfl))L
(A3)
102
{(
)
Taking 1/N times the natural logarithm of the
right side of Eq. B1, and defining & =Pb-Pa >
QL(r)
O,
exp ( - (M/2)(rl-r)2 /rl.2n )
gives
Iexp( - (L/2)(rl-r)/2n(1-R)rl
1 )
Pb
(B4)
+ b loge(l+6/%b)
loge(1-S/)
= exp ( - (L/2)(r 1 -r)2 /r
References
(B2)
(S/J)
cs
We take qb > Pb
coefficient of 6
( )J
1
)
ED.
q1J.
Pbb-
1.
Elias, P., Coding for Noisy Channels", I.R.E.
Convention record part 4, p.37-46 (1955).
Then 1/qb < 1/Pb
in Eq. B2 is >
0.
so the
2. Elias, P., Coding for Two Noisy Channels", in
Cherry, (editor) Informtion Theory, p.61-74,
Butterworth (1956).
Then the
sum in Eq. B2 is more negative than its leading
3. Slepian, D., "A Class of Binary Signalling
Alphabets", Bell Syst. Tech. J. 25, p.203-234
(1956).
term, and
4. Slepian, D., A Note on Two Binary Signalling
Alphabets', Trans. I.R.E. PGIT IT-2, 61-74 (1956).
N
5.
L
(k)
k=kb
Pa qb
2
= exp ( - N(
/2)((lpb)
Shannon, C. E., 'The Mathematical Theory of
Communication , University of Illinois Press
(1949) see p.bO .
-
6. Feller, W., Probability Theory and its
Applications , Wiley (1950) see .126, Eq. (8.1)
+,(l/qb)) )
(B3)
<= ep ( - s 6/2pbqb)
< exp ( - N
2
/2pb)
In applying this to QL(r), we have pa = r/2
Pb
=
rl/2n
n .
= (rl-r)/2 n , and we may use N = M,
,
since increasixg M-1 to M in Eq. 16 increases the
sum.
Thus, using Eqs. 5 and 6,
7. Chernoff, H. 'A Measure of Asymptotic
Efficiency for Tests of a ypothesis Based
on a Sum of Observations', Ann. Math. Stat. 2,
(1952).
8. Cramer, . Sur un Nouveau Theoreme-Limite
de la Theorie des Probabilites' in Actualites Sci.
et Indust., no. 736, Hermsnn et cie (1938).
9. Shannon, C. ., Information Seminar Notes
M.I.T. (1956) (unpublished).
10. Wozencraft, J. M., Sequential Decoding for
Reliable Communications , Dissertation submitted
to Elect. Eng. Dept., M.I.T. (1957).
q
INPUT
OUTPUT
SYMBOLS
I
_
IE--C - H(P)
R,C t
SYMBOLS
q
THE BINARY SYMMETRIC CHANNEL
O ' I
Fig. la.
P
-
CHANNEL CAPACITY
Fig. lb.
103
·I__·I
I_^L_
I_
n SYMBOLS
O11
n SYMBOLS
L MESSAGES
TRANSMITTED SEQUENCE
NOISE SEQUENCE
RECEIVED SEQUENCE
W 010
=
1,2,3
001
OI
NOISY
CHANNEL
OPERATION
DECODED
MESSAGE LIST
RECEIVED
SEQUENCE
TRANSMITTED
SEQUENCE
MESSAGE
2
001
BLOCK CODING AND LIST DECODING
Fig. 2.
ENCODING
MESSAGE
m
I
2
,
3
4
NOISE
CODE BOOK
_
TRANSMITTER
SEQUENCE um
RANK
SEQUENCE
r
Wr
u
000
I
W
=
I
u2
0 11
2
-
w2
u3
I101
3
-
w3
U4"
I 10
4
,
5
,
6
DECODING CODE BOOK
SEQUENCES
w4
W
5 -
w6
7
-
w7
8
_
w
000
001
010
1 00
011
101
10
I II
RECEIVED
SEQUENCE v,
VI
V2
V3
V4=
V5
V6
V7
=
000
001
010
1 00
011
101
110
Vs
8
DECODED
MESSAGE LIST
1,2,3
1,2,3
1, 2,4
I, 3,4
1,2,4
1,3,4
2,3,4
2,3,4
EXAMPLE OF LIST DECODING SYSTEM
Fig. 3a.
NOISE
SEQUENCES
PROB [wr]
CIRCLED
ENTRIES
ARE
CORRECTLY
TRANSMITTER
SEQUENCES
DECODED
EXAMPLE
OF LIST DECODING SYSTEM
Fig. 3b.
104
Download