Documenti di Didattica
Documenti di Professioni
Documenti di Cultura
Many thanks to Dan Costello, Shrinivas Kudekar, Alon Orlitsky, and Thomas Riedel
for their help with these slides.
Storing Shannon
Networking Shannon
Completing Shannon
Compressing Shannon
Reading Shannon
Coding
Disclaimer
Classes of Codes
convolutional codes
polar codes
How Do We Compare?
capacity
rate
block length
PN (R, C)
complexity
1
log(PN (R, C))
N N
E(R, C) = lim
error exponent
10-2
PN (R, )
10-3
10-4
10
1/
0
N in
cre
asin
-5
10-6
10-7
10
threshold
10-1
-8
N 1/ ( )
scaling exponent
channel quality
PN (R, C)
10-2
10-3
10-4
10
1/
0
N in
cre
asin
-5
10-6
10-7
threshold
10-1
10-8
N 1/ (C
C C)
C
scaling exponent
rate
Finite-Length Scaling
lim
N :N 1/ (CR)=z
PN (R, C) = f (z)
f (z)
>0
scaling exponent
V. Privman, Finite-size scaling theory, in Finite Size Scaling and Numerical Simulation of Statistical
Systems, V. Privman, ed., World Scientific Publ., Singapore, 1990, pp. 198.
Complexity
gap to capacity
Block Codes
1
log(PN (R, C)) error exponent
N N
E(R, C) = lim
quadratic
PN
0
=0
distribution of erasures
PN
=1R
erasure fraction
=1
E[E] = N
2 ] = N (1 )
E[(E E)
E N (N , N (1 ))
N ((1 R) )
N ((1 R) )
Q
=Q
=Q
N (1 )
(1 )
(1 )
z = N ((1 R) )
n1
i=0
square binary
random matrix
of dimension n
n1
2n 2i
in n
=
(1
2
) 0.28878809508 . . .
2n
i=0
00101010001
01110010010
10101010101
01000101001
if we have k rows
less then probability
decays by roughly
(k+1
2 )
hence for random linear block codes the transition is of constant (on an absolute scale) width
log A(N, P ) = N C
N V Q1 (P ) + O(log(N ))
block length
error probability
A(N, P )
C = E[i(x, y)]
i(x, y) = log
V = V[i(x, y)]
dp(y | x)
dp(y)
convolutional codes
Convolutional Codes
Convolutional Codes
Figures ``borrowed from
affine
scaling behavior?
K constraint length
Scaling behavior
digrams such as TH, ED, etc. In the second-order approximation, digram structure is introduced. After a
letter is chosen, the next one is chosen in accordance with the frequencies with which the various letters
follow the first one. This requires a table of digram frequencies pi j . In the third-order approximation,
trigram structure is introduced. Each letter is chosen with probabilities which depend on the preceding two
letters.
3. T HE S ERIES OF A PPROXIMATIONS TO E NGLISH
To give a visual idea of how this series of processes approaches a language, typical sequences in the approximations to English have been constructed and are given below. In all cases we have assumed a 27-symbol
alphabet, the 26 letters and a space.
1. Zero-order approximation (symbols independent and equiprobable).
XFOML RXKHRJFFJUJ
QPAAMKBZAACIBZLUDREVOIGRES
LDPC ZLPWCFWKCYJ FFJEYVKCQSGHYD
HJQD.
2. First-order approximation (symbols independent but with frequencies of English text).
OCRO HLI RGWR NMIELWIS EU LL NBNESEBYA TH EEI ALHENHTTPA OOBTTVA
NAH BRL.
3. Second-order approximation (digram structure as in English).
ON IE ANTSOUTINYS ARE T INCTORE ST BE S DEAMY ACHIN D ILONASIVE TUCOOWE AT TEASONARE FUSO TIZIN ANDY TOBE SEACE CTISBE.
4. Third-order approximation (trigram structure as in English).
IN NO IST LAT WHEY CRATICT FROURE BIRS GROCID PONDENOME OF DEMONSTURES OF THE REPTAGIN IS REGOACTIONA OF CRE.
5. First-order word approximation. Rather than continue with tetragram, , n-gram structure it is easier
and better to jump at this point to word units. Here words are chosen independently but with their
appropriate frequencies.
REPRESENTING AND SPEEDILY IS AN GOOD APT OR COME CAN DIFFERENT NATURAL HERE HE THE A IN CAME THE TO OF TO EXPERT GRAY COME TO FURNISHES
THE LINE MESSAGE HAD BE THESE.
6. Second-order word approximation. The word transition probabilities are correct but no further structure is included.
THE HEAD AND IN FRONTAL ATTACK ON AN ENGLISH WRITER THAT THE CHARACTER OF THIS POINT IS THEREFORE ANOTHER METHOD FOR THE LETTERS THAT
THE TIME OF WHO EVER TOLD THE PROBLEM FOR AN UNEXPECTED.
The resemblance to ordinary English text increases quite noticeably at each of the above steps. Note that
these samples have reasonably good structure out to about twice the range that is taken into account in their
construction. Thus in (3) the statistical process insures reasonable text for two-letter sequences, but fourletter sequences from the sample can usually be fitted into good sentences. In (6) sequences of four or more
words can easily be placed in sentences without unusual or strained constructions. The particular sequence
of ten words attack on an English writer that the character of this is not at all unreasonable. It appears then
that a sufficiently complex stochastic process will give a satisfactory representation of a discrete source.
The first two samples were constructed by the use of a book of random numbers in conjunction with
(for example 2) a table of letter frequencies. This method might have been continued for (3), (4) and (5),
since digram, trigram and word frequency tables are available, but a simpler equivalent method was used.
LDPC Ensemble
n
o
d
e
s
x1+x4+x8 = 0
p
e
r
m
u
t
a
t
i
o
n
c
h
e
c
k
n
o
d
e
s
Hx=0
sparse
v
a
r
i
a
b
l
e
(1-(1-x)5)3
34
(1-(1-x)5)3
MAP versus BP
35
graph
channel parameter
# of iterations
Expansion
(dl, dr)-regular cannot
have expansion beyond
(dl-1)/dl
|C|
at most size
dl |V|
|V|
13
))
z=
(x) =
1
5
x + x3
6
6
N (BP N 2/3 )
(x) = x5
R = 3/7
PN
10-1
10-2
= 0.5791
/ = 0.6887
10-3
10-4
10-5
0.3
0.35
0.4
0.45
0.5
BP = 0.4828
Optimization
10-1
10 11 12 13
10
40.58 %
0.0
10-2
rate/capacity
1.0
10 12 14 16 18 20 22 24 26
10-4
10-5
10-6
0.2
0.25
0.3
0.35
0.4
0.45
0.5
0.55
0.6
0.65
0.7
0.75
(3, 6) BSC
(3, 4) BAWGNC
N :N 1/ (CR)=z
PN (R, C) = f (z)
1/
N1/
N
(C R) = z
N = (z/)
=2
fixes error
additive gap
Polar Codes
patterns
Reed-Muller Codes
Definition of Channels
Polar Codes
W -- BMS channel
Definition of Channels
Channel Polarization
0
0
bad channels
Definition of Channels
Successive Decoding
Successive Decoding
threshold
Channel Polarization
known
noise
X2
BEC()
BEC()
U1
U2
+
X1
1-(1-)2
much worse
X2
Y1 Y2
2 much better
X2
parity-check node
X1+U1=X1
Y2 Y1
repetition code
Definition of Channels
How Do Channels Polarize?
0.5
0.8789
0.43750.8086 0.25
0.3164
0.5625 0.6836
0.1914
0.1211
0.0625
0.0039
0.5
MAC channel:
E. Sasoglu, E. Telatar and Edmund Yeh, Polar codes for the two-user multiple-access channel, http://
arxiv.org/pdf/1006.4255
E. Abbe and E. Telatar, Polar Codes for the m-User MAC and Matroids, http://arxiv.org/pdf/
1002.0777
Compound channel:
S. H. Hassani, S. B. Korada, and R. Urbanke, The Compound Capacity of Polar Codes, http://
arxiv.org/pdf/0907.3291
Wire-tap channel and security:
H. Mahdavifar and A. Vardy, Achieving the Secrecy Capacity of Wiretap Channels Using Polar Codes,
http://arxiv.org/pdf/1007.3568
E. Hof and S. Shamai, Secrecy-Achieving Polar-Coding for Binary-Input Memoryless Symmetric WireTap Channels, http://arxiv.org/pdf/1005.2759
Mattias Andersson, Vishwambhar Rathi, Ragnar Thobaben, Joerg Kliewer, Mikael Skoglund, Nested
Polar Codes for Wiretap and Relay Channels, http://arxiv.org/pdf/1006.3573
O. O. Koyluoglu and H. El Gamal, Polar Coding for Secure Transmission and Key Agreement, http://
arxiv.org/pdf/1003.1422
Constructions:
R. Mori and T. Tanaka, Performance and Construction of Polar Codes on Symmetric Binary-Input
Memoryless Channels, http://arxiv.org/pdf/0901.2207
M. Bakshi, S. Jaggi, and M. Effros, Concatenated Polar Codes, http://arxiv.org/pdf/1001.2545
Scaling:
S. H. Hassani and R. Urbanke, On the scaling of Polar codes: I. The behavior of polarized channels,
http://arxiv.org/pdf/1001.2766
T. Tanaka and R. Mori, Refined rate of channel polarization, http://arxiv.org/pdf/1001.2067
S. H. Hassani, K. Alishahi and R. Urbanke, On the scaling of Polar Codes: II. The behavior of unpolarized channels, http://arxiv.org/pdf/1002.3187
Z ,
1 (1 Z)2 ,
wp
wp
1
2,
1
2.
Z 2,
2Z Z 2 ,
wp 12 ,
wp 12 .
Y = log(Z)
2Y,
Y 1,
wp 12 ,
wp 12 .
X = log(Y )
X + 1,
X,
wp 12 ,
wp 12 .
X + 1,
X,
wp 12 ,
wp 12 .
m/2
=2
lim P (Zm 2
2m/2+
mQ1 (R/C)/2+o(
m)
)=R
wp 12 ,
wp 21 .
Z 2,
1 (1 Z)2 ,
Battacharyya process
z=0
z=1
1
2
= 1/2
symmetry of distribution
(general case follows in similar manner)
QN (x) =
z=0
1
1
(i)
|{i : x E(WN ) }|
N
4
z=1
z = 2x
1
2
2Q2N (x) = QN (1/2 1/4 x/2)+(121{x1/8} )QN (min x/2, 1/2 x/2)
1
Q(x) = 2
1
1
Q(1/2
scaling assumption
BEC
3.62
BAWGNC
log10 PN (R, C)
PN (R, C) Q1 (N (C R))
N 1/ (C R)
C R
lim
N :N 1/ (CR)=z
PN (R, C) = f (z)
1/
N1/
N
(C R) = z
N = (z/)
fixes error
additive gap
= O(log(1/))
Message
sparse graph codes -- best codes in ``practice; still miss
some theory; error floor region is tricky; still somewhat of an
art to construct
polar codes -- nice for theory; not (yet) ready for applications
but the field is young; how do we improve finite-length
performance
scaling behavior is the next best thing to exact analysis;
probably more meaningful characterization for practical
case than error exponent