1cm Introduction to channel coding -0 - IIT Hyderabad
Transcript of 1cm Introduction to channel coding -0 - IIT Hyderabad
Introduction to channelcodingShashank Vatedka
1 / 18
What is the maximum rate at which we canreliably communicate across a noisy channel?
2 / 18
3 / 18
3 / 18
3 / 18
3 / 18
3 / 18
Not just cellular...
§ WiFi/deep space/wireless
§ Wireline/optical
§ Storage
Very general!
4 / 18
Digital Communication system
Source signal
Sampling
A/D convertSourcecoding
Channelcoding
ModulatorD/A
Demodulator
A/D
Channeldecoder
Sourcedecoder
Reconstruction
D/A convert
Decodedsignal
Noisycommunication
medium
Effective discrete "channel"
Transmitter
Receiver
5 / 18
What is the maximum rate at which we canreliably communicate across a discrete
memoryless channel?
6 / 18
Discrete memoryless channel
ChannelEncoder Decoder
§ Mk „ iid Unif(t0,1uk)§ Memoryless channel:
pY n|Xnpyn|xnq “n
ź
i“1pY |X pyi |xiq
7 / 18
Discrete memoryless channel
ChannelEncoder Decoder
§ Mk „ iid Unif(t0,1uk)§ Memoryless channel:
pY n|Xnpyn|xnq “n
ź
i“1pY |X pyi |xiq
7 / 18
Common channels
Binary symmetric channel: BSC(p)
X “ Y “ t0,1u, and
pY |X py|xq “#
1´ p if x “ yp if x ‰ y.
8 / 18
Common channels
Binary erasure channel: BEC(p)
X “ t0,1u, Y “ t0,1,eu
pY |X py|xq “
$
’
&
’
%
p if y “ e1´ p if x “ y0 otherwise.
8 / 18
Common channels
Additive white Gaussian noise (AWGN) channel
X “ Y “ RYi “ xi ` Zi , i “ 1,2, . . . ,n
where pZ1, . . . ,Znq are iid with N p0, σ2q components.Power constraint:
}xn}2def“
nÿ
i“1x2
i ď nP
8 / 18
Common channels
Complex slow/quasi-static fading channel
X “ Y “ CYi “ hXi ` Zi ,
8 / 18
Common channels
Fast fading channel
Yi “ hiXi ` Zi ,
8 / 18
Common channels
Multiple antenna/multi-input multi-output (MIMO) channels
X “ Rts , Y “ Rtr .
Y i “ HiX i ` Z i , i “ 1, . . . ,n
8 / 18
Common channels
A simple channel with memory
Yi “ a0Xi ` a1Xi´1 ` . . .` akXi´k ` Zi
8 / 18
Common channelsInsertion/deletion channels
2 2 1 1 0 1 0 2 2 1
2 2 1 1 1 0 1 0 2 2 1
insertion
Tx
Rx
8 / 18
Channel codes
ChannelEncoder Decoder
§ Encoder: f : t0,1uk Ñ X n
§ Decoder: g : Yn Ñ t0,1uk§ Rate:
R “ kn
§ Probability of error:
Pe “ Prr pMk ‰ Mks
9 / 18
Mutual information
IpX ; Y q def“
ÿ
xPX ,yPYpXY px, yq log2
pXY px, yqpX pxqpY pyq
,
§ Mutual information is symmetric
§ Measures the information that X gives about Y , or Y givesabout X .
§ What happens if X and Y are independent?
10 / 18
Mutual information
IpX ; Y q def“
ÿ
xPX ,yPYpXY px, yq log2
pXY px, yqpX pxqpY pyq
,
§ Mutual information is symmetric
§ Measures the information that X gives about Y , or Y givesabout X .
§ What happens if X and Y are independent?
10 / 18
Channel capacity
Maximum rate R for which limnÑ8 Pe “ 0.
Theorem (Shannon)
C “ maxpX
IpX ; Y q.
11 / 18
12 / 18
A brief history of channelcoding
13 / 18
The early codes
§ BCH, Reed Solomon, Reed Muller codes (1960s)
§ Convolutional codes (1955-1967)
§ Concatenated codes (1966): deep space communication
§ Trellis coded modulation (1982?): telephone lines
14 / 18
Performance
0Costello and Forney, “Channel Coding: The Road to Channel Capacity,”Proceedings of the IEEE, 2007. Link:https://arxiv.org/pdf/cs/0611112.pdf 15 / 18
Performance
0Costello and Forney, “Channel Coding: The Road to Channel Capacity,”Proceedings of the IEEE, 2007. Link:https://arxiv.org/pdf/cs/0611112.pdf 16 / 18
Modern codes
§ Turbo codes (1993)
§ LDPC codes (Gallager 1960, rediscovered 2000s)
§ Polar codes (2009)
17 / 18
Performance
0Costello and Forney, “Channel Coding: The Road to Channel Capacity,”Proceedings of the IEEE, 2007. Link:https://arxiv.org/pdf/cs/0611112.pdf 18 / 18
Performance
0Costello and Forney, “Channel Coding: The Road to Channel Capacity,”Proceedings of the IEEE, 2007. Link:https://arxiv.org/pdf/cs/0611112.pdf 19 / 18