Lecture 18: Gaussian Channel
Lecture 18: Gaussian Channel
? Gaussian channel ? Gaussian channel capacity
Dr. Yao Xie, ECE587, Information Theory, Duke University
Mona Lisa in AWGN
Mona Lisa
Noisy Mona Lisa
100 200 300 400 500 600 700 800 900 1000 1100
200
400
600
100 200 300 400 500 600 700 800 900 1000 1100
200
400
600
Dr. Yao Xie, ECE587, Information Theory, Duke University
1
Gaussian channel
? the most important continuous alphabet channel: AWGN ? Yi = Xi + Zi, noise Zi N (0, N ), independent of Xi ? model for communication channels: satellite links, wireless phone
Zi
Xi
Dr. Yao Xie, ECE587, Information Theory, Duke University
Yi
2
Channel capacity of AWGN
? intuition: C = log number of distinguishable inputs
? if N = 0, C =
? if no power constraint on the input, C =
? to make it more meaningful, impose average power constraint: for any
codewords (x1, . . . , xn)
1 n
n
x2i
P
i=1
Dr. Yao Xie, ECE587, Information Theory, Duke University
3
Naive way of using Gaussian channel
? Binary phase-shift keying (BPSK)
? transmit 1 bit over the channel
? 1 + P, 0 - P
? Y =? P +Z
? Probability of error
Pe = 1 - ( P/N )
normal
cumulative
probability
function
(CDF):
(x)
=
x
-
1
e-
t2 2
dt
2
? convert Gaussian channel into a discrete BSC with p = Pe. Lose information in quantization
Dr. Yao Xie, ECE587, Information Theory, Duke University
4
Definition: Gaussian channel capacity
? C = max I(X; Y )
f (x):EX2P
? we can calculate from here
(
)
1
P
C = log 1 +
2
N
maximum attained when X N (0, P )
Dr. Yao Xie, ECE587, Information Theory, Duke University
5
C as maximum data rate
? we can also show this C is the supremum of rate achievable for AWGN
? definition: a rate R is achievable for Gaussian channel with power constraint P : if there exists a (2nR, n) codes with maximum probability of error n = max2i=nR1 i 0 as n .
Dr. Yao Xie, ECE587, Information Theory, Duke University
6
Sphere packing
why we may construct (2nC, n) codes with a small probability of error?
Fix one codeword ? consider any codeword of length n ? received vector N (true codeword, N ) ? with high probability, received vector contained in a sphere of radius
n(N + ) around true codeword ? assign each ball a codeword
Dr. Yao Xie, ECE587, Information Theory, Duke University
7
................
................
In order to avoid copyright disputes, this page is only a partial summary.
To fulfill the demand for quickly locating and searching documents.
It is intelligent file search solution for home and business.
Related download
- hd3ss3411 one channel differential 2 1 mux demux datasheet
- mimo ii capacity and multiplexing architectures
- chapter 3 open channel hydraulics
- 2 1 channel high efficiency digital audio system sound
- lecture 18 gaussian channel
- fhp3350 fhp3450 triple and quad voltage feedback amplifiers
- stormwater conveyance channel michigan
- united states army corps of engineers engineering manual
- lorawan channel plans
- how to bridge hdmi dvi to lvds oldi rev c
Related searches
- gaussian elimination matrix
- gaussian elimination rules
- how to do gaussian elimination
- gaussian elimination calculator
- gaussian elimination method
- gaussian elimination 2x2 matrix
- how to use gaussian elimination
- the gaussian elimination method
- solve matrix by gaussian elimination
- gaussian elimination 3x3
- gaussian elimination explained
- gaussian elimination method example