Notes
Notes
Course Notes
c
°Dr. A. Anpalagan
Fall 2008
Contents
1 Overview of a Digital Communication System 5
3 Probability Theory 9
3.1 Event Space and Probability . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9
3.2 Random Variables . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11
3.3 Statistics of Random Variables . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12
3.4 Useful Probability Distributions . . . . . . . . . . . . . . . . . . . . . . . . . . . 14
4 Random Processes 14
4.1 Notion of Random Process . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15
4.2 Stationary Random Process . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15
4.3 Statistical Averages . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16
4.4 Wide Sense Stationary: . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17
4.5 Power Spectrum Density . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17
4.6 Input/Output Relationship in LTI Systems . . . . . . . . . . . . . . . . . . . . . 18
4.7 Fourier Transform and Properties: . . . . . . . . . . . . . . . . . . . . . . . . . . 19
7 Source Coding 31
2
7.1 Information Source Models . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31
7.2 Nyquist Sampling Theorem . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32
7.3 Measure of Information . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 33
7.3.1 Mutual Information . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 33
7.3.2 Entropy . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 36
7.4 Coding for Discrete Sources . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 37
7.4.1 Fixed Length Coding . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 38
7.4.2 Variable Length Coding . . . . . . . . . . . . . . . . . . . . . . . . . . . 39
7.5 Coding for Analog Sources . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41
8 Channel Capacity 43
8.1 Channel Coding . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 43
8.2 Channel Models . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 44
8.3 Capacity of BSC and AWGN Channels . . . . . . . . . . . . . . . . . . . . . . . 46
8.4 Shannon’s Channel Capacity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 48
3
10.3 PSD of Bandpass Signals . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 66
4
1 Overview of a Digital Communication System
• Source (or sink) can be analog (e.g., audio) or digital (e.g., fax).
5
and transmit one of M = 2b possible waveforms (a symbol is transmitted).
♠ Example with b = 2.
Rate is b-fold decreased. How about average power and bandwidth ?.
Communication channel: Examples are microwave radio, fiber or cable.
Thermal noise in the channel corrupts the waveforms and symbol by symbol
detection is done to declare the transmitted symbol in digital communication.
Multiple access interference (MAI) is also a cause for impairment, for example,
wireless cellular systems and Ethernet.
System performance: A digital communication system can be characterized
by performance measure such as
1. bit rate (copper wire (kbps), optical fiber (Gbps), cellular radio (kbps))
6
2 Communication Channel Models
• Problem arises when the noise and the signal of interest occupy the same
frequency band. Unfortunately, noise occupy the entire band!
• Proper system design (i.e., waveform, modulation, receiver etc.) can min-
imize the adverse effect of the noise.
7
Linear time−
s(t) r(t)
variant filter +
c(τ ; t)
channel n(t)
where c(τ ; t) is the impulse response of the channel (filter) at time t for an
impulse applied at time t − τ .
where ak (t) represents time-varying attenuation factor for the L multipaths and
τk are the corresponding delays. Hence,
L
X
r(t) = ak (t)s(t − τk ) + n(t)
k=1
c(τ ; t)
1
Multipath
0.4 0.3
channel
0.2
0 to t 0 τ1 τ 2 τ 3 t
8
3 Probability Theory
9
1, 2, ...n are mutually exclusive, then
n
X
P (Ai , Bj ) = P (Bj )
i=1
♠ Example - BSC:
Assuming equal probability for input bits,
Pe = P (X = 0|Y = 1)
10
1−p
0 0
p
Input (X) Output (Y)
p
1 1
1−p
P (Y = 1|X = 0)P (X = 0)
Pe = =p
P (Y = 1|X = 0)P (X = 0) + P (Y = 1|X = 1)P (X = 1)
E. Statistical Independence: Occurrence of A does not depend on the
occurrence of B. Hence, P (A|B) = P (A) ⇒ P (A, B) = P (A)P (B)
X(s)
Real line
Sample space
B. Distributions:
11
• Cumulative Density Function (CDF): F (x) = P (X ≤ x), x ∈ <
dF (x)
• Probability Density Function (PDF): p(x) = dx , x ∈<
C. Multiple RVs:
Z x1 Z x2
F (x1 , x2 ) = P (X1 ≤ x1 , X2 ≤ x2 ) = p(u1 , u2 )du2 du1
−∞ −∞
12
R∞
• If n = 1 and Y = X, then mean of X, E(X) ≡ mx = −∞ xp(x)dx
R∞
• If n = 2 and Y = X − mx , then variance of X, σx2 = −∞ (x −
mx )2 p(x)dx
B. Joint Moments:
• Correlation:
Z ∞ Z ∞
E(Xi Xj ) = xi xj p(xi , xj )dxi dxj
−∞ −∞
• Covariance:
dn ψX (jw)
E(X n ) = (−j)n |w=0
dwn
13
They are useful in determining PDF of a sum of statistically independent
P
RVs. Let Y = ni=1 Xi , then it can be shown that,
n
Y
ψY (jw) = ψXi (jw)
i=1
be written as, µ ¶
x − mx
F (x) = 1 − erfc √
2σ
R∞ 2
where erfc(x) = √2 e−t dt. Another function often used in Q(x) =
π x
1
2 erfc ( √x2 ).
4 Random Processes
14
• thermal noise voltages at radio receivers
X(t)
θ= 0
x x
x x
X1(t) X2(t)
θ = π/2
t
x
x X4(t)
X3(t)
θ= π x
x
θ =3/2π t = t t = t2
1
Let X(t) be a random process. Then, X(ti ), (i = 1, 2, ...., n) are random vari-
ables and can be characterized by their joint PDF, p(xt1 , ..., xtn ).
15
If
p(xt1 , ..., xtn ) = p(xt1 +t , ..., xtn +t ), ∀t, n
then X(t) is said to be strict sense stationary (SSS). That is, the process is
statistically indifferent with respect to any time shift.
Ensemble average: Let X(t) be a random process and X(ti ) ≡ Xti . The nth
moment of the RV Xti is:
Z ∞
E(Xtni ) = xnti p(xti )dxti
−∞
If the process is SSS, PDF as well as the moment (hence, mXti )is indepen-
dent of time.
φ(0) = E(Xt2 )
16
Autocovariance:
Wide sense stationary (WSS) is less stringent than strict-sense stationary (SSS)
in that mean and autocorrelation of the process are constant and de-
pendent on time difference respectively, but that is non-stationary process
(i.e., statistics such as PDF may not be stationary).
17
• Fourier transform of the (deterministic) signal and Fourier transform of the
autocorrelation of the (random) signal are respectively used in frequency-
domain analysis.
Z ∞
φ(τ ) = Ψ(f )ej2πf τ df
−∞
R∞
The average power of the random signal is, φ(0) = −∞ Ψ(f )df = E(|Xt2 |).
Whether random process is real or complex, Ψ(f ) is real, that is, it contains
real information about the actual frequency contents of the (base-band or pass-
band) signal.
Let x(t) be the sample function of the stationary random process X(t) and
h(t) be the impulse response of a LTI filter. Note that X(t) may be a complex
process in general. The output y(t) is a sample function of a process Y (t).
Hence,
Z ∞
y(t) = h(τ )x(t − τ )dτ
−∞
18
where H(0) is the frequency response of the system at f = 0. Therefore, the
mean of the output process is constant.
The autocorrelation function of the output signal,
1
φyy (t1 , t2 ) ≡ E(Yt1 , Yt∗2 )
2
Z Z
1 ∞ ∞
φyy (t1 , t2 ) = h(β)h∗ (α)E[X(t1 − β)X ∗ (t2 − α)]dαdβ
2 −∞ −∞
Since X(t) is stationary so is Y (t). Therefore,
Z ∞Z ∞
φyy (τ ) = h(β)h∗ (α)φxx (τ + α − β)dαdβ
−∞ −∞
It can be shown that the power spectrum density of the output is given by
19
5 Complex Signals and Systems
• Many digital signals are transmitted over a channel using some kind of
carrier modulation (e.g., DSB). Modulation is to generate bandpass (or
passband) signals from baseband signals.
|S(f)|
f
−fc 0 +f c
Note that
(b) Re(ζ) = 12 (ζ + ζ ∗ )
20
(d) cos2 (θ) − sin2 (θ) = cos(2θ) sin(2θ) = 2 sinθcosθ
Let s(t) be a real-valued bandpass signal and fc be the carrier frequency. Also
let s+ (t) = s(t) + jŝ(t), where ŝ(t) is the Hilbert transform of s(t). Therefore,
S+ (f ) = 2u(f )S(f )
Hilbert transformer (or filter) is a 90o phase-shifter for all frequencies in the
input signal. It is given as,
−j, f > 0,
1
h(t) = and H(f ) =
πt
+j, f < 0.
where x(t) and y(t) are lowpass real signals. Note that s(t) is bandpass real signal
and sl (t) is lowpass equivalent of bandpass signal s(t).
21
s l (t)
s(t) + j s(t)
fc
22
1
S(f ) = [Sl (f − fc ) + Sl∗ (−f − fc )]
2
R +∞
Note that Sl∗ (f ) = −∞ s∗l (t)ej2πf t dt.
where the equivalent lowpass channel (or system or filter), hl (t), is in general,
complex-valued.
Let s(t) be bandpass (input) signal and sl (t) is the equivalent lowpass signal.
Also, let h(t) be bandpass (system) response and hl (t) be equivalent lowpass
response.
23
s(t) r(t)
h(t)
sl(t) rl (t)
hl (t)
We have just seen how s(t) and sl (t) (or h(t) and hl (t) are related in time
and frequency domain. The output of the filter is given by
Also ,
R(f ) = S(f )H(f ) and Rl (f ) = Sl (f )Hl (f )
1
R(f ) = [Rl (f − fc ) + Rl∗ (−f − fc )]
2
Note: The above equivalent characterization of signals and systems are valid
for narrowband signals and systems only. Otherwise, the energy in both sig-
nals/systems will not be equal.
24
Let n(t) be a sample function of a narrowband WSS process with zero mean
and PSD of Φnn (f ).
Autocorrelation
= x(t)cos2πfc t − y(t)sin2πfc t
= Re[z(t)ej2πfc t ],
p
where a(t) = x2 (t) + y 2 (t). Since n(t) is stationary,
Note that z(t) is low pass equivalent of n(t). It can be shown that
Power Spectrum
1 1
φzz (τ ) = E[z ∗ (t)z(t + τ )] = [φxx (τ ) + φyy (τ ) − jφxy (τ ) + jφyx (τ )]
2 2
Hence,
φnn (τ ) = Re[φzz (τ )ej2πfc τ ]
25
That is, autocorrelation function of the bandpass random process is uniquely
determined from the autocorrelation of the equivalent lowpass process.
The power spectrum density is given by,
Z ∞
Φnn (f ) = {Re[φzz (τ )ej2πfc τ ]}e−j2πf τ dτ
−∞
1
Φnn (f ) = [Φzz (f − fc ) + Φzz (−f − fc )]
2
Note that the autocorrelation function, φzz (τ ), is an even function.
φnn(f)
No
2
B B
−f c +f c f
• White noise has a constant power spectrum density over the entire fre-
quency range, i.e., wideband noise signal. Hence, it cannot be expressed
in terms of quadrature components.
• Now that we have a bandpass signal, we can use any of the above three
equivalent lowpass random process representation to analyze it.
26
6 Vector and Signal Space
k v1 + v2 k2 =k v1 k2 + k v2 k2 +2v1 .v2
If v1 .v2 = 0, v1 and v2 are orthogonal. Example is (1, 0, 0), (0, 2, 0), (0, 0, 1) in
3-D vector space.
Norm (or Length) of v is denoted by k v k and defined by
v
u n
uX
k v k= (v.v)1/2 = t vi2
i=1
Orthonormal: If two vectors are orthogonal and each has a unity norm, they
are orthonormal.
27
Gram-Schmidt Procedure for Orthonormalization of Vectors
v1
1. Select any vector (v1 ) and normalize it, u1 = kv1 k
2. Select the next vector (v2 ), subtract the projection of v2 onto u1 . Thus,
0
0 0 u2
u2 = v2 − (v2 .u1 )u1 and then normalize u2 . Hence, u2 = 0
ku2 k
Let x1 (t) and x2 (t) be (generally complex) signals. The inner product,
Z ∞
(x1 (t), x2 (t)) ≡ x1 (t)x∗2 (t)dt
−∞
28
Also let fn (t) be a set of K orthonormal functions. That is,
Z ∞
0, m 6= n,
fn (t)fm (t)dt =
−∞
1, m = n.
PK
s(t) is approximated by s̄(t) = k=1 sk fk (t), where {sk , 1 ≤ k ≤ K} are
coefficients in the approximation. Hence, the error in the approximation is
s̄(t) is obtained by projecting the signal s(t) onto K-dimensional signal space.
When ²e = 0, error signal has no energy and the set fn (t), (n = 1, ..., K) is a
complete set. Then,
Z ∞ K
X
2
²s = |s(t)| dt = s2k
−∞ k=1
29
Gram-Schmidt Procedure for Orthonormalization of Signals
Let {si (t), i = 1, ..., M } be a set of finite energy signal waveforms. In general,
the orthonormalization of the kth function gives us,
0
fk (t)
fk (t) = √
²k
where
k−1
X
0
fk (t) = sk (t) − cik fi (t)
i=1
and
Z ∞
cik = sk (t)fi (t)dt, i = 1, ..., k − 1
−∞
Once orthonormal signal set is identified, then any signal in the signal space
can be expressed in terms of a linear combination of the orthonormal signals.
30
√
3. c13 = 2 and c23 = 0, hence,
−1, (2 ≤ t ≤ 3),
0 √
f3 (t) = s3 (t) − 2f1 (t) =
0, otherwise.
0
f3 (t) = f3 (t)
0
4. f4 (t) = 0, eliminate this signal from the orthogonal set.
In signal form,
√
s4 (t) = − 2f1 (t) + f3 (t)
7 Source Coding
• Source encoder generates efficient digital sequence (i.e., binary: 0’s and
1’s) for transmission/storage.
They are known as Discrete Memoryless Sources (DMS) for which joint
PDFs can be expressed as the product of (individual) marginal PDFs.
Let X(t) be a stationary random process with power spectral density ΦXX (f )
and ΦXX (f ) = 0 for |f | ≥ W (i.e., band-limited). By sampling theorem,
X∞ µ ¶ n
n sin [2πW (t − 2W )]
X(t) = X n
n=−∞
2W [2πW (t − 2W )]
where discrete samples X(m), m = 1, 2, ... are taken at the Nyquist sampling
rate of 2W samples/s.
¶ µ
n
♠ Bandwidth, sinc(x), construction of X(t) from X .
2W
32
It is important to note the following:
P (X = xi |Y = yj ) ≡ P (xi |yj ).
What if X and Y are independent (e.g., raining in Toronto and NASDAQ index
going up or down) or dependent (e.g., wearing winter jackets in a snowy day)?
If X and Y are independent, then
33
− log P (xi ) amount of information. Note that 0 ≤ P (xi ) ≤ 1. Also it can be
seen that, I(X; Y ) = I(Y ; X).
1
= −log2 = 1 bit (note the base of 2)
2
Let us now consider a block of K binary digits that are statistically in-
dependent (DMS model). There are M = 2K possible K-bit blocks, each
with equal occurrence probability of 2−K . Note that
34
P(X=0) = P(X=1) = 0.5
1−po
0 0
p
1
Input (X) Output (Y)
po
1 1
1−p1
P (Y = 0|X = 0)
I(X = 0; Y = 0) = I(Y = 0; X = 0) = log2
P (Y = 0)
Also,
1
⇒ P (Y = 0) = (1 − po + p1 )
2
· ¸
2(1 − po )
Hence, I(0; 0) = log2
1 − po + p1
If po = p1 = 0 (noiseless channel), then I(0, 0) = 1.
If po = p1 = 21 , then I(0, 0) = 0 (useless channel).
35
7.3.2 Entropy
36
Figure 3.1−1
Binary entropy function.
37
• Coding efficiency : compare the average number of binary digits per output
symbol with source entropy, H(X). It is defined as H(X)/R̄ where R̄ is
the average number of bits required to represent a symbol.
H(X) ≤ log2 L.
38
Shannon Source Coding Theorem:
Let X be an alphabet set (consisting of output symbols) with entropy H(X).
Block of J symbols are encoded into code words of length N digits from a binary
alphabet. The symbols are not uniquely encoded and therefore decoding error
occurs.
The probability of a block decoding failure (Pb ) can be made very small if
N
R̂ = ≥ H(X)
J
When source symbols are not equally probable, use variable-length code words.
Idea is to assign short code words for high probable source symbols (known as
entropy coding).
39
♠ Example: Variable Length Coding
Code I has problem in decoding. What is the decoded sequence if the received
sequence is 001001 ? a2 a1 a2 a1 or a2 a4 a3 ?
The prefix condition requires that no code word is a prefix of any other code
word. Code III is also not uniquely decodable.
(a) Variable-length code words, (b) prefix condition will be satisfied, (c) aver-
age number of binary digits to represent the source symbols is minimum, (d)
uniquely and instantaneously decodable, (e) symbol probabilities are needed
and (f) code words may not be unique but all achieve the same efficiency.
40
Symbol Probability Code word
x1 0.35 00
x2 0.30 01
x3 0.20 10
x4 0.10 110
x5 0.04 1110
x6 0.005 11110
x7 0.005 11111
Figure 3.3−4
An example of variable−length source
encoding for a DMS.
41
• Samples are then quantized in amplitude and encoded with finite number of
bits. Output is grouped into L levels with R bits/sample where R = log 2 L.
Example, in PCM (used in speech coding), L = 4096, R = 12.
Rate Distortion Function: Let xk and x̂k be actual and quantized values.
Squared-error distortion is: d(xk , x̂k ) = (xk − x̂k )2 . With n samples,
n
1X
d(Xn , X̂n ) = d(xk , x̂k )
n
k=1
where σx2 is the variance of the Gaussian source output. Distortion-rate function
(for the above Gaussian source) is, Dg (R) = 2−2R σx2 . In dB, −6R + 10log 2
10 σx .
42
8 Channel Capacity
n
Code rate is denoted
µ ¶ as (k,n). The raw rate becomes k with the amount of
redundancy = n−k k %. Example: a1 a2 → a1 a2 a1 ⊕ a2 .
• Sequence of coded bits are input into the modulator which transmits one
symbol for each r bits (using one waveform from M = 2r possible wave-
forms). Orthogonal signals can make Pe smaller but with larger bandwidth
0 T 0 T
M = 4, r = 2
0 T 0 T
B=W B = 2W
W
expansion factor, Be = R, where R is the channel bit rate. Be grows ex-
ponentially with block length, r.
43
• Detector outputs 0’s and 1’s (based on soft/hard decision rule).
BIBO and BSC are examples. In general, we can have q−ary inputs and Q-ary
outputs. q = Q = 2 gives BIBO channel.
Let a channel encoder output consists of symbols selected from a finite alphabet
set, X = {x0 , ..., xq−1 } and Q = ∞ (un-quantized input to the channel decoder).
The composite channel can be characterized by the conditional PDFs as ,
P (y|X = xk ), k = 0, 1, ..., q − 1.
Y =X +G
44
where G is a zero-mean Gaussian RV with variance σ 2 . For a given X = xk , Y
is Gaussian with mean xk and variance σ 2 . Therefore,
1 2 2
P (y|X = xk ) = √ e−(y−xk ) /2σ
2πσ
If we transmit n symbols through the memoryless channel, we can say,
n
Y
P (y1 , ..., yn |X = x1 , ..., Xn = xn ) = P (yi |Xi = xi )
i=1
Let us suppose that the inputs to the channel and outputs from the channel
are continuous. Let x(t) be a band-limited (to W ) input process to a AWGN
channel (with PSD of 21 N0 ). Let n(t) be the sample function of the noise process.
Hence, the output process,
where {xi }, {ni }, {yi } are the sets of coefficients in the corresponding expansion.
The functions {fi (t)} form a complete orthonormal set over (0, T ), i.e.,
Z T
1, i = j,
fi (t)fj (t)dt =
0
0, i 6= j.
45
• Note that ni ’s are uncorrelated and hence statistically independent Gaussian
RVs. Hence, yi ’s are also statistically independent Gaussian RVs. Hence,
Z T Z T
yi = y(t)fi (t)dt = [x(t) + n(t)]fi (t)dt = xi + ni
0 0
• Now we can work with the coefficients in characterizing the channel. Hence
1 2 2
P (yi |xi ) = √ e−(yi −xi ) /2σi , i = 1, 2, ....
2πσi
• Continuous waveform channel has been reduced to equivalent discrete
channel.
Consider a discrete memoryless channel having an input alphabet X = {x0 , ..., xq−1 }
and an output alphabet Y = {y0 , ..., yQ−1 } with the set of transition probabili-
ties, P (yi |xi ).
The mutual information provided about X = xi by the occurrence of Y = yi
is, · ¸
P (yi |xi )
I(xi ; yi ) = log
P (yi )
The probability of Y = yi is,
q−1
X
P (yi ) = P (xk )P (yi |xk ).
k=0
46
Therefore, the average mutual information provided by the output Y about the
input X is given by,
q−1 X
X Q−1 · ¸
P (yi |xj )
I(X; Y ) = P (xj )P (yi |xj )log
j=0 i=0
P (yi )
C = max I(X; Y )
P (xj )
Pq−1
under the constraints that P (xj ) ≥ 0 and j=0 P (xj ) = 1. The units of C is
bits/symbol and the channel rate is C/T , if a symbol enters the channel every
T seconds.
♠ Capacity of a BSC
Given that P (0|1) = P (1|0) = p and P (0) = P (1) = 0.5. The capacity of a
BSC is
C = plog 2p + (1 − p)log 2(1 − p) = 1 − H(p)
48
Let us use the samples (or coefficients) to determine I(X; Y ). We want to
find the mutual information between XN = [x1 ...xN ] and YN = [y1 ...yN ], where
N = 2W T and yi = xi + ni .
The capacity of the band-limited AWGN channel with a band-limited and
average power-limited input is shown to be,
µ ¶
Pav
C = W log 2 1 + = W log 2 (1 + SNR)
W N0
where Pav is the average power of x(t). The relationship between SNR and C
Figure 7.1−7
Normalized channel capacity as a function
of SNR for band−limited AWGN channel.
49
Figure 7.1−8
Channel capacity as a function
of bandwidth with a fixed
transmitted average power.
Noisy Channel Coding Theorem: There exist channel codes that can
achieve reliable communication with Pe → 0, if the transmission rate R < C.
Modulator with memory operates on current and previous digits (or waveforms).
The waveforms may differ in:
50
Phase: BPSK, QPSK; and QAM (PAM-PSK)
• The waveform g(t) is a real-valued signal pulse whose shape influences the
spectrum of the transmitted signal (flexible in spectrum shaping)
• If input rate to the modulator is R bits/s, then the symbol rate is R/k
symbols/s
51
The energy of signal, sm (t) is
Z T Z
A2m T 2 A2
εm = 2
sm (t)dt = g (t)dt = m εg
0 2 0 2
(e) p
dmin = d 2εg
The signal-space diagrams are shown in the figure for different M. They are
also known as constellations.
Antipodal Signal: M = 2, s1 (t) = −s2 (t), ε1 = ε2
52
9.1.2 Phase Shift Keying (PSK)
• θm ’s are possible phases of the carrier that convey the transmitted infor-
mation.
εg
• These PSK signals have equal energy of 2 for all the transmitted waveforms
where εg is the signal energy of the pulse g(t).
53
• These signals are two-dimensional and hence can be represented as a linear
combination of two orthonormal signal waveforms, f1 (t) and f2 (t).
Therefore,
sm (t) = sm1 f1 (t) + sm2 f2 (t)
where, s
2
f1 (t) = g(t) cos 2πfc t
εg
s
2
f2 (t) = − g(t) sin 2πfc t
εg
The 2D vector is given by,
·r r ¸
εg 2π(m − 1) εg 2π(m − 1)
sm = cos sin
2 M 2 M
The minimum Euclidean distance between adjacent phases of the signals is,
s · µ ¶¸
(e) 2π
dmin = εg 1 − cos
M
Signal space diagrams are shown in the figure. M = 2 (BPSK) corresponds
to binary PAM (also antipodal signals). The preferred mapping of symbols to
waveforms is using the Gray coding.
It can be viewed as the combination of ASK and PSK. In QAM, two sym-
bols are efficiently modulated using two quadrature carriers, cos (2πfc t) and
sin (2πfc t). Hence,
54
♠Constellation and Waveforms forM = 8
π
Note the 4 shift to increase the distance between signal points.
55
frequency separation between orthogonal signals. M (orthogonal) FSK signals
can be represented as vectors
√
s1 = [ ε 0 0... 0 0]
√
sM = [0 0 0... 0 ε]
(e)
√
dkm = 2ε, ∀m, k
1
• Since 4f = 2T , the smaller pulse (i.e., high speed communication) re-
quires larger frequency separation, and hence the FSK signal bandwidth
is increased.
56
• ♠Constellation for M = 4, In = ±1, ±3
4f 4f 34f 34f
4f1 = , 4f2 = − , 4f3 = , 4f4 = −
2 2 2 2
Let us consider a modulation scheme where ”0” and ”1” are represented by a
rectangular pulse of −A and A respectively.
NRZ
NRZI
data 1 0 1 1 0 0 0 1
57
• Transition occurs only when a ”1” is transmitted. This is differential en-
coding and is described mathematically as:
bk = ak ⊕ bk−1
where ak is the binary information sequence into the encoder and bk is the
output sequence of the encoder.
• The output of the encoder (b0k s) is mapped into one of two waveforms.
1
fn = 4f In , In = ±1, ±3, ..., ±M − 1.
2
58
1
where g(t) is a rectangle pulse of duration T with amplitude 2T . {In } is the
sequence of amplitudes obtained by mapping k-bit blocks of binary digits from
symbols {an } into different levels, M = 2k . Therefore, the carrier-modulated
signal with equal-energy of ε,
r
2ε
s(t) = cos [2πfc t + φ(t; I)],
T
where
Z t Z t X
φ(t; I) = 4πT fd d(τ )dτ = 4πT fd [ In g(τ − nT )]dτ
−∞ −∞ n
where,
h = 2fd T
n−1
X
θn = πh Ik
k=−∞
0 t < 0,
q(t) = t/2T 0 ≤ t ≤ T,
1/2 t > T.
• Note the accumulation of memory in θn .
• h is called modulation index.
• In general,
Z t
q(t) = g(τ )dτ.
0
59
Pulse shapes for CPFSK signals.
Phase Trees: Phase of the PSK signals are described using phase-evolution
tree.
Let In = ±1 be binary symbols and the corresponding phase tree is shown in
the figure. Assume θ0 = 0, g(t) is a rectangle pulse with duration T.
Smoother phase trajectories can be obtained using pulses that do not contain
discontinuities.
and µ ¶
t − nT
φ(t; I) = θn + πIn , nT ≤ t ≤ (n + 1)T
2T
Therefore, · µ ¶¸
t − nT
s(t) = A cos 2πfc t + θn + πIn
2T
60
Phase trajectory for binary CPFSK signal with
rectangular pulse.
· µ ¶ ¸
1 π
s(t) = A cos 2π fc + In t − nIn + θn
4T 2
The binary (In = ±1) CPFSK signals can be expressed as a sinusoid having
one of two possible frequencies in the interval nT ≤ t ≤ (n + 1)T . Let
1 1
f1 = fc − f2 = fc +
4T 4T
1 1
Minimum frequency 4f = 2T . Hence, binary CPFSK with h = 2 is called MSK.
Comparison of MSK and QPSK signals:
61
10 Spectra of Digital Signals
62
The Fourier transform of φ̄vv (τ ) gives average PSD of v(t) as,
1
Φvv (f ) = |G(f )|2 Φii (f )
T
where G(f ) is the Fourier transform of g(t) and Φii (f ) is the PSD of the infor-
mation sequence which is defined as,
∞
X
Φii (f ) = φii (m)e−j2πf mT .
m=−∞
The first term is continuous spectrum and second term consists of discrete
frequency components spaced 1/T apart in frequency.
What if {In } are equi-probable and symmetrically positioned in the signal
space? µi = 0.
63
Rectangular pulse and its energy density spectrum
64
Raised cosine pulse and its energy density spectrum
Let
In = bn + bn−1
Assume {bn }’s are uncorrelated random variables each having zero mean and
unit variance. Then
2, m = 0,
φii (m) = 1, m = ±1,
0, otherwise.
4
Φvv (f ) = |G(f )|2 cos 2 (πf T )
T
65
10.3 PSD of Bandpass Signals
s(t) = Re[v(t)ej2πfc t ],
where v(t) is the equivalent low-pass signal of s(t). The autocorrelation function
of s(t) is,
φss (τ ) = Re[φvv (τ )ej2πfc τ ].
1
Φss (f ) = [Φvv (f − fc ) + Φ∗vv (−f − fc )].
2
sm (t), 0 ≤ t ≤ T, m = 1, 2, ..., M.
66
Let the dimension of the M-ary signal set be N , i.e., the orthonormal functions
fn (t), n = 1, 2, ..., N, span the signal space.
Received signal in the symbol interval T is,
where n(t) is the sample function of the AWGN process with PSD of Φnn (f ) =
N0
2 .
r(t)
S1 (t)
lT (l+1)T
S2 (t)
Binary anti−podal signals:
S2 (t) = − S1 (t)
R
In the figure, intuitively, if r(t)dt > 0, then we can declare that the trans-
mitted pulse is s1 (t)
• Observe r(t) over T and then decide what was transmitted to the best of
knowledge. That is, design a receiver that is optimum in the sense that it
minimizes the probability of making an error.
67
11.1 Demodulator
11.1.1 Correlator
• The signal sm (t) and the noise n(t) are expanded into a series of linearly
weighted orthonormal functions.
• Note that {fn (t)} does not span the noise space.
• The received signal r(t) is passed through a parallel bank of N cross cor-
relators that compute the projection of r(t) onto N orthonormal functions
{fn (t)} over T , as shown in the figure.
68
⇒ rk = smk + nk , k = 1, 2, ...N,
where
Z (l+1)T Z (l+1)T
smk = sm (t)fk (t)dt and nk = n(t)fk (t)dt.
lT lT
• The components {nk } are random variables that arise due to the presence
of additive noise.
0
It can be shown that n (t) and rk are statistically independent and consequently,
0
n (t) is irrelevant to the decision as to what was transmitted. Therefore, the
decision can be based on the correlator output (signal and noise) components,
rk = smk + nk .
N noise components {nk } are zero-mean uncorrelated Gaussian RVs with equal
variance of σn2 = 21 N0 . Therefore, the correlator outputs {rk } conditioned on the
mth signal being transmitted are Gaussian RVs with mean and variance given
by,
E(rk ) = E(smk + nk ) = smk
1
σr2 = σn2 = N0
2
69
The correlator outputs (rk ) are statistically independent Gaussian RVs. There-
fore,
N
Y
P (r|sm ) = P (rk |smk ), m = 1, 2, ..., M,
k=1
where, · ¸
1 (rk − smk )2
P (rk |smk ) = √ exp −
πN0 N0
Joint conditional PDFs,
· X N ¸
1 (rk − smk )2
P (r|sm ) = exp −
(πN0 )N/2 k=1
N0
Sufficient Statistics
All the relevant information to make the decision as to which of sm (t) was
transmitted is contained in the correlator outputs {rk } where rk = smk + nk .
♠ M-ary baseband PAM signal
M -ary baseband PAM with basic pulse shape g(t). Am = (2m − 1 − M )d.
PAM signal has dimension N = 1. Let
A, 0 ≤ t ≤ T,
g(t) =
0, otherwise.
Hence, εg = A2 T.
√
1 1/ T , 0 ≤ t ≤ T,
f (t) = √ g(t) =
A T
0, otherwise.
70
⇒ r = sm + n.
y(t) = s(t)*h(t)
s(t) h(t) = s(T−t)
A A y(T)
t t t
0 T 0 T 0 T 2T
(a) (b) (c)
Let
hk (t) = fk (T − t), 0 ≤ t ≤ T,
be the impulse response of the N filters where {fk (t)} are the N basis functions
and hk (t) = 0 outside the symbol interval T . The output of the filter is,
Z t Z t
yk (t) = r(τ )hk (t − τ )dτ = r(τ )fk (T − t + τ )dτ, k = 1, 2, ..., N.
0 0
71
If we sample at t = T,
Z T
yk (T ) = r(τ )fk (τ )dτ = rk
0
Therefore, the sampled outputs of the filters at t = T are the exact values
obtained from N correlators.
H(f ) = S ∗ (f )e−j2πf T
72
The output spectrum is,
Hence,
Z ∞
ys (t) = |S(f )|2 e−j2πf T ej2πf t df
−∞
By sampling at t = T , we obtain,
Z ∞ Z T
2
ys (T ) = |S(f )| df = s(t)2 dt = εs .
−∞ 0
R∞ 2
Note that Parseval’s theorem states that, the signal energy is, −∞ |s(t)| dt =
R∞ 2
−∞ |S(f )| df.
Hence, the output power (of the signal sampled at t = T ) is Ps = ε2s . The
noise PSD at the output of the matched filter is
N0
Φo (f ) = |H(f )|2
2
Ps ε2 2εs
SNRo = = 1 s =
Pn 2 εs N0
N0
11.2 Detector
73
• Design a detector that makes a decision on sm (t) in each symbol interval
T based on observing r.
74
N
N 1 X
⇒ ln P(r|sm ) = − ln (πN0 ) − (rk − smk )2
2 N0
k=1
Therefore, the maximum of P (r|sm ) over sm is equivalent to finding the signal
P
that minimizes the Euclidean distance, D(r, sm ) = N 2
k=1 (rk − smk ) .
O O
s 1 s 3
r
X
s 2 s 4
O O
N
X N
X N
X
D(r, sm ) = rk2 −2 rk smk + s2mk
k=1 k=1 k=1
⇒k r k2 −2r.sm + k sm k2
ksm k2
Therefore, let us minimize −2r.sm + k sm k2 or maximize r.sm − 2
and selects the signal corresponding to the smallest and the largest metric
respectively.
Note that the above development of optimum detector (ML and MD) as-
sumes equiprobable symbols. Otherwise, the optimum MAP detector bases its
75
decision on P (sm |r), m = 1, 2, ..., M, or equivalently P (r|sm )P (sm ) since P (r)
is irrelevant in the optimization.
The received signal vector for binary PAM is, (note that N = 1),
√
r = ± εb + n.
N0
We have seen that n is a zero-mean Gaussian RV with variance σn2 = 2 .
76
Let β() ≡ P (r|sm )P (sm ). Based on MAP criterion, the decision rule is
β(r, s1 ) s1
≷ 1.
β(r, s2 ) s2
Since · ¸
√ √
β(r, s1 ) p (r + εb )2 − (r − εb )2
= exp
β(r, s2 ) 1 − p 2σn2
µ ¶
N 0 1 − p
⇒ r ≷ss12 √ ln
4 εb p
√
• Since C(r, s1 ) = r εb , the optimum detector
µ ¶compares the correlation
metric with the threshold χ(p, N0 ) = N40 ln 1−pp and decides as
• The decision is influenced by noise PSD, pulse energy and symbol proba-
bility when symbols are not equiprobable.
1
• When p = 2, χ(.) = 0 and in this case, minimum Euclidean distance
criterion can be used.
π
♠ Consider 4 −shifted QPSK and its constellations. What are the decision
regions assuming equiprobable symbols?
77
or
S2 X X S1
or
S3 X S4
X
π
Figure 3: 4
−shifted QPSK Constellations and Decision Regions for Equiprobable Symbols
78
Let us consider an example with NRZI signal whose memory is shown in the
trellis for binary PAM signaling scheme.
The output of the matched-filter (or correlation modulator) in the lth symbol
interval is
√
rl = ± εb + nl
Let us now consider a sequence of outputs {r1 , r2 , ..., rL }. The noise sequence
{n1 , n2 , ..., nL } is white and Gaussian. Therefore, for any transmitted sequence
s(m) , L marginal PDFs are
L
Y
(m) (m)
P (r1 , r2 , ..., rL |s )= P (rl |sl )
l=1
· X L (m) ¸
1 (rl − sl )2
= exp −
(2πσn2 )L/2 2σn2
l=1
√
with sl = ± εb .
Therefore, with the received sequence {r1 , ..., rL }, the detector determines
(m) (m)
the sequence s(m) = {s1 , ..., sL } that maximizes the above conditional PDF
(maximum-likelihood sequence detector).
Equivalently, the above MLSD selects the sequence s(m) that minimizes the
Euclidean metric,
L
X
(m) (m)
D(r, s )= (rl − sl )2 .
l=1
79
Viterbi algorithm - a trellis search algorithm for performing maximum likelihood
sequence detection
At steady state (and thereafter), at t = nT, n ≥ 2, there are two paths entering
and leaving each node.
For example, at t = 2T , two paths entering node s0 corresponding to the
information bits (0, 0) and (1, 1). The Euclidean distances are
√ √
D0 (0, 0) = (r1 + εb )2 + (r2 + εb )2
√ √
D0 (1, 1) = (r1 − εb )2 + (r2 + εb )2
80
12.1.1 Binary PAM
Let us consider equiprobable symbols with signals s1 (t) and s2 (t). We have seen
in AWGN channel,
1 √ 2
P (r|s1 ) = √ e−(r− εg ) /N0
πN0
Given that s1 (t) was transmitted,
Z 0 Z 0
1 √ 2
P (e|s1 ) = P (r|s1 )dr = √ e−(r− εg ) /N0 dr
−∞ πN0 −∞
Z −√2εg /N0 Z ∞ µr ¶
1 2 1 2 2εg
P (e|s1 ) = √ e−x /2 dx = √ √ e−x /2 dx = Q
2π −∞ 2π 2εg /N0 N0
R∞ 2
where Q(x) = √12π x et /2 dt.
Similarly, P (e|s2 ) can be computed. Hence, the average probability of error
for equiprobable equal energy binary PAM signals (i.e., antipodal signals) is
µr ¶
antipodal 1 1 2εg p
Pb = P (e|s1 ) + P (e|s2 ) = Q = Q( 2γb )
2 2 N0
81
Hence,
µs 2 ¶ µr ¶
orthogonal d12 εg √
Pb =Q =Q = Q( γb )
2N0 N0
The larger the argument the smaller the Pb in Q(.) function. Therefore,
orthogonal signals are 3 dB poorer than antipodal signals. That is, in order to
get the same error performance, the orthogonal signals have to increase their
energy by a factor of 2 compared to antipodal signals.
For equal-energy (εs ) orthogonal signals, the optimum detector selects the signal
with the largest cross correlation between the received vector r and each of the
M possible transmitted signal vector {sm }.
M
X
C(r,sm ) = r.sm = rk smk , m = 1, 2, ..., M.
k=1
82
Let us suppose that the signal s1 was transmitted. Therefore, the received
signal vector is,
√
r = [ εs + n1 n2 n3 ... nM ]
83
• Since M = 2k , εs = kεg .
For equiprobable orthogonal signals, all symbol errors occur with equal prob-
ability of,
PM PM
= k
M −1 2 −1
PM
♠P (sm |s1 ) = . For M = 2, PM = Pb .
M −1
¡k ¢
There are n ways in which n bits out of k, may be in error. Therefore, the
average number of bit errors per k−bit symbol is,
X k µ ¶
k PM k2k−1
Pb = n k −1
= k PM .
n=1
n 2 2 − 1
2k−1 PM
Hence, the average bit error probability is Pb = P
2k −1 M
≈ 2 , k À 1.
p
1 − M )d. dmin = d 2εg . Assuming equiprobable symbols, the average energy
is,
M
1 X d2
εav = εm = (M 2 − 1)εg
M m=1 6
84
Figure 4: Probability of bit error for orthogonal signals.
85
• The decision is made in favor of the amplitude level closer to r.
q
M −1
PM = P (|r − sm | > d εg /2)
M
µs 2 ¶
2(M − 1) d εg
PM = Q
M N0
With k = ln(M ), the symbol error can be expressed in terms of SNR per bit (εg,av /N0 ) as,
µs ¶
2(M − 1) 6 ln(M ) εg,av
PM = Q ,
M (M 2 − 1) N0
6εav
where εav = ln(M )εg,av and εg = (M 2 −1)d2
.
♠ As M is increased, the effect in symbol error probability for FSK and PSK?
86
12.1.4 M-ary PSK
Since the signal waveforms have equal energy (and assuming equiprobable
symbols), the optimum detector for AWGN channel computes the correlation
metrics, C(r, sm ). The (corrupted) received signal vector is r = [r1 r2 ]. The
phase of r is, θr = tan−1 ( rr21 ). We need to compute the PDF of θr to compute
PM .
Z +π/M
PM = 1 − pθr (Θr )dΘr
−π/M
One can compare the SNR required to achieve a specified Pe for different modu-
lation schemes. However, it is not just enough or fair. Bandwidth (and achiev-
able data rate) requirement has also to be taken into account.
87
12.2.1 Bandwidth Limited Case
For multi-phase signals (such as PSK, PAM), the bandwidth requirement is the
bandwidth of the equivalent lowpass signal, g(t).
1
Let us suppose that the bandwidth of signal g(t) is W = T (approximately).
k
Since T = R and k = ln (M ),
R
W = .
ln M
The bandwidth efficiency is measured in terms of bit/s/Hz. Hence, as M is
increased, when R is fixed, the bandwidth requirement decreases.
The bandwidth efficient method to transmit PAM signal is to use SSB modu-
1
lation with bandwidth of 2T . Hence, the efficiency is 2 × ln (M ). This is 3 dB
better than DSB.
Orthogonal Signals
88
shown in the figure.
• In the case of PAM, QAM and PSK, increasing M results in higher ratio
of R/W with the cost of increased SNR per bit (i.e., signal power).
89
• The maximum possible achievable ratio of C/W (where C = R bit/s) is
Shannon capacity limit for AWGN channel. It is an upperbound on the
bandwidth efficiency of any type of modulation (either band-limited or
power-limited case).
Analog Repeaters
• If K analog repeaters are used, the bit-error probability for binary PAM
transmitted signal is given as
µr ¶
analog 2εg
Pb ≈Q ,
KN0
where α >>
µ 0¶ and the output SNR at the last repeater can be approxi-
2ε
mated as KNg0 .
90
Regenerative Repeaters
• The only error that occurs is in the detection of the symbol and this error
may propagate up to the last repeater.
• Suppose that the binary PAM is used. Hence, the probability of (bit) error
is, µr ¶
2εg
Pb = Q
N0
2ε
Note that the output SNR of any repeater is ( N0g ).
91
13 Signal Design for Band-limited Channels
• How to design a pulse g(t) to transmit through the (AWGN) channel with
minimum distortion?
In general,
C(f ) = |C(f )|ejθ(f )
where |C(f )| and θ(f ) are amplitude and phase responses respectively.
• One can transmit a sequence of pulses every T seconds since g(t) has zeros
at ±T, ±2T, ...
92
• Compensation for non-ideal frequency response characteristics can be done
using equalizers.
• Let us suppose that g(t) has a band-limited frequency response of G(f )(=
0, |f | > W ). The signal to be transmitted is v(t) as given in (1).
93
• v(t) is transmitted over the band-limited channel whose lowpass equivalent
frequency response is C(f ).
• Suppose that r(t) is passed through a filter and sampled at a rate 1/T
samples/s.
*
v(t) c(t), C(f) r(t) h(T−t), H (f) y(t)
(band−limited channel) (matched filter) k
n(t), Sn(f)
(AWGN channel)
• In the presence of AWGN, the optimum filter is one that is matched to the
signal at the receiver. What is the signal input to the filter in a symbol
interval?
94
where x(t) and z(t) are responses of the matched filter to h(t) and n(t)
respectively.
Hence, assuming zero transmission delay through the channel, the k th sampled
output is,
X X
y(kT ) ≡ yk = In x(kT − nT ) + z(kT ) = In xk−n + zk
n n
X
yk = Ik x0 + Ik xk−n + zk . (3)
n6=k
Eye Diagram
What should x(t) (or equivalently g(t)), be such that there is no ISI. That is,
we require (in time domain),
1, k = 0,
x(t = kT ) = xk =
0, k 6= 0.
95
96
Let us assume an ideal channel, i.e., C(f ) = 1 for |f | < W. Then, by looking at
the previous block diagram in Figure 8,
Hence, from x(t) we can find X(f ) and then G(f ) and finally g(t)!
The necessary and sufficient condition on X(f ) in order for x(t) to satisfy
the above relationship is also known as Nyquist pulse shaping criterion, which
is stated as (in frequency domain),
∞
X
B(f ) = X(f + n/T ) = T
n=−∞
It can be seen from (4), X(f ) = G(f ) = 0, |f | > W, since the channel bandwidth
is W , i.e., C(f ) = 0, |f | > W. The pulse is,
Z ∞p
g(t) = |X(f )|ej2πf t df
−∞
A note of caution: There will still be distortion due to noise n(t); in order
to minimize the effect due to (AWGN) noise, we used the matched filter.
97
cannot design a system with no ISI.
1
Case 2: T = 1/2W or symbol rate, T = 2W .
When T = 1/2W, (i.e., at the Nyquist rate that is required for no frequency-
aliasing), the replicas of X(f ) are separated by only 1/T . Hence, there exists
only one X(f ) that results in B(f ) = T. That is,
T, |f | < W,
X(f ) =
0, otherwise.
98
1
Case 3: T > 1/2W or symbol rate, T < 2W .
99
µ ¶
T, 0 ≤ |f | ≤ 1−β
2T
½ · µ ¶¸¾ µ ¶
T πT 1−β 1−β 1+β
Xrc (f ) = 1 + cos |f | − , ≤ |f | ≤
2
β 2T 2T 2T
µ ¶
1+β
0 |f | > 2T
Note that x(t) satisfies the time domain constraint (i.e., x(kT ) = 0, k 6= 0,)
100
for no ISI.
1
• To realize practical filters with zero ISI, T < 2W . That is, the symbol rate
is below the Nyquist rate of 2W samples/s.
• Relax the zero ISI constraint and with controlled ISI, then the symbol rate
comparable to Nyquist rate is achievable.
Hence,
T n = 0, −1,
bn =
0 otherwise.
Then,
B(f ) = T + T e−j2πf T
For T = 1/2W ,
Note that the spectrum of x(t) decays smoothly and as a result, is realizable.
Thus, a symbol rate of 2W samples/s can be achieved.
♠ Example with spectrum shaping: Modified Duobinary Signal Pulse:
101
1, n = −1,
x(nT ) = −1, n = 1,
0, otherwise.
102
103
Assignments
104
Problem Set: 1
2. Additional Problems:
Time Domain Signal: Find the inverse Fourier transform of the spectra
shown in the figure (use properties of the Fourier transform such as
frequency-shifting).
G( ω) G( ω)
1 1
ω ω
−4 −2 2 4 −6 −2 2 6
(a) (b)
105
BSC: Over a binary communication channel, the symbol 0 is transmitted
with probability 0.4 and 1 is transmitted with probability 0.6. It
is known that P (², 0) = 10−6 and P (², 1) = 10−4 , where P (², xi ) is
the probability of detecting the error given that xi is transmitted.
Determine P (²), the error probability of the channel.
106
Problem Set: 2
1. Encoding of Information: From the text book (Proakis), 3.1, 3.2, 3.7;
and
A source has an alphabet {a1 , a2 , a3 , a4 } with corresponding probabilities
{0.1, 0.2, 0.3, 0.4}.
(c) Design a Huffman code for the source and compare the average length
of the Huffman code with the entropy of the source?
(d) Design a Huffman code for the second extension of the source (take
two symbols at a time)? What is the average codeword length? What
is the average required binary letters per each source output symbol?
107
1-p
x1 y1
p
p
1-p
x2 y2
p
x3 1-p y3
Problem Set: 3
2
• 4.29 (a) with h = 3
• Additional problems:
s1 s2 s3 s4
3A
s5 s6 s7 s8
A
−3A −A A 3A
s9 s 10 s 11 s 12
−A
s 13 s 14 s 15 s 16
−3A
108
S1 (t) S2 (t) S3 (t) S4 (t)
2A 2A 2A 2A
A A A A
0 0 0 0
T/2 T t T/2 T t T/2 T t T/2 T t
−A −A −A −A
109
Problem Set: 4
Note that you have to do only the above problems for credit. However, you
are encouraged to do the following problems for which the solutions will also
be provided.
110
Sample Questions
111
1. Probability and Random Process: (15 pts)
Determine the mean and the autocorrelation function of the random process
1 1 1
t t t
0 2 4 0 4 0 1 2 3 4
−1 −1
112
(a) the output power spectral density Φyy (f ) ?
(c) from the above results in (a) and (b), what can you conclude about
the coding of symbols ?
6. Digital Transmission:
113
a bit rate of Rb = 20 × 103 bits/second, M -ary modulation is used.
Find M .
(b) [15 pts] A video signal is to be stored in a hard drive. The signal
is composed of 24 frames/second with 728 × 512 pixels/frame. The
grayness level of each pixel is represented by R bits. The quantization
process results in signal-to-noise ratio (SNR) of γ(R) = 1.8 + 6R (in
dB). If the output quality requirement dictates an SNR of at least 40
dB, determine how many minutes of video signal can be stored in a 40
GB hard drive?
X(t) = Acos(2πf t + θo ),
8. Signal Space [10 pts]: Signals s1 (t) and s2 (t) are as shown in Figure 13.
Draw two unit-energy signals s3 (t) and s4 (t) which are orthogonal to both
s1 (t) and s2 (t).
114
s (t) s 2(t)
1
A A
T T/2 T
X(t) + d Y(t)
dt
Delay by T
10. Source Coding : [20 pts] A discrete memoryless source emits messages
m1 , m2 and m3 with probabilities 0.2, 0.3 and 0.5 respectively.
115
(a) If two output messages are encoded at a time, determine the optimum
code assignment with fixed code length?
(b) If two output messages are encoded at a time, determine the optimum
code assignment with variable code length?
(c) For both cases above, determine the coding efficiency? What can you
conclude about the coding of symbols from the efficiency?
11. Signals and Vectors [15 pts]: Consider four waveforms as shown in
Fig. 15. (i) Show that {gi (t)}, i = 1, 2, 3, forms an orthogonal set. (ii)
Find the minimal basis functions for the above set. (iii) Express x(t) shown
in Fig. 15 (d) as a vector using the above basis functions.
g (t)
1
g (t)
2
g (t)
3
x(t)
1
1 1 1
t
t t t 0 2 4
0 2 4 0 4 0 1 2 3 4
−1
−1 −1 −1
(d)
(a) (b) (c)
116
13. Random Process [15 pts]: A random process v(t) is defined as
14. (a) Source Coding: [20 pts] A binary erase channel (BEC) has the
transition probabilities as shown in Fig. 16. Assuming equiprobable
inputs, determine (i) the source entropy, H(X) and (ii) conditional mu-
tual information, H(X|Y ), (iii) average mutual information I(X; Y ),
in bits/symbol.
q
x1 y1
p
y3
p
x2 q y2
15. Power Spectral Density [20 pts]: A noise signal ni (t) with PSD Sni (f ) =
K(a constant), for all f, is first filtered using an ideal low pass filter h(t)
117
and then is passed through an ideal differentiator
as shown in Fig. 17. The
1, |f | ≤ B,
filter response is characterized as: H(f ) =
0, otherwise.
(i) Determine the PSD and the power of the noise signal no (t). (ii) Given
3×10−12
that K = 8π 2 , what should be the (ideal) filter bandwidth that will
d
give 1 Watt of output power? Aid: dt ↔ j2πf .
n i (t) d n o(t)
h(t)
dt
118
and is sampled at the Nyquist rate, encoded into a PCM format using
8 bits/sample, and then transmitted through the AWGN channel using
M −ary baseband PAM. Determine the bandwidth required for distortion-
less transmission when M = 16 assuming that the bandwidth of the basic
1
pulse can be approximated by W = T Hz where T is the symbol duration
in seconds. (6 pts)
sin( πt πt
T ) cos( 2T )
(e) Inter Symbol Interference (ISI): Show that x(t) = πt
( T ) 1− t22
T
satisfies the Nyquist criterion for no ISI. Hint: plot x(t) for t = kT, k=
0, 1, 2, .. (8 pts)
(a) The signal constellation for 8-QAM is shown in Fig. 18. If a transmit-
ted signal is x(t) = s1 (t) + s8 (t − T ) where T is the symbol duration,
sketch x(t) assuming the carrier frequency of fc = T1 . (10 pts)
sin 2 πfc t
s1 s2
3A
s5 s6
A
−3A
s3 s4
119
(b) A digital source produces binary data with equal probability. Two
types of modulation schemes are considered for transmission of the
above information through a AWGN channel.
Scheme A: 0’s and 1’s are transmitted using +f (t) and −f (t) respec-
tively where
A 0 ≤ t ≤ T,
f (t) =
0 otherwise.
Scheme B: 0’s and 1’s are transmitted using g(t) and h(t) respec-
tively where
Bsin(2πt/T ) 0 ≤ t ≤ T,
Bcos(2πt/T ) 0 ≤ t ≤ T,
g(t) = h(t) =
0
0
otherwise, otherwise.
(a) Find the overall channel capacity of the cascaded connection in terms
of p assuming that both channels have the same transition probability
as shown in Fig 19. You can also assume that the maximum capacity
is achieved when input symbols are equiprobable, i.e., P (X = 0) =
120
P (X = 1) = 12 . (10 pts)
input output
BSC BSC
composite channel
1−p 1−p
0 0
0 p
p
p p
1−p 1−p
1 1 1
(b) It is known apriori that on the average, 60% of the binary source
outputs are 1’s and the rest are 0’s. A baseband PAM scheme is
employed with rectangular pulse g(t) to transmit through the AWGN
N0
channel with noise PSD of 2 = 5 × 10−12 W/Hz and channel the
bandwidth of 1 kHz. As a system designer, your goal is to achieve
maximum possible reliable transmission rate of 104 bps through the
channel. You are constrained to use only the following waveforms for
transmission: f1 (t) = Ag(t) and f0 (t) = 2Ag(t) where f1 (t) and f0 (t)
are transmitted for 1 and 0 respectively. What should be the value of
A? Hint: use Shannon’s channel capacity limit. (10 pts)
121
by,
1
H(f ) = ,
1 + j2πβf
where β is a constant.
(b) Determine the power spectral density Φxx (f ) of the process x(t). (4
pts)
(e) If the required average output power is Pav = 1Watt, show that β =
√
α. (4 pts)
x(t)
n(t) + h(t) y(t)
filter
T
delay
(a) Find E(bk ), E(b2k ) and variance σb2k of the sequence bk . (5 pts)
122
(b) Find and plot the discrete autocorrelation φbb (m) = E(bk bk+m ). (5
pts)
(c) If g(t) has the raised-cosine spectrum with 100% excess bandwidth,
sketch G(f ). (5 pts)
The noise n(t) is assumed to be zero-mean AWGN with PSD of N0 /2. Due
0
to distortion in the channel, the received pulses have different shapes g (t)
as shown in Fig. 21(c). The receiver is unaware of the distortion in the
channel and therefore it uses the filter that is matched to the transmit-
ted pulse g(t). Note that an optimal receiver should have matched to the
0
received pulse g (t).
(a) What is the impulse response h(t) of the filter used in the above re-
0
ceiver? Show that g(t) and g (t) have equal energy; i.e., Eg = Eg0 = Eb .
Find Eb . (4 pts)
123
Given that a ”1” was transmitted, find (i) s and (ii) the mean and the
variance of r. (4 pts)
(c) Compute and plot pR (r|1) and pR (r|0) on the same graph where pR (.)
denotes the probability density function. (3 pts)
(a)
g(t) g’(t)
8 A
5
A 1 8A
2 5
t 0 t
0 T T/2 T
(b) (c)
Figure 21: (a) A baseband binary communication system through a channel that introduces
0
distortion, (b) transmitted pulse g(t) and (c) channel distorted pulse g (t).
(b) [6 pts] For the QAM signal constellation shown in Figure 1, specify
the Gray code.
124
x
7
5 x
x
3
1 x
−7 −5 −3 −1 1 3 5 7
x x x x x x x x
−1 x
x
−3
x
−5
x
−7
125
(a) [3 pts] Determine the entropy of the source.
(c) [6 pts] Design a Huffman code that encodes two output levels (J = 2)
at a time and determine the average bit rate.
(d) [3 pts] Determine the efficiency of the encoding scheme in (b) and
(c). From the efficiency, what can you conclude about the encoding of
blocks ?
(e) [4 pts] What is the lower bound on the bit rate obtained by encoding
J output levels as J → ∞ ?
(a) [3 pts] Find E(bk ), E(b2k ) and variance, σb2 , of the sequence bk where
E(.) denotes expectation.
(b) [7 pts] Plot the discrete autocorrelation function, φbb (m) = E(bk bk+m )
for A = 12.
126
(d) [7 pts] The power spectral density expression is given as
∞
σb2 2 (E(bk ))2 X m 2 m
Φxx (f ) = |G(f )| + |G( )| δ(f − ).
T T2 m=−∞
T T
For the above raised-cosine pulse and with A = 12, sketch Φxx (f ).
s1 (t) = A, 0 ≤ t ≤ T,
(a) [8 pts] Determine (i) the optimum detector for an AWGN channel and
(ii) the optimum threshold assuming that the signals are equiprobable.
(b) [8 pts] Given that The noise power spectral density is N0 /2, determine
the probability of symbol error as a function of SNR.
(c) [4 pts] How does this signaling scheme compare with the antipodal
signaling ? Hint: Compare the energy required to achieve the same
error performance.
127
A, 0 ≤ t ≤ T,
s1 (t) =
0, otherwise,
A, 0 ≤ t ≤ T /2,
s2 (t) = −s3 (t) = −A, T /2 < t ≤ T,
0, otherwise,
where A is a constant.
(b) [3 pts] Find a basis for the signal space. Hint: Can be done by inspec-
tion.
(c) [4 pts] Give a vector representation of the signals and draw the signal
constellation for this problem.
(e) [8 pts] Prove that the message m1 is more vulnerable to errors. Hint:
Consider P (error|mi transmitted), i = 1, 2, 3.
(a) [3 pts] Find the set of basis function(s), {fi }, needed to represent the
above 4 signals.
128
S (t) S (t) S (t) S (t)
1 2 3 4
2A
0 t 0 t 0 t 0 t
T T T T
−A
hr(t)
x(t)
B
y = m i+ n Decision
t Device
0 T
129
First determine x(t) and then consider minimum-distance criterion.
(c) [4 pts] Sketch fY (y|1), fY (y|2), fY (y|3) and fY (y|4), on the same fig-
ure. Here, fY (y|i) denotes the conditional probability for i th signal
transmitted.
(d) [6 pts] Find the probability of symbol error. Hint: Consider for exam-
ple,
P (error|1st signal transmitted) and its relationship to the overall error
probability.
130
f (t) f (t) f (t)
1 2 3
1 1 1
t t t
0 2 4 0 4 0 1 2 3 4
−1 −1
(c) from the above results in (a) and (b), what can you conclude about
the coding of symbols ?
131
(a) determine the capacity of the channel ?
132