Professional Documents
Culture Documents
Raphael Seebacher
raphasee@ee.ethz.ch
Contents
1 Random Processes 1
1.1 Mathematical Definition . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
1.2 Stationary Processes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
1.3 Mean, Correlation and Covariance . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
1.4 Ergodic Processes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2
1.5 Transmission of a Random Process Through a LTI Filter . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2
1.6 Power Spectral Density . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2
1.7 Gaussian Processes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3
1.8 Noise . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3
1.8.1 White Noise . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3
3 Signal-Space Analysis 9
3.1 Geometric Representation of Signals . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9
3.1.1 Gram-Schmidt Orthogonalization Procedure . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9
3.2 Conversion of the Continuous AWGN Channel into a Vector Channel . . . . . . . . . . . . . . . . . . . . . . . 10
3.2.1 Statistical Characterization of the Correlator Outputs . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
3.3 Likelihood Functions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
3.4 Coherent Detection of Signals in Noise: Maximum Likelihood Decoding . . . . . . . . . . . . . . . . . . . . . 10
3.4.1 Maximum a Posteriori Decoding . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11
3.4.2 Maximum Likelihood Decoding . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11
3.5 Correlation Receiver . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11
3.6 Proability of Error . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11
3.6.1 Invariance of the Probability of Error to Rotation and Translation . . . . . . . . . . . . . . . . . . . . 11
3.6.2 Minimum Energy Signals . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12
3.6.3 Union Bound on the Probability of Error . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12
3.6.4 Bit versus Symbol Error Proababilities . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12
i
4.3.6 Bandwidth Efficiency of M-ary PSK Signals . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17
4.4 Hybrid A/P Modulation Schemes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17
4.4.1 M-ary Quadrature Amplitude Modulation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17
4.4.2 Carrierless Amplitude/Phase Modulation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18
4.5 Coherent Frequency-Shift Keying . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18
4.5.1 Binary FSK . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18
4.5.2 Minimum Shift Keying . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19
4.5.3 Gaussian-Filtered MSK . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21
4.5.4 M-ary FSK . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21
4.6 Detection of Signals with Unknown Phase . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22
4.6.1 Optimum Quadratic Receiver . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22
4.7 Noncoherent Orthogonal Modulation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22
4.7.1 Noncoherent Binary FSK . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 23
4.7.2 Differential PSK . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 23
4.8 Comparison of Digital Modulation Schemes Using a Single Carrier . . . . . . . . . . . . . . . . . . . . . . . . 23
7 Error-Control Coding 39
7.1 Discrete-Memoryless Channels . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 39
7.1.1 Notation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 39
7.2 Linear Block Codes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 39
7.2.1 Syndrome: Definition and Properties . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40
7.2.2 Minimum Distance Considerations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40
7.2.3 Syndrome Detecting . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40
7.2.4 Dual Code . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41
7.3 Cyclic Codes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41
7.3.1 Generator Polynomial . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41
7.3.2 Parity-Check Polynomial . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41
7.3.3 Calculation of the Syndrome . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 42
ii
8 Multiple Access Protocols 43
8.1 ALOHA Throughput . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 43
8.1.1 Assumptions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 43
8.1.2 Throughput pure . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 43
8.1.3 Throughput slotted . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 43
8.1.4 Mean Delay . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 43
8.2 Carrier Sense Multiple Access . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 43
8.2.1 1-persistent CSMA . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 43
8.2.2 Nonpersistent CSMA . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 43
8.2.3 P-persistent CSMA . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 43
8.2.4 Comparison . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 43
A Line Codes 44
B Hilbert Transform 44
iii
iv
Communication Systems Summary CONFIDENTIAL Raphael Seebacher
1 Random Processes
1.1 Mathematical Definition where fX(t) (x) is the first-order probability density function
of the process.
Def 1.1 Sample Space S
The totality of sample points si S corresponding to the ag-
If X(t) is strictly stationary, the mean is a constant:
gregate of all possible outcomes of the experiment is called
the sample space S. X (t) = X t
Special Cases are not sufficient to guarantee that the random process X(t)
is strictly stationary.
For k = 1, we have Such processes are referred to as wide-sense stationary or
FX(t) (x) = FX(t+ ) (x) = FX (x) t, weakly stationary. The class of strictly stationary processes
with finite second-order moments forms a subclass of the
hence the first-order distribution function of a station- class of all stationary processes.
ary random process is independent of time.
Properties of the Autocorrelation Function
For k = 2 and = t1 , we have
RX (0) = E[X 2 (t)]
FX(t1 ),X(t2 ) (x1 , x2 ) = FX(0),X(t2 t1 ) (x1 , x2 ) t1 , t2
RX ( ) = RX ( )
hence the second-order distribution function of a sta-
tionary random process depends only on the time dif- |RX ( )| RX (0)
ference between the observation times.
Def 1.8 Cross-Correlation Function
Consider two random processes X(t) and Y (t) with autocor-
1.3 Mean, Correlation and Covariance relation functions RX (t, u) and RY (t, u).
Def 1.5 Mean The two cross-correlation functions of X(t) and Y (t) are de-
We define the mean of the process X(t) as the expectation fined by
of the random variable obtained by observing the process at
RXY (t, u) = E[X(t)Y (u)] and RY X (t, u) = E[Y (t)X(u)]
some time t.
Z where t and u denote the two values of tim eat which the
X (t) = E[X(t)] = xfX(t) (x)dx processes are observed.
1
Communication Systems Summary CONFIDENTIAL Raphael Seebacher
Def 1.9 Correlation Matrix If X(t) is stationary the mean is constant, and hence
Let X(t) and Y (t) denote two random processes. The cor- Z
relation matrix is given by Y = X h(1 )d1 = X H(0)
" #
RX (t, u) RXY (t, u)
R(t, u) = The autocorrelation function of Y (t) is, provided that
RY X (t, u) RY (t, u)
E[X 2 (t)] < , t and the system is stable
If the random processes X(t) and Y (t) are each stationary
and jointly stationary, the correlation matrix simplifies to RY (t, u) = E[Y (t)Y (u)]
Z Z
d2 h(2 )RX (t 1 , u 2 )
" #
RX ( ) RXY ( ) = d1 h(1 )
R(t, u) =
RY X ( ) RY ( )
If X(t) is stationary, let = t u
where = t u. Z Z
RY ( ) = h(1 )h(2 )RX ( 1 + 2 )d1 d2
RXY ( ) = RY X ( )
The process X(t) is ergodic in the autocorrelation function if Properties of the PSD
limT Rx (, T ) = RX (T ) R
SX (0) = RX ( )d
limT Var[Rx (, T )] = 0 R
E[X 2 (t)] = SX (f )df = RX (0)
For a random process to be ergodic, it has to be stationary.
SX (f ) 0, f
For the mean of Y (t) we get, provided that E[X(t)] < , t The power spectral density of the output process Y (t) equals
and the system is stable the power spectral density of the input process X(t) multi-
Z plied by the squared magnitude response of the filter:
Y (t) = E[Y (t)] = h(1 )X (t 1 )d1 SY (f ) = |H(f )|2 SX (f )
2
Communication Systems Summary CONFIDENTIAL Raphael Seebacher
We obtain the relation between the power spectral density 2. Consider the set of random variables X(t1 ), . . . , X(tn ),
SX (f ) of an ergodic process and the squared magnitude spec- obtained by observing a random process X(t) at times
trum |X(f, T )|2 of a truncated sample function of the pro- t1 , . . . , tn . If the process X(t) is Gaussian, then this set
cess: of random variables is jointly Gaussian for any n, with
1 their n-fold joint probability density function being com-
SX (f ) = lim E[|X(f, T )|2 ] pletely determined by specifying the set of means
T 2T
2
1 T
Z
X(tk ) = E[X(ti )] i = 1, . . . , n
= lim E x(t) exp(j2f t)dt
T 2T T
and the set of covariance functions
Theorem 1.1 Central Limit Theorem then these random variables are statistically independent.
Let Xi , i = 1, . . . , N be a set of independently and identically
distributed (i.i.d.) random variables. 1.8 Noise
They are normalized as follows
1 The term noise is used customarily to designate unwanted
Yi = (Xi X ) i = 1, . . . , N signals that tend to disturb the transmission and processing
X
of signals in communication systems and over which we have
so that E[Yi ] = 0 and Var[Yi ] = 1. incomplete control.
Define furthermore the random variable
N 1.8.1 White Noise
1 X
VN = Yi
N i=1 The noise analysis of communication systems is customarily
based on an idealized form of noise called white noise.
The central limit theorem states that the probability distri-
bution of VN approaches a normalized Gausian distribution N0 N0
N (0, 1) in the limit as the number of random variables N SW (f ) = RW ( ) = ( )
2 2
approaches infinity.
where the dimension of N0 is watts per Hertz, and
Properties of a Gaussian Process
N0 = kB Te
1. If a Gaussian process X(t) is applied to a stable linear fil-
ter, then the rando process Y (t) developed at the output where kB is the Boltzmanns constant and Te is the equiva-
of the filter is also Gaussian. lent noise temperature of the receiver.
3
Communication Systems Summary CONFIDENTIAL Raphael Seebacher
By applying Schwarzs inequality, we get for the above ex- Let y denote the sample value obtained at the end of a sig-
pression Z naling interval. The sample value y is compared to a preset
2 threshold in the decision device.
|G(f )|2 df
N0 There are two possible kinds of error:
For the optimum value of H(f we get (as a result of Schwarzs 1. Symbol 1 is chosen when a 0 was actually transmitted
inequality) (error of the first kind).
Hopt (f ) = kG (f ) exp(j2f T ) df 2. Symbol 0 is chosen when a 1 was actually transmitted
(error of the second kind).
where G (f ) is the complex conjugate of the Fourier trans-
form of the input signal g(t) and k is a scaling factor. Suppose that symbol 0 was sent. The matched filter output,
Since for a real signal g(t) we have G (f ) = G(f ), we get sampled at time t = Tb is then given by
Z Z Tb Z Tb
1
hopt (t) = k G(f ) exp(j2f (T t)) df y= x(t) dt = A + w(t) dt
0 Tb 0
= kg(T t) which represents the sample value of a random variable Y .
5
Communication Systems Summary CONFIDENTIAL Raphael Seebacher
We may characterize the random variable Y as follows: Def 2.4 Binary Symmetric Channel
A channel for which the conditional probabilities of error p01
Y is Gaussian distributed with E[Y ] = A. and p10 are equal is said to be binary symmetric.
The variance of Y is
Def 2.5 Transmitted Signal Energy per Bit
2 2
Y = E[(Y + A) ]
Z Tb Z Tb Eb = A2 Tb
1
= 2 RW (t, u) dt du
Tb 0 0 Def 2.6 Average Probability of Symbol Error in a
N0
Binary Symmetric Channel
with RW (t, u) = 2 (t u) we get
r !
1 Eb
N0 Pe = erfc
Y 2 = 2 N0
2Tb
We then have for p10 , the conditional probability of error 2.3 Intersymbol Interference
given that symbol 0 was sent Intersymbol Interference (ISI) is another source of bit errors,
which arises when the communication channel is dispersive.
p10 = P (y > | symbol 0 was sent) For the baseband transmission of digital data, discrete pulse-
Z
amplitude modulation (PAM) is one of the most efficient
= fY (y|0) dy
schemes in terms of power and bandwidth utilization.
Z
(y + A)2
1
=p exp dy Consider a baseband binary PAM system. The incoming bi-
N0 /Tb N0 /Tb
nary sequence {bk } consists of symbols 1 and 0, each of dura-
tion Tb . The pulse-amplitude modulator modifies this binary
Def 2.3 Complementary Error Function sequence into (
2
Z +1 if bk = 1
erfc(u) = exp(z 2 ) dz ak =
u 1 if bk = 0
This sequence of short pulses is applied to a transmit filter
By defining the new variable of impulse response g(t), producing the transmitted signal
X
y+A s(t) = ak g(t kTb )
z=p
N0 /Tb k
N0
p0 2.4 Nyquists Criterion
opt = log
4ATb p1 By controlling the overall pulse p(t), we can ensure perfect
reception in the absence of noise
For the special case when both symbols are equiprobable, we (
have 1 i=k
p(iTb kTb ) =
0 i 6= k
1
p0 = p1 = = opt = 0 = p01 = p10
2 where p(0) = 1 by normalization.
6
Communication Systems Summary CONFIDENTIAL Raphael Seebacher
Def 2.9 Ideal Nyquist Channel This time response has two extremely useful properties
By applying the inverse Fourier transform we get the ideal
Nyquist channel At t = Tb /2 = 1/4W we have p(t) = 0.5.
Def 2.13 Rolloff Factor For the overall frequency response of the duobinary signaling
The parameter is called the rolloff factor scheme we have
(
f1 2 cos(f Tb ) exp(jf Tb ) |f | 2T1 b
=1 HI (f ) =
W 0 otherwise
7
Communication Systems Summary CONFIDENTIAL Raphael Seebacher
The original two-level sequence {ak } may be detected by in- In an M-ary PAM system, 1 baud is equal to log2 M bits
voking the use of per second, and the symbol duration of the M-ary PAM sys-
ak = ck ak1 tem is related to the bit duration Tb of the equivalent binary
where ak denotes the estimate of the original pulse ak . The PAM system as
technique of using a stored estimate of the previous symbol T = Tb log2 M
is called decision feedback.
Therefore, in a given channel bandwidth, we find that by
However, a major drawback of this detection procedure is
using an M-ary PAM system, we are able to transmit infor-
that once errors are made, they tend to propagate through
mation at a rate that is log2 M faster, than the corresponding
the output. A practical means of avoiding this error-
binary PAM system.
propagation phenomenon is to use precoding before the duobi-
However, to realize the same average probability of error, an
nary coding.
M-ary PAM system requires more transmitted power.
dk = bk dk1
where denotes modulo-two addition.
This nonlinear precoding leads to
(
0 if data symbol bk = 1
ck =
2 if data symbol bk = 0
8
Communication Systems Summary CONFIDENTIAL Raphael Seebacher
3 Signal-Space Analysis
A message source emits one symbol every T seconds, with The set of coefficients {sij }N j=1 may naturally be viewed as
the symbols belonging to an alphabet of M symbols denoted an N -dimensional vector, denoted by si . Hence, every signal
by m1 , m2 , . . . , mM . in the set {si (t)} is completely characterized by the vector
It is customary to assume that the M symbols of the alpha- of its coefficients, i.e. its signal vector
bet are equally likely
h iT
1 si = si1 si2 ... siN i = 1, . . . , M
pi = P (mi ) = i = 1, . . . , M
M
The transmitter takes the message source output mi and We may visualize the set of signal vectors {si | i = 1, . . . , M }
codes it into a distinct signal si (t), which occupies the full du- as defining a corresponding set of M points in an N -
ration T allotted to symbol mi . si (t) is a real-valued energy- dimensional Euclidean space, with N mutually prependic-
signal ular axes labeled 1 , . . . , N . This N -dimensional Euclidean
Z T space is called the signal space.
Ei = si 2 (t) dt i = 1, . . . , M
0
Def 3.2 Inner Product
Def 3.1 AWGN Channel The squared-length of any signal vector si is defined to be
A channel having the following two characteristics is referred the inner product or dot product of si with itself
to as an additive white Gaussian noise (AWGN) channel:
N
X
The channel is linear, with a bandwidth that is wide ||si ||2 = si T si = sij 2 i = 1, . . . , M
enough. j=1
The receiver has the task of observing the received signal Def 3.4 Euclidean Distance between Signal Vectors
x(t) for a duration of T seconds and making a best estimate
of the transmitted signal si (t). N
X Z T
The requirement is therefore to desing the receiver so as to dik = ||si sk ||2 = (sij skj )2 = (si (t) sk (t))2 dt
0
minimize average probability of symbol error defined as j=1
M
X Def 3.5 Angle between Signal Vectors
Pe = pi P (m 6= mi |mi )
i=1
si T sk
where mi is the transmitted symbol and m is the estimate. cos(ik ) =
||si || ||sk ||
where the coefficients of the expansion are defined by where the coefficients sij are defined by
Z T Z T
sij = si (t)j (t) dt i, j = 1, . . . , M sij = si (t)j (t) dt j = 1, . . . , i 1
0 0
9
Communication Systems Summary CONFIDENTIAL Raphael Seebacher
3.2 Conversion of the Continuous AWGN where the vector x is called the observation vector, and xj is
Channel into a Vector Channel called an observable element.
By substituting Gaussian probability densitiy function in the
Suppose that the input to the bank of N product integrators above equation, we get
or correlators is not the transmitted signal si (t) but rather
the received signal x(t).
N
1 X
fX (x|mi ) = (N0 )N/2 exp (xj sij )2
x(t) = si (t) + w(t) 0 t T, i = 1, . . . , M N0 j=1
where w(t) is a sample function of a white Gaussian noise
process W (t) of zero mean and power spectral density N0 /2. Finally, we note that any random variable W 0 (tk ), derived
At the output of correlator j, we have a sample of a random from the noise process W 0 (t) by sampling it at time tk , is in
variable Xj fact statistically independent of the set of random variables
Z T {Xj }. The random variable W 0 (tk ) is irrelevant to the de-
cision as to which particular signal was actually transmitted.
xj = x(t)j (t) dt = sij + wj j = 1, . . . , N
0
with RW (t, u) = N0
u) we get Def 3.8 Log-Likelihood Function
2 (t
N0
Z T
N0 l(mi ) = log L(mi ) i = 1, . . . , M
= j 2 (t) dt =
2 0 2
Def 3.9 Log-Likelihood Function for an AWGN
The Xj are mutually uncorrelated, since Channel
cov[Xj Xk ] = 0 j 6= k N
1 X
and because the Xj are Gaussian random variables, they are l(mi ) = (xj sij )2 i = 1, . . . , M
N0 j=1
also statistically independent.
Define the vector of N random variables
h iT 3.4 Coherent Detection of Signals in Noise:
X = X1 X2 . . . XN Maximum Likelihood Decoding
whose elements are independent Gaussian random variables Suppose that in each time slot of duration T seconds, one of
with mean values equal to sij and variances equal to N0 /2. the M possible signals s1 (t), . . . , sM (t) is transmitted with
1
equal probability M .
We may therefore express the conditional probability density The corresponding point of si (t) in the Euclidean space of
function of the vector X, given the transmitted symbol mi , dimension N M is referred to as the transmitted signal
as the product of the conditional probability density func- point or message point. The set of message points is called a
tions of its individual elements signal constellation.
N Based on the observation vector x, we may represent the re-
Y
fX (x|mi ) = fXj (xj |mi ) i = 1, . . . , M ceived signal x(t) by a point in the same Euclidean space.
j=1 This point is referred to as the received signal point.
10
Communication Systems Summary CONFIDENTIAL Raphael Seebacher
Def 3.11 Maximum a Posteriori Probability Rule 2. The signal transmission decoder is implemented in form
Set m = mi if of a maximum-likelihood decoder.
The optimum receiver is commonly reffered to as a correla-
P (mi sent | x) P (mK sent | x) k 6= i tion receiver.
Using Bayes rule we get for the MAP Rule
Def 3.13 Equivalence of Correlation and Matched
Set m = mi if
Filter Receivers
pk fX (x|mk ) The correlation receiver involves a set of correlators. Alter-
is maximum for k = i natively, we may use a corresponding set of matched filters.
fX (x)
The maximum likelihood decision rule hence is simply to Def 3.15 Principle of Translational Invariance
choose the message point closest to the received signal point. If a signal constellation is translated by a constant vector
amount, then the probability of symbol error Pe incurred in
Observation vector x lies in region Zi if maximum likelihood signal detection over an AWGN channel
PN 1
j=1 xj skj 2 Ek is maximum for k = i is completely unchanged.
11
Communication Systems Summary CONFIDENTIAL Raphael Seebacher
3.6.2 Minimum Energy Signals The probability of symbol error, averaged over all the M
symbols, is therefore overbounded as follows:
A ueful application of the principle of translational invari-
ance is in the translation of a given signal constellation in M
X
such a way that the average energy is minimized. Pe = pi Pe (mi )
Given a signal constellation {si }M
i=1 , the corresponding sig- i=1
nal constellation with minimum average energy is obtained M M
1X d
ik
X
by subtracting from each signal vector si in the given con- erfc
stellation an amount equal to the constant vector E[s] with 2 i=1 2 N0
k=1, k6=i
M
X where pi is the probability of transmitting symbol mi .
E[s] = si pi
i=1 1. Suppose that the signal constellation is circularly sym-
where pi is the probability that symbol mi was been emitted. metric about the origin. Hence
The minimum average energy of the translated signal con-
M
stellation is then 1 d
ik
X
Pe erfc i
2 2 N0
Etranslate,min = E ||E[s]||2 k=1, k6=i
12
Communication Systems Summary CONFIDENTIAL Raphael Seebacher
4.1.3 Power Spectra The receiver, which consists of a detector followed by a signal
transmission decoder, performs two functions:
The power spectra of the resulting modulated signals is par- 1. It reverses the operations performed in the transmitter.
ticularly important in two contexts: occupancy of the chan-
nel bandwidth and co-channel interference in multiplexed 2. It minimizes the effect of channel noise on the estimate m
systems. computed for the transmitted symbol mi .m
13
Communication Systems Summary CONFIDENTIAL Raphael Seebacher
4.3 Coherent Phase-Shift Keying The conditional probability density function of random vari-
able X1 , given that symbol 0 was transmitted is defined by
4.3.1 Binary Phase-Shift Keying
1 1 2
In a coherent BPSK system, the pair of signals below is used fX1 (x1 |0) = exp (x1 s21 )
to represent binary symbols 1 and 0 N0 N0
p 2
1 1
r = exp x1 Eb
2Eb N0 N0
s1 (t) = cos(2fc t)
T
r b r The conditional probability of the receiver deciding in favor
2Eb 2Eb of symbol 1, given that symbol 0 was transmitted, is
s2 (t) = cos(2fc t + ) = cos(2fc t)
Tb Tb r !
Z
1 Eb
where 0 t Tb and Eb is the transmitted signal energy p10 = fX1 (x1 |0) dx1 = erfc
0 2 N 0
per bit. The carrier frequency fc is chosen equal to nc /Tb for
some fixed integer nc . Consider now an error of the second kind. We note that the
signal space is symmetric with respect to the origin.
Def 4.1 Antipodal Signals
A pair of sinusoidal waves that differ only in a relative phase- Thus, averaging the conditional error probabilities p10 and
shift of 180 degrees are referred to as antipodal signals. p01 , we find that the average probability of symbol error or
equivalently the bit error rate for coherent BPSK is (assum-
Signal Space ing equiprobable symbols)
In BPSK there is only one basis function r !
1 Eb
r
2 Pe = erfc
1 (t) = cos(2fc t) 0 t < Tb 2 N0
Tb
Power Spectral Density
For the signals s1 (t) and s2 (t) we then have The complex envelope of a BPSK wave consists of an in-
p p phase component only. The signal shaping function g is de-
s1 (t) = Eb 1 (t) s2 (t) = Eb 1 (t) 0 t < Tb fined by (q
2Eb
0 t Tb
The one-dimensional signal space consists of two message g(t) Tb
points 0 otherwise
We assume that the input binary wave is random, with sym-
bols 1 and 0 equally likely and the symbols transmitted dur-
ing the different time slots being statistically independent.
The baseband power spectral density of a BPSK signal equals
14
Communication Systems Summary CONFIDENTIAL Raphael Seebacher
There are four message points, and the associated signal vec- The average probability of a correct decision from the com-
tors are defined by bined action of two channels working together is
" # r ! r !
E cos (2i 1) 4 E 1 E
si = i = 1, 2, 3, 4 Pc = (1 P 0 )2 = 1 erfc + erfc2
E sin (2i 1) 4 2N0 4 2N0
Accordingly, a QPSK signal has a two-dimensional signal The average probability of symbol error for coherent QPSK
constellation and four message points whose phase angles is therefore
increase in a counterclockwise direction. As with BPSK, the r ! r !
E 1 2 E
QPSK signal has minimum average energy. Pe = 1 Pc = erfc erfc
2N0 4 2N0
15
Communication Systems Summary CONFIDENTIAL Raphael Seebacher
1 1 Error Probability
0 0 Suppose that the transmitted signal corresponds to message
point m1 . Suppose that the ratio E/N0 is large enough to
consider the nearest two message points as potential candi-
dates for being mistaken for m1 due to channel noise.
Attractive features of the /4-shifted QPSK scheme include The average probability of symbol error for coherent M-ary
the following PSK is r !
E
The phase transitions from one symbol to the next are Pe ' erfc sin
N0 M
restricted to /4 and 3/4. Consequently, envelope
variations of /4-shifted QPSK signals due to filtering where it is assumed that M 4. The approximation be-
are significantly reduced, compared to those in QPSK. comes extremely tight, for fixed M , as E/N0 is increased.
Unlike offset QPSK signals, /4-shifted QPSK signals Power Spectral Density
can be noncoherently detected. The symbol duration of M-ary PSK is defined by
Like QPSK signals, /4-shifted QPSK signals can be T = Tb log2 M
differently encoded, in which case we should really
speak of /4-shifted DQPSK. where Tb is the bit duration.
16
Communication Systems Summary CONFIDENTIAL Raphael Seebacher
The baseband power spectral density of an M-ary PSK signal where E0 denotes the energy of the signal with the lowest
is given by amplitude. The transmitted M-ary QAM signal for symbol
k is then defined by
SB (f ) = 2E sinc2 (T f ) = 2Eb log2 (M ) sinc2 Tb f log2 (M )
r r
2E0 2E0
sk (t) = ak cos(2fc t) bk sin(2fc t)
4.3.6 Bandwidth Efficiency of M-ary PSK Signals T T
The power spectra of M-ary PSk signals possess a main lobe with 0 t T, k Z.
bounded by well-defined spectral nulls. Depending on the number of possible symbols M , we may
distinguish two distinct QAM constellations:
Def 4.2 Null-To-Null Bandwidth
The null-to-null bandwidth is defined as the spectral width of QAM Square Constellations
the main lobe, wich provides a simple and popular measure With an even number of bits per symbol, we may write
for the bandwidth of M-ary PSK signals.
L= M
The channel bandwidth required to pass M-ary PSK signals where L is a positive integer. Under this condition, an M-ary
(more precisely, the main spectral lobe of M-ary signals) is QAM square constellation can always e viewed as the Carte-
given by sian product of a one-dimensional L-ary PAM constellation
2 2Rb with itself.
B= =
T log2 M In the case of a QAM square consteallation, the ordered pairs
where T is the symbol duration. of coordinates naturally forms a square matrix {ai , bi } =
The bandwidth efficiency of M-ary PSK signals is then given
by (L + 1, L 1) (L + 3, L 1) . . . (L 1, L 1)
Rb log2 M (L + 1, L 3) (L + 3, L 3) . . . (L 1, L 3)
= = .. .. ..
B 2
. . .
Observation
As the number of states M is increased, the bandwidth effi- (L + 1, L + 1) (L + 3, L + 1) . . . (L 1, L + 1)
ciency is improved at the expense of error performance. To To calculate the probability of symbol error for M-ary QAM,
ensure that there is no degradation in error performance, we we exploit the property that a QAM square constellation can
have to increase Eb /N0 to compensate for the increase in M . be factored into the product of the corresponding PAM con-
stellation with itself.
4.4 Hybrid A/P Modulation Schemes The probability of correct detection for M-ary QAM may be
written as
In an M-ary PSK system, the in-phase and quadrature com-
Pc = (1 Pe0 )2
ponents of the modulated signal are interrelated in such a
way that the envelope is constrained to remain constant. If where r !
this constraint is removed, and the in-phase and quadrature 1 E0
Pe0 = 1 erfc
components are thereby permitted to be independent, we get M N0
a new modulation scheme called M-ary quadrature amplitude
Assuming that Pe0 is small enough compared to unity, we
modulation (QAM).
find that the probability of symbol error for M-ary QAM is
The passband basis functions in M-ary QAM may not be pe-
approximately given by
riodic for an arbitrary choice of the carrier frequency fc with
respect to the symbol rate 1/T . Ordinarily, this aperiodicity
1
r !
E0
is of no real concern. Pe ' 2 1 erfc
M N0
4.4.1 M-ary Quadrature Amplitude Modulation The transmitted energy in M-ary QAM is variable in that its
M-ary QAM is a two-dimensional generalization of M-ary instantaneous value depends on the particular symbol trans-
PAM in that its formulation involves two orthogonal pass- mitted. It is therefore more logical to express Pe in terms of
band basis functions the average value of the transmitted energy rather than E0 .
r Assuming that the L amplitude levels of the in-phase or
2 quadrature component are equally likely, we have
1 (t) = cos(2fc t) 0tT
T
r L/2
2 2E 0
X
2 (t) = sin(2fc t) 0tT Eav = 2 (2i 1)2
T L i=1
Let the message point si in the (1 , 2 ) plane be denoted by
Accordingly we may rewrite the equation for Pe using Eav
as
dmin dmin
ai , bi s !
2 2 1 3Eav
Pe ' 2 1 erfc
M 2(M 1)N0
where dmin is the minimum distance between any two mes-
sage points in the constellation; ai and bi are integers, and QAM Cross Constellations
i = 1, . . . , M . Let p To generate an M-ary QAM signal with an odd number of
dmin = E0 bits per symbol, we require the use of a cross constellation.
17
Communication Systems Summary CONFIDENTIAL Raphael Seebacher
The probability of symbol error Pe can approximately be 3. The transmitted signal s(t) represents a symbol-time-
written as invariant realization of this hybrid modulation process.
1
r !
E0 The transmitted signal
Pe ' 2 1 erfc
M N 0 X
s(t) = ak p(t kT ) bk p(t kT )
k=
for high E0 /N0 . Note that it is not possible to perfectly Gray
code a QAM cross constellation. is known as carrierless amplitude/phase modulation (CAP).
4.4.2 Carrierless Amplitude/Phase Modulation Properties of the Passband In-phase and Quadrature
Pulses
So far a rectangular pulse has been used for the pulse-shaping
function. 1. The passband in-phase pulse p(t) and quadrature pulse
We redefine the transmitted M-ary QAM signal in terms of p(t) are even and odd functions of time t, given that the
a general pulse-shaping function g(t) as baseband pulse g(t) is an even function of time t.
2. The passband pulses p(t) and p(t) form an orthogonal set
sk (t) = ak g(t kT ) cos(2fc t) bk g(t kT ) sin(2fc t) over the entire interval (, ) as shown by
Z
0 t T , k Z. The transmitted signal s(t) can then be
p(t)p(t) dt = 0
written as
X
s(t) = sk (t)
3. The passband pulses u(t) = p(t)h(t) and u(t) = p(t)h(t)
k=
where h(t) is the impulse response of a LTI channel, form
This equation shows that for an arbitrary fc , the passband a Hilbert-transform pair and are therefore orthogonal over
functions g(t kT ) cos(2fc t) and g(t kT ) sin(2fc t) are the entire interval (, ) for any h(t).
aperiodic in that they vary from one symbol to another.
We can further write s(t) in a complex notation 4.5 Coherent Frequency-Shift Keying
!
X M-ary PSK as well as QAM are examples of linear mod-
s(t) = Re Ak g(t kT ) exp(j2fc t) ulation, whereas coherent frequency-shift keying (FSK) is a
k= nonlinear method of passband data transmission.
where Ak is a complex number defined by Ak = ak + jbk .
Define 4.5.1 Binary FSK
In a binary FSK system, symbols 1 and 0 are distinguished
Ak = Ak exp(j2fc kT ) and g+ (t) = g(t) exp(j2fc t) from each other by transmitting one of two sinusoidal waves
that differ in frequency by a fixed amount.
The scalar Ak is a rotated version of the complex represen-
tation of the coordinates of the kth transmitted symbol in
The FSK signal
the (1 , 2 )-plane. r
2Eb
cos(2fi t) 0 t Tb
We assume that the pulse-shaping function g(t) is a low-pass si (t) = Tb
signal whose highest frequency component is smaller than
0 elsewhere
the carrier frequency fc . We then recognize g+ (t) as the an-
alytic signal, or pre-envelope, representation of the band-pass where i = 1, 2, and Eb is the transmitted signal energy per bit
signal g(t) cos(2fc t). is known as Sundes FSK, which is a continuous-phase sig-
nal in the sense that phase continuity is always maintained.
g+ (t) = g(t) cos(2fc t) + jg(t) sin(2fc t) = p(t) + j p(t) It is, hence, an example of cintinuous-phase frequency-shift
keying (CPFSK).
We may then say that the quadrature (imaginary) compo- The transmitted frequency is
nent p(t) of the analytic signal g+ (t) is the Hilbert transform nc + i
of the in-phase (real) component p(t). fi =
Tb
g(t) is referred to as the baseband pulse, and p(t) and p(t)
are referred to as passband in-phase and passband quadrature for some fixed integer nc and i = 1, 2.
pulses.
Signal Space
!
X
s(t) = Re Ak g+ (t kT ) The set of orthonormal basis functions is given by
k= (q
2
Important Observations Tb cos(2fi t) 0 t Tb
i (t) = i = 1, 2
0 elsewhere
1. The transmitted signal s(t) appears to be carrierless.
Correspondingly the coefficient sij is defined by
2. The new formulation of s(t) fully retains the hybridized (
amplitude and phase modulation characterizing the orig- Eb i = j
sij = i, j {1, 2}
inal M-ary QAM signal 0 i 6= j
18
Communication Systems Summary CONFIDENTIAL Raphael Seebacher
The two message points are defined by symbols transmitted in adjacent time slots are statistically
" # " # independent.
Eb 0 The baseband power spectral density of Sundes FSK signal
s1 = and s2 =
0 Eb equals the sum of the power spectral densities of the in-phase
and the quadrature component.
with the Euclidean distance between them equal to 2Eb . 8E cos2 (T f )
Eb
b b
SB (f ) = f 2T1 b + f + 2T1 b + 2
Error Probability 2Tb (4Tb 2 f 2 1)2
The observation vector x has two elements The presence of the two discrete frequency components ((.))
Z Tb Z Tb provides a means of synchronizing the receiver with the
x1 = x(t)1 (t) dt and x2 = x(t)2 (t) dt transmitter.
0 0
Since the baseband PSD falls of as the inverse fourth power
where x(t) denotes the received signal. of frequency, an FSK signal with continuous phase does not
produce as much interference outsite the signal band of in-
Given that symbol 1 was transmitted, x(t) equals s1 (t)+w(t), terest as an FSK signal with discontinuous phase.
where w(t) is the sample function of a white Gaussian noise
process of zero mean and power spectral density N0 /2. If 4.5.2 Minimum Shift Keying
symbol 0 was transmitted, x(t) equals s2 (t) + w(t).
In the coherent detection of binary FSK signal, the phase
information contained in the received signal is not fully ex-
Define a new Gaussian random variable Y whose sample
ploited, other than to provide for synchronization of the re-
value is defined by
ceiver to the transmitter.
y = x1 x2
Another useful way of representing the CPFSK signal s(t) is
The mean value of the random variable Y depends on which to express it in the conventional form of an angle-modulated
binary symbol was transmitted. The conditional means of Y signal: r
are 2Eb
s(t) = cos 2fc t + (t)
p p Tb
E[Y |1] = + Eb and E[Y |0] = Eb
where (t) is the phase of s(t).
The variance of random variable Y is given by The phase (t) of a CPFSK signal increases or decreases lin-
early with time during each bit duration Tb seconds:
Var[Y ] = Var[X1 ] + Var[X2 ] = N0
h
(t) = (0) t 0 t Tb
We get for the conditional probability of error, given that Tb
symbol 0 was transmitted
where the plus sign corresponds to sending symbol 1, and
r ! the minus sign corresponds to sending symbol 0; the dimen-
1 Eb
p10 = erfc sionless parameter h is referred to as the deviation ratio.
2 2Tb ) (
f1 = fc + 2Thb fc = 12 (f1 + f2 )
Similarly, we may show that p01 has the same value as p10 .
f2 = fc 2Thb h = Tb (f1 f2 )
We hence find that the average probability of bit error, or the
bit error rate for coherent binary FSK is (assuming equiprob- Phase Trellis
able symbols) ! The sending of symbol 1 increases the phase of a CPFSK
r
1 Eb signal s(t) by h radians, whereas the sending of symbol 0
Pe = erfc
2 2Tb reduces it by an equal amount
(
In a coherent binary FSK system, we have to double the bit h for symbol 1
energy-to-noise density ratio Eb /N0 , to maintain the same (Tb ) (0) =
h for symbol 0
bit error rate as in a coherent BPSK system.
The phase of a CPFSK signal is an odd or even multiple of
Power Spectral Density h radiands at odd or even multiples of the bit duration Tb ,
Consider the case of Sundes FSK. We may express this spe- respectively.
cial BFSK signal as follows: For Sundes FSK the deviation ratio h is exactly unity.
r Hence, the phase change over one bit interval is radians.
2Eb t But, a change of + radians is exactly the same as a change
s(t) = cos 2fc t 0 t Tb
Tb Tb of radians, modulo 2. It follows therefore that in the
case of Sundes FSK there is no memory.
r
2Eb t t
= cos cos(2fc t) sin sin(2fc t)
Tb Tb Tb
In contrast, we have a completely different situation when
where, in the last line, the plus sign corresponds to trans- the deviation ratio h is assigned the special value of 1/2,
mitting symbol 0, and the minus sign corresponds to trans- where the difference between the frequencies is half the bi-
mitting symbol 1. rate. This is, hence, the minimum difference, for which
We assume that the symbols 1 and 0 in the random binary s1 1 (t) and s2 2 (t) are orthogonal.
wave at the modulator input are equally likely, and that the
19
Communication Systems Summary CONFIDENTIAL Raphael Seebacher
Signal Space
With the deviation ratio defined as above (h = 1/2), we have
(t) = (0) t 0 t Tb
2Tb
The time interval 0 t Tb , for which the phase states If x2 > 0, the receiver chooses the estimate (Tb ) = /2.
(0 and (Tb are defined, is common to both signals. If x2 < 0, it chooses the estimate (Tb ) = /2.
Accordingly, the signal constellation for an MSK signal is
The bit error rate for coherent MSK is given by
two-dimensional with four possible message points with co- r !
ordinates 1 Eb
( Pe = erfc
( 2 N0
( Eb , Eb ) ( Eb , Eb )
symbol 0 symbol 1 which is exactly the same as that for BPSK and QPSK.
( Eb , Eb ) ( Eb , Eb )
Note that the coordinates of the message points s1 and s2 Power Spectral Density
have opposite signs when symbol 1 is sent in this interval, As with the BFSK signal, we assume that the input binary
but the same sign when symbol 0 is sent. wave is random, with symbols 1 and 0 equally likely, and
In QPSK the transmitted symbol is represented by any one the symbols transmitted during different time slots being
of the four message points, whereas in MSK one of two mes- statistically independent.
sage points is used to represent the transmitted symbol at
any one time, depending on the value of (0). The baseband power spectral density of the MSK signal is
given by
2
Transmitted (0) (Tb ) s1 s2 32Eb cos(2Tb f )
SB (f ) =
Binary Symbol 2 16Tb 2 f 2 1
0 0 2 + Eb + Eb For f 1/Tb , the baseband PSD of the MSK signal falls off
1 2 Eb + Eb as the inverse fourth power of frequency, whereas in the case
of the QPSK signal it falls off as the inverse square of fre-
0 + 2 Eb Eb
quency. Hence, MSK does not produce as much interference
1 0 + 2 + Eb Eb outsite the signal band of interest as QPSK.
20
Communication Systems Summary CONFIDENTIAL Raphael Seebacher
Coherent detection performance equialent to QPSK where Eb is the signal energy per bit and N0 /2 is the noise
spectral density. The constant depends on the time-
Bringing the power spectrum of the MSK signal in a compact bandwidth product W Tb .
form, while maintaining the constant-envelope property, can
be achieved through the use of a premodulation low-pass fil- We may view 10 log10 (/2), expressed in decibels, as a mea-
ter, hereafter referred to as a baseband pulse-shaping filter. sure of performance degradation of GMSK compared to or-
It should satisfy the follwing properties dinary MSK.
1. Frequency response with narrow bandwidth and sharp
4.5.4 M-ary FSK
cutoff characteristics.
Consider the M-ary version of FSK, for which the transmit-
2. Impulse response with relatively low overshoot. ted signals are defined by
3. Evolution of a phase trellis where the carrier phase of the
r
2E
modulated signal assumes the two values /2 at odd si (t) = cos (nc + i)t 0tT
T T
multiples of Tb and the two values 0 and at even multi-
ples of Tb as in MSK. where i = 1, . . . , M , and the carrier frequency fc = nc /2T
for some fixed integer nc . The transmitted symbols are of
These desirable properties can be achieved by passing a equal duration T and have equal energy E.
nonreturn-to-zero (NRZ) binary data stream through a base-
band pulse-shaping filter whose impulse response is defined Signal Space
by a Gaussian function. The resulting method of binary Since the individual signals in M-ary FSK are orthogonal,
frequency modulation is naturally referred to as Gaussian- we may use the transmitted signals in normalized form as a
filtered MSK or just GMSK. complete orthonormal set of basis functions
1
Let W denote the 3dB baseband bandwidth of the pulse- i (t) = si (t) 0 t T, i = 1, . . . , M
shaping filter. E
Accordingly, the M-ary FSK is described by an M -
2 !
log 2 f dimensional signal-space diagram.
H(f ) = exp
2 W
Error Probability
and For coherent M-ary FSK, the optimum receiver consists of
2 2 2 2
r
2 a bank of M correlators or matched filters, with the i (t)
h(t) = W exp W t providing the pertinent reference signals.
log 2 log 2
21
Communication Systems Summary CONFIDENTIAL Raphael Seebacher
With the minimum distance dmin in M-ary FSK being 2E In order to design an optimum receiver for detecting sym-
and equiprobable symbols, we get bol si , we may formulate the conditional likelihood function,
r ! given the carrier phase
1 E r Z T !
Pe (M 1) erfc E
2 2N0 L si () = exp x(t) cos(2fi t + ) dt
N0 T 0
Power Spectral Density By integrating the conditional likelihood function over all
The spectral analysis of M-ary FSK signals is much more possible values of , we remove the phase dependence of
complicated than that of M-ary PSK signals. L(si ()) and obtain
Z
Bandwidth Efficiency L(si ) = L si () f () d
We may define the channel bandwidth required to transmit
Z r Z T !
M-ary FSK signals as 1 E
= exp x(t) cos(2fi t + ) dt d
2 N0 T 0
M Rb M
B= =
2T 2 log2 M Using trigonometric formulas, and defining
v
u T 2 T 2
The bandwidth efficiency of M-ary signals is therefore u Z
u
Z
li = t x(t) cos(2fi t)dt + x(t) sin(2fi t)dt
Rb 2 log2 M
= = 0 0
B M RT !
0
x(t) sin(2fi t)dt
We see that increasing the number of levels M tends to in- i = arctan RT
0
x(t) cos(2fi t)dt
crease the bandwidth efficiency of M-ary PSK signals, but
it also tends to decrease the bandwidth efficiency of M-ary we obtain
FSK signals. Z T
x(t) cos(2fi t + ) dt = li cos( + i )
0
4.6 Detection of Signals with Unknown and can simplify the above likelihood function to
Phase Z r !
1 E
L(si ) = exp li cos d
Up to this point, we have assumed that the receiver is per- 2 N0 T
fectly synchronized to the transmitter, and the only channel
impairment is noise. Perhaps the most common random sig- which we recognize as modified Bessel function of zero order.
nal parameter is the carrier phase, which is especially true Hence r !
for narrowband signals. E
L(si ) = I0 li
A digital communication receiver with no provision made for N0 T
carrier phase recovery is said to be noncoherent. The binary hypothesis test (i.e. the hypothesis that signal
s1 (t) or signal s2 (t) was transmitted) can now be written as
4.6.1 Optimum Quadratic Receiver r ! r !
E H1 E
I0 l1 I 0 l2
Consider a binary digital communication system in which N0 T H2 N0 T
the transmitted signal is
where hypothesis H1 and H2 correspond to signals s1 (t) and
r s2 (t), respectively.
2E
si (t) = cos(2fi t) 0 t T, i = 1, 2 Since the modified Bessel function is a monotonically increas-
T ing function, we can carry out the test in terms of li 2
where E is the signal energy, T is the duration of the signal- H1
ing interval, and the carrier frequency fi for symbol i is an l1 2 l2 2
H2
integral multiple of T2 .
The received signal, assuming an AWGN channel and a non- 4.7 Noncoherent Orthogonal Modulation
coherent system, may be written as
r The noise performance of noncoherent orthogonal modulation
2E is now studied for two noncoherent receivers as special cases:
x(t) = cos(2fi t + ) + w(t) 0 t T, i = 1, 2 noncoherent BFSK and DPSK.
T
Consider a binary signaling scheme that involves the use
where is the unknown carrier phase, and w(t) is the sample of two orthogonal signals s1 (t) and s2 (t) of equal energy.
function of a white Gaussian noise process of zero mean and During the interval 0 t T , one of these two signals is
PSD N0 /2. sent over an imperfect AWGN channel that shifts the carrier
We furthermore assume uniform distribution of phase by an unknown amount.
We may thus express the received signal as
1 < (
f () = 2 g1 (t) + w(t) s1 (t) sent, 0 t T
x(t) =
0 otherwise g2 (t) + w(t) s2 (t) sent, 0 t T
22
Communication Systems Summary CONFIDENTIAL Raphael Seebacher
23
Communication Systems Summary CONFIDENTIAL Raphael Seebacher
25
Communication Systems Summary CONFIDENTIAL Raphael Seebacher
The performance of a digital communication system, in the = d2 (d) watts per steradian
presence of channel noise modeled as additive white Gaussian
noise, is defined by a formula having the shape of a water- We may express the radiation intensity as (, ), and so
fall curve. The probability of symbol error Pe is usually speak of a radiation-intensity pattern. The power radiated
plotted versus the bit energy-to-noise spectral density ratio inside an infinitesimal solid angle d is given by
Eb /N0 . The first design task is to specify two particular
(, )d d = sin d d
values of Eb /N0 :
26
Communication Systems Summary CONFIDENTIAL Raphael Seebacher
27
Communication Systems Summary CONFIDENTIAL Raphael Seebacher
It is often necessary to evaluate the noise figure of a cascade In North America, the band of radio frequencies assigned to
connection of two-port networks whose individual noise fig- the cellular system is 800-900 MHz. The subband 824-849
ures are known. It is assumed that the devices are matched. MHz is used to receive signals from the mobile units, and
The overall noise figure of a cascade connection is the subband 869-894 MHz is used to transmit signals to the
F2 1 F3 1 F4 1 mobile units. In Europe and elsewhere, the base-mobile and
F = F1 + + + + ... mobile-base subbands are reversed.
G1 G1 G2 G1 G2 G3
where F1 , F2 , . . . are the individual noise figures, and
G1 , G2 , . . . are the available power gains. 5.4.1 Propagation Effects
The overall equivalent noise temperature of the cascade con-
The major propagation problems are due to the fact that the
nection is (Friis formula)
antenna of a mobile unit may lie well below the surround-
T2 T3 T4 ing buildings. Simply put, there is no line-of-sight path
Te = T1 + + + + ...
G1 G1 G2 G1 G2 G3 to the base station. As a result of scattering, reflection and
where T1 , T2 , . . . are the equivalent noise temperatures. diffraction, we speak of a multipath phenomenon in that the
various incoming waves reach their destination from different
Def 5.13 Equivalent Noise Temperature of an directions and with differnet time delays.
Earth-Terminal Receiver The effect of the differential time delay is to introduce a rela-
tive phase shift between the two components of the received
Tmixer + TIF signal, with two extreme cases:
Te = Tantenna + TRF +
GRF
1. The relative phase shift is zero, in which case the two
5.4 Wireless Communications components add constructively.
A second type of multiuser radio communication system is 2. The relative phase shift is 180 degrees, in which case the
wireless communications, which is synonymous with mobile two components add destructively.
radio. In such a system the radio transmitter or receiver
is capable of being moved, regardless of whether it actually Due to motion of the receiver, there is a continuous change
moves or not. in the length of each propagation path. Hence, the rela-
The aim of a characterization of the mobile radio channel is tive phase shift between the two components of the received
to quantify two factors of primary concern: signal is a function of spatial location of the receiver. The
received amplitude (envelope) is no longer constant as was
1. median signal strength
the case in a static environment.
2. signal variability, which characterizes the fading nature of
the channel. Def 5.14 Signal Fading
An idealized model of the cellular radio system consists of There is constructive addition at some locations, and almost
an array of hexagonal cells with a base station located at the complete cancellation at some other locations. This phe-
center of each cell. A typical cell has a radius of 1 to 12 miles. nomenon is referred to as signal fading.
The function of the base stations is to act as an interface be-
tween mobile subscribers and the cellular radio system. The The net result is that the envelope of the received signal
base stations are themselves connected to a switching center varies with location in a complicated fashion.
by dedicated wirelines.
The mobile switching center has two important roles:
5.5 Binary Signaling over a Rayleigh Fad-
1. It acts as an interface between the cellular radio system
and the public switched telephone network. ing Channel
2. It performs overall supervision and control of the mobile Consider the transmission of binary data over a Rayleigh
communications. fading channel, for which the (low-pass) complex envelope of
the received signal is modeled as follows
The switching process, called a handover or handoff, is de-
signed to move a mobile subscriber from one base station x(t) = exp(j)s(t) + w(t)
to another during a call in a transparent fashion, without
interruption of service. where s(t) is the complex envelope of the transmitted (band-
This cellular concept relies on two essential features pass) signal, is a Rayleigh-distributed random variable de-
1. Frequency reuse refers to the use of radio channels on the scribing the attenuation in transmission, is a uniformly dis-
same carrier frequency to cover different areas, which are tributed random variable describing the phase-shift in trans-
physically sepearted from each other sufficiently ot ensure mission, and w(t) is a complex-valued white Gaussian noise
that co-channel interference is not objectionable. process. It is assumed that the channel is flat in both time
and frequency, so that we can estimate the phase-shift from
2. Cell splitting. When demand for service exceeds the num-
the received signal without error.
ber of channels allocated to a particular cell, cell splitting
Suppose then that coherent binary phase-shift keying is used
is used to handle the additional growth in traffic within
to do the data transmission. Under the condition that is
that particular cell. The new cells, which have a smaller
fixed or constant over a bit interval, we may express the av-
radius than the original cell, are called microcells.
erage probability of symbol error (i.e. bit error rate) due to
28
Communication Systems Summary CONFIDENTIAL Raphael Seebacher
the additive white Gaussian noise acting alone as follows: 5.5.1 Diversity Techniques
which defines the bit error rate for coherent binary phase-
shift keying (PSK) over a flat-flat Rayleigh fading channel.
Following a similar approach, we may derive the correspond-
ing bit error rates for coherent binary frequency-shift keying,
binary differential phase-shift keying and noncoherent binary
frequency-shift keying.
29
Communication Systems Summary CONFIDENTIAL Raphael Seebacher
Def 6.2 Information 2. H(S) = log2 K, if and only if pk = 1/K for all k (i.e., all
The amount of information is related to the inverse of the symbols in the alphabet S are equiprobable); this upper
probability of occurrence. We define the amount of informa- bound on entropy corresponds to the maximum uncer-
tion gained after observing the event S = sk , which occurs tainty.
with probability pk , as the logarithmic function
In discussing information-theoretic concepts, we often find
1 it useful to consider blocks rathar than individual symbols,
I(sk ) = log
pk with each block consisting of n successive source symbols.
with the following properties We may view each such block as being produced by an ex-
tended source with a source alphabet S n that has K n dis-
1. If we are absolutely certain of the outcome of an event, tinct blocks, where K is the number of distinct symbols in
even before it occurs, there is no information gained. the source alphabet S of the original source.
I(sk ) = 0 for pk = 1
H(S n ) = n H(S)
2. The occurence of an event S = sk either provides some or
no information, but never brings about a loss of informa-
tion. 6.2 Source-Coding Theorem
I(sk ) 0 0 pk 1
An important problem in communications is the efficient rep-
3. The less probable an event is, the more information we resentation of data generated by a discrete source. The pro-
gain when it occurs. cess by which this representation is accomplished is called
source encoding. The device that performs the representa-
I(sk ) > I(si ) pk < pi tion is called a source encoder.
31
Communication Systems Summary CONFIDENTIAL Raphael Seebacher
Theorem 6.1 Source-Coding Theorem where the factor 2 refers to the radix (number of symbols)
Given a discrete memoryless source of entropy H(S), the in the binary alphabet.
average code-word length L for any distortionless source en-
coding scheme is bounded as Def 6.8 Instantaneous Codes
Prefix codes are distinguished from other uniquely decodable
L H(S) codes by the fact that the end of a code word is always rec-
ognizable. For this reason, prefix codes are also referred to
Thus with Lmin = H(S) we may rewrite the efficiency of a as instantaneous codes.
source encoder in termy of the entropy H(S) as
Given a discrete memoryless source of entropy H(S), a prefix
H(S)
= code can be constructed with an average code-word length
L L, which is bounded as follows
A common characteristic of signals generated by physical For the case that pk = 2lk , the average code-word length is
sources is that, in their natural form, they contain a signifi- K1
X lk
cant amount of information that is redundant, the transmis- L=
2lk
sion of which is therefore wasteful of primary communica- k=0
tion resources. For efficient signal transmission, the redun- and the corresponding entropy of the source is
dant information should be removed from the signal prior to
K1 K1
transmission. This operation, with no loss of information, is X 1
lk
X lk
ordinarily performed on a signal in digital form, in which case H(S) = log 2 (2 ) =
2lk 2lk
k=0 k=0
we refer to it as data compaction or lossless data compres-
sion. The entropy of the source establishes the fundamental Hence we find that the prefix code is matched to the source
limit on the removal of redundancy from the data. in that L = H(S).
32
Communication Systems Summary CONFIDENTIAL Raphael Seebacher
Let Ln denote the average code-word length of the extended where p0 , . . . , pK1 are the source statistics, and lk is the
prefix code. For a uniquely decodable code, Ln is the small- length of the code word assigned to source symbol sk .
est possible.
It is usually found that when a combined symbol is moved
H(S n ) Ln < H(S n ) + 1
as high as possible, the resulting Huffman code has a signif-
nH(S) Ln < nH(S) + 1 icantly smaller variance 2 than when it is moved as low as
possible. On this basis, it is reasonable to choose the former
In the limit, as n approaches infinity, the lower and upper Huffman code over the latter.
bounds converge. We may therefore state that by making the
order n of an extended prefix source encoder large enough, Drawbacks of Huffman Coding
we can make the code faithfully represent the discrete mem-
oryless source S as closely as desired. The Huffman code requires knowledge of a probabilistic
However, the price we have to pay for decreasing the average model of the source; unfortunately, in practice, source
code-word length is increased decoding complexity, which is statistics are not always known a priori.
brought about by the high order of the extended prefix code.
In modeling text we find that storage requirements pre-
vent the Huffman code from capturing the higher-order
6.3.2 Huffman Coding
relationships between words and phrases, thereby compro-
An important class of prefix codes is known as Huffman mising the efficiency of the code.
codes. The basic idea behind Huffman coding is to assign
to each symbol of an alphabet a sequence of bits roughly 6.3.3 Lempel-Ziv Coding
equal in length to the amount of information conveyed by
the symbol in question. To overcome the practical limitations of the Huffman code,
The essence of the algorithm used to synthesize Huffman we may use the Lempel-Ziv algorithm, which is intrinsically
code is to replace the prescribed set of source statistics of a adaptive and simpler to implement than Huffman coding.
discrete memoryless source with a simpler one.
The Huffman encoding algorithm proceeds as follows: Basically, encoding in the Lempel-Ziv algorithm is accom-
plished by parsing the source data stream into segments that
1. The source symbols are listed in order of decreasing prob- are the shortest subsequences not encountered previously.
ability. The two source symbols of lowest probability are
assigned a 0 and a 1. This part of the step is referred to The last symbol of each subsequence in the code book is an
as a splitting stage. innovation symbol, which is so called in recognition of the
fact that its appendage to a particular subsequence distin-
2. These two symbols are regarded as being combined into a
guishes it from all previous subsequences stored in the code
new source symbol with probability equal to the sum of
book. Correspondingly, the last bit of each uniform block of
the two original probabilities. (The list of source symbols,
bits in the binary encoded representation of the data stream
and therefore source statistics, is thereby reduced in size
represents the innovation symbol for the particular subse-
by one.) The probability of the new symbol is placed in
quence under consideration.
the list in accordance with its value.
The remaining bits provide the equivalent binary represen-
3. The procedure is repeated until we are left with a final tation of the pointer to the root subsequence that matches
list of source statistics (symbols) of only two for which a the one in question except for the innovation symbol.
0 and a 1 are assigned.
In contrast to Huffman coding, the Lempel-Ziv algorithm
The code for each (original) source symbol is found by work- uses fixed-length codes to represent a variable number of
ing backward and tracing the sequence of 0s and 1s assigned source symbols; this feature makes the Lempel-Ziv code suit-
to that symbol as well as its successors. able for synchronous transmission.
It is noteworthy that the Huffman encoding process (i.e., the For a long time, Huffman coding was unchallenged as the al-
Huffman tree) is not unique: gorithm of choise for data compaction. However, the Lempel-
1. At each splitting stage in the construction of a Huffman Ziv algorithm has taken over almost completely and is now
code, there is arbitrariness in the way a 0 and a 1 are the standard algorithm for file compression.
assigned to the last two source symbols.
6.4 Discrete Memoryless Channels
2. Ambiguity arises when the probability of a combined sym-
bol is found to equal another probability in the list. A discrete memoryless channel is a statistical model with
an input X and an output Y that is a noisy version of X;
Note that the average code-word length remains the same. both X and Y are random variables. Every unit of time, the
channel accepts an input symbol X selected from an alpha-
Def 6.9 bet X and, in response, it emits an output symbol Y from
As a measure of the variability in code-word lengths of a an alphabet Y.
source code, we define the variance of the average code-word The channel is said to be discrete when both of the alpha-
length L over the ensemble of source symbols as bets have finite sizes. It is said to be memoryless when the
K1 current output symbol depends only on the current input
symbol and not any of the previous ones.
X
2 = pk (lk L)2
k=0
33
Communication Systems Summary CONFIDENTIAL Raphael Seebacher
The channel is described in terms of an input alphabet a measure of the prior uncertainty about X, how can we
measure the uncertainty about X after observing Y ?
X = {x0 , . . . , xJ1 }
J1
X 1
and output alphabet H(X |Y = yk ) = p(xj |yk ) log2
j=0
p(xj |yk )
Y = {y0 , . . . , yK1 }
The mean of entropy H(X |Y = yk ) over the output alphabet
and a set of transition probabilities Y is given by
p(yk |xj ) = P (Y = yk |X = xj ) j, k K1
X
H(X |Y) = H(X |Y = yk )p(yk )
with 0 p(yk |xj ) 1, j, k. k=0
K1
X J1
X 1
A convenient way of describing a discrete memoryless chan- = p(xj |yk )p(yk ) log2
p(xj |yk )
nel is to arrange the various transition probabilities of the k=0 j=0
channel in the form of a matrix as follows: K1
X J1
X
1
= p(xj , yk ) log2
p(xj |yk )
p(y0 |x0 ) p(y1 |x0 ) ... p(yK1 |x0 ) k=0 j=0
p(y0 |x1 ) p(y1 |x1 ) ... p(yK1 |x1 )
P = .. .. ..
The quantity H(X |Y) is called a conditional entropy. It
. . . represents the amount of uncertainty remaining about the
p(y0 |xJ1 ) p(y1 |xJ1 ) . . . p(yK1 |xJ1 ) channel input after the channel output has been observed.
The J-by-K matrix P is called the channel matrix, or tran- Another important quantity is called the mutual information
sition matrix. Not that each row of the channel matrix of the channel, and given by
corresponds to a fixed channel input, whereas each column
of the matrix corresponds to a fixed channel output. I(X ; Y) = H(X ) H(X |Y)
Suppose now that the inputs to a discrete memoryless chan- Similarly, we may write
nel are selected according to the probability distribution p(xj )
I(Y; X ) = H(Y) H(Y|X )
p(xj ) = P (X = xj ) j = 0, . . . , J 1
where H(Y) si the entropy of the channel output and H(Y|X
The joint probability distribution of the random variables X is the conditional entropy of the channel output given the
and Y is given by channel input.
p(xj , yk ) = P (X = xj , Y = yk )
= P (Y = yk |X = xj )P (X = xj ) 6.5.1 Properties of Mutual Information
= p(yk |xj )p(xj ) 1. The mutual information of a channel is symmetric; that
is
The marginal probability distribution of the output random
I(X ; Y) = I(Y; X )
variable Y is obtained by
J1
where the mutual information I(X ; Y) is a measure of the
p(yk ) =
X
p(yk |xj )p(xj ) k = 0, . . . , K 1 uncertainty about the channel input that is resolved by
j=0
observing the channel output, and the mutual information
I(Y|X ) is a measure of the uncertainty about the channel
The probabilities p(xj for j = 0, . . . , J 1, are known as the output that is resolved by sending the channel input.
a priori probabilities of the various input symbols.
2. The mutual information is always nonnegative; that is
Binary Symmetric Channel
I(X ; Y) 0
1p
x0 = 0 y0 = 0
p 3. The mutual information of a channel is related to the joint
entropy of the channel input and channel output by
p
x1 = 1 y1 = 1 I(X ; Y) = H(X ) + H(Y) H(X , Y)
1p
where the joint entropy H(X , Y) is defined by
6.5 Mutual Information K1
X J1
X
1
Given that we think of the channel output Y as a noisy ver- H(X , Y) = p(xj , yk ) log2
p(xj , yk )
sion of the channel input X, and that the entropy H(X ) is k=0 j=0
34
Communication Systems Summary CONFIDENTIAL Raphael Seebacher
6.6 Channel Capacity the destination from the source alphabet S and at the same
source rate of one symbol every Ts seconds.
Consider a discrete memoryless channel with input alphabet
We assume that the channel is capable of being used once
X , output alphabet Y, and transition probabilities p(yk |xj ),
every Tc seconds. Hence, the channel capacity per unit time
where j = 0, . . . , J 1 and k = 0, . . . , K 1.
is C/Tc bits per second, which represents the maximum rate
The mutual information of a channel depends not only on
of information transfer over the channel.
the channel but also on the way in which the channel is used
(i.e., the input probability distribution).
Theorem 6.2 Channel Coding Theorem
The channel coding theorem for a discrete memoryless chan-
Def 6.10 Channel Capacity
nel is state in two parts:
We define the channel capacity of a discrete memoryless
channel as the maximum mutual information I(X ; Y) in any 1. let a discrete memoryless source with an alphabet S have
single use of the channel (i.e., signaling interval), where the entropy H(S) and produce source symbols once every Ts
maximization is over all possible input probability distribu- seconds. Let a discrete memoryless channel have capacity
tions {p(xj )} on X . C and be used once every Tc seconds. Then, if
C = max I(X ; Y) H(S) C
{p(xj )}
Ts Tc
The channel capacity C is measured in bits per channel use,
or bits per transmission. there exists a coding scheme for which the source output
can be transmitted over the channel and be reconstructed
Note that the channel capacity C is a function only of the with an arbitrarily small probability of error. When the
transition probabilities p(yk |xj ), which define the channel. above equation is satisfied with the equality sign, the sys-
The calculation of C involves maximization subject to two tem is said to be signaling at the critical rate.
constraints:
2. Conversely, if
J1 H(S) C
X <
p(xj ) 0 j and p(xj ) = 1 Ts Tc
j=0
it is not possible to transmit information over the channel
and reconstruct it with an arbitrarily small probability of
6.7 Channel-Coding Theorem error.
The inevitable presence of noise in a channel causes discrep-
ancies (errors) between the output and input data sequences The theorem specifies the channel capacity C as a funda-
of a digital communication system. For many applications, mental limit on the rate at which the transmission of reliable
a probability of error equal to 106 or even lower is often error-free messages can take place over a discrete memoryless
a necessary requirement. To achieve such a high level of channel. It is important to note the following
performance, we resort to the use of channel coding.
Specifically, channel coding consists of mapping the incom- The theorem should be viewed as an existence proof.
ing data sequence into a channel input sequence, and inverse
mapping the channel output sequence into an output data It tells us that the probability of symbol error tends to
sequence in such a way that the overall effect of channel zero as the length of the code increases.
noise on the system is minimized. Not also that power and bandwidth constraints were hidden
in the discussion.
The channel encoder and channel decoder are both under
the designers control and should be designed to optimize
the overall reliability of the communication system. The 6.7.1 Application of the Channel Coding Theorem
approach taken is to introduce redundancy in the channel to Binary Symmetric Channels
encoder so as to reconstruct the original source sequence as Consider a discrete memoryless source that emits equally
accurately as possible. likely binary symbols once every Ts seconds. With the source
It suffices to confine our attention to block codes. In this class entropy equal to one bit per souce symbol, the information
of codes, the message sequence is subdivided into sequential rate of the source is (1/Ts ) bits per second. The source se-
blocks each k bits long, and each k-bit block is mapped into quence is applied to a channel encoder with code rate r. The
an n-bit block, where n > k. The number of redundant bits channel encoder produces a symbol once every Tc seconds.
added by the encoder to each transmitted block is n k bits. Hence, the encoded symbol transmission rate is (1/Tc ) sym-
The ratio bols per second.
k
r= <1 The channel encoder engages a binary symmetric channel
n
once every Tc seconds. Hence, the channel capacity per unit
is called the code rate.
time is (C/Tc ) bits per second, where C is determined by
Suppose then the discrete memoryless source has the source
the prescribed channel transition probability p. The channel
alphabet S and entropy H(S) bits per source symbol. We
coding theorem implies that if
assume that the source emits symbols once every Ts seconds.
hence, the average information rate of the source is H(S)/Ts 1 C
bits per second. The decoder delivers decoded symbols to
Ts Tc
35
Communication Systems Summary CONFIDENTIAL Raphael Seebacher
the probability of error can be made arbitrarily low by the 6.8.1 Mutual Information
use of a suitable channel encoding scheme. With
We define the mutual information between random variables
Tc X and Y as follows:
r= Z Z
Ts fX (x|y)
I(X; Y ) = fX,Y (x, y) log2 dx dy
fX (x)
we may restate the condition simply as
where fX,Y (x, y) is the joint probability density function of
rC X and Y , and fX (x|y) is the conditional probability density
function of X, given that Y = y.
That is, for r C there exists a code capable of achieving
We find that the mutual information has the properties
an arbitrarily low propability of error.
1. I(X; Y ) = I(Y ; X)
6.8 Differential Entropy and Mutual Infor- 2. I(X; Y ) 0
mation for Continuous Ensembles
3. I(X; Y ) = h(X) h(X|Y ) = h(Y ) h(Y |X)
The sources and channels considered thus far have involved
The parameter h(x) is the differential entropy of X; likewise
ensembles of random variables that are discrete in amplitude.
for h(Y ). The parameter h(X|Y ) is the conditional differen-
We extend some of these concepts to continuous random vari-
tial entropy of X, given Y ; it is defined by
ables and random vectors. Z Z
1
h(X|Y ) = fX,Y (x, y) log2 dx dy
Def 6.11 Differential Entropy fX (x|y)
Consider a continuous random variable X with probability
density function fX (x). 6.9 Information Capacity Theorem
We define h(X) as the differential entropy of X
Consider a zero-mean stationary process X(t) that is band-
Z limited to B hertz. Let Xk , k = 1, . . . , K, denote the con-
1
h(X) = fX (x) log2 dx tinuous random variables obtainted by uniform sampling of
fX (x)
the process X(t) at the Nyquist rate of 2B samples per sec-
When we have a continuous random vector X consisting of ond. These samples are transmitted in T seconds over a
n random variables X1 , . . . , Xn , we define the differential en- noisy channel, also band-limited to B hertz. The number of
tropy of X as the n-fold integral samples K is given by
Z
1
K = 2BT
h(X) = fX (x) log2 dx
fX (x) We refer to Xk as a sample of the transmitted signal. The
channel output is perturbed by additive white Gaussian noise
where fX (x) is the joint probability density function of X. (AWGN) of zero mean and power spectral density N0 /2. The
noise is band-limited to B hertz. Let the continuous random
Consider a random variable X uniformly distributed over the variables Yk , k = 1, . . . , K denote samples of the received
interval (0, a). We then get signal
Yk = Xk + Nk k = 1, . . . , K
h(X) = log a
The noise sample Nk is Gaussian with zero mean and vari-
Note that the differential entropy of a continuous random ance given by
variable can be negative. 2 = N0 B
Consider two random variables X and Y that have the same We assume that the samples Yk are statistically independent
mean and the same variance 2 . Furthermore X is Gaus- for all k. Such a channel is called a discrete-time, memory-
sian distributed. For the differential entropy of the Gaussian less Gaussian channel.
random variable X we get
1 To assign a cost to each channel input, it is reasonable to
h(X) = log2 (2e 2 define
2
E[Xk 2 ] = p k = 1, . . . , K
and finally as the transmitter is typically limited in power; with P de-
( noting the average transmitted power.
X Gaussian random variable
h(Y ) h(X)
Y another random variable Let I(Xk ; Yk ) denote the mutual information between Xk
and Yk . We may then define the information capacity of the
For a finite variance 2 , the Gaussian rnadom variable has channel as
the largest differential entropy attainable by any random
C = max I(Xk ; Yk ) : E[Xk 2 ] = P
variable. fXk (x)
The entropy of a Gaussian random variable X is uniquely where the maximization is performed with respect to fXk (x),
determined by the variance of X (i.e., it is independent of the probability density function of Xk .
the mean of X).
36
Communication Systems Summary CONFIDENTIAL Raphael Seebacher
By using I(Xk ; Yk ) = h(Yk ) h(Nk ) we may reformulate the A plot of bandwidth efficiency Rb /B versus Eb /N0 is called
channel capacity as the bandwidth-efficiency diagram.
C = I(Xk ; Yk ) : Xk Gaussian E[Xk 2 ] = P
The variance of sample Yk of the received signal equals
P + 2 . Hence
1
h(Yk ) = log2 2e(P + 2 )
2
The variance of the noise sample Nk equals 2 .
1
h(Nk ) = log2 (2e 2 )
2
For the desired channel capacity C we then get
1 P
C = log2 1 + 2 bits per transmission
2
With the cannel used K times for the transmission of K
samples of the process X(t) in T seconds, we find that the
information capacity per unit time is (K/T ) times the result
given above. The number K equals 2BT .
37
Communication Systems Summary CONFIDENTIAL Raphael Seebacher
7 Error-Control Coding
For a fixed Eb /N0 , the only practical option available for the mathematical structure of the code. Accordingly, these
changing data quality from problematic to acceptable is to n k bits are referred to as parity bits.
use error-control coding. Block codes in which the message bits are transmittd in
Error control for data integrity may be exercised by means of unaltered form are called systematic codes. For applications
forward error correction (FEC). The discrete source gener- requiring both error detection and error correction, the use
ates information in the form of binary symbols. The channel of systematic block codes simplifies implementation of the
encoder in the transmitter accepts message bits and adds re- decoder.
dundancy according to a prescribed rule, thereby producing
encoded data at a higher bit rate. The channel decoder in Let m0 , . . . , mk1 constitute a block of k arbitrary message
the receiver exploits the redundancy to decide which message bits. Thus we have 2k distinct message blocks. Let this se-
bits were actually transmitted. quence of message bits be applied to a linear block encoder,
For a fixed modulation scheme, the adition of redundancy in producing an n-bit code word whose elements are denoted
the coded messages implies the need for increased transmis- by c0 , . . . , cn1 . Let b0 , . . . , bnk1 denote the (n k) parity
sion bandwidth. Moreover, the use of error-control coding bits in the code word.
adds complexity to the system. We may then write
(
bi i = 0, . . . , n k 1
Historically, the different error-correcting codes have been ci =
classified into block codes and convolutional codes. The dis- mi+kn i = n k, . . . , n 1
tinguishing feature for this particular classification is the The (n k) parity bits are linear sums of the k message bits,
presence or absence of memory in the encoders of the two as shown by the generalized relation
codes.
k1
X
bi = pji mj
7.1 Discrete-Memoryless Channels j=0
39
Communication Systems Summary CONFIDENTIAL Raphael Seebacher
40
Communication Systems Summary CONFIDENTIAL Raphael Seebacher
2. An error pattern e2 is picked and placed under c1 , and a 7.3.1 Generator Polynomial
second row is formed by adding e2 to each of the remain-
Let g(X) be a polynomial of degree n k that is a factor of
ing code vectors in the first row; it is important that the
X n + 1; as such, g(X) is the polynomial of least degree in the
error pattern chosen as the first element in a row not have
code. In general, g(X) may be expanded as follows:
previously appeared in the standard array.
nk1
3. Step 2 is repeated until all the possible error patterns have
X
g(X) = 1 + gi X i + X nk
been accounted for. i=1
For a given channel, the probability of decoding error is min- where the coefficient gi is equal to 0 or 1.
imized when the most likely error patterns (i.e., those with The polynomial g(X) is called the generator polynomial of a
the largest probability of occurrence) are chosen as the coset cyclic code. A cyclic code is uniquely determined by the gen-
leaders. In the case of a binary symmetric channel, the erator polynomial g(X) in that each code polynomial in the
smaller the Hamming weight of an error pattern the more code can be expressed in the form of a polynomial product
likely it is to occur. as follows:
We may now describe a decoding procedure for a linear block c(X) = a(X)g(X)
code:
where a(X) is a polynomial in X with degree k 1.
1. For the received vector r, compute the syndrome s =
rHT . Suppose we are given the generator polynomial g(X)
and the requirement is to encode the message sequence
2. Within the coset characterized by the syndrome s, iden-
(m0 , . . . , mk1 ) into an (n, k) systematic cyclic code.
tify the coset leader (i.e., the error pattern with the largest
Let the message polynomial be defined by
probability of occurrence); call it e0 .
m(X) = m0 + m1 X + . . . + mk1 X k1
3. Compute the code vector
and let
c = r + e0
b(X) = b0 + b1 X + . . . + bnk1 X nk1
as the decoded version of the received vector r.
We want the code polynomial to be in the form
This procedure is called syndrome decoding.
c(X) = b(X) + X nk m(X) = a(X)g(X)
7.2.4 Dual Code
In light of modulo-2 addition, we may write
Given a linear block code, we may define its dual as follows: X nk m(X) b(X)
T = a(X) +
GH = 0 g(X) g(X)
T We may now summarize the steps involved in the encoding
where H is the transpose of the parity-check matrix of the
code, and 0 is a new zero matrix. This equation suggests procedure for an (n, k) cyclic code assured of a systematic
that every (n, k) linear block code with generator matrix G structure.
and parity-check matrix H has a dual code with parameters 1. Multiply the message polynomial m(X) by X nk .
(n, n k), generator matrix H and parity-check matrix G.
2. Divide X nk m(X) by the generator polynomial g(X), ob-
taining the remainder b(X).
7.3 Cyclic Codes
3. Add b(X) to X nk m(X), obtaining the code polynomial
Cyclic codes form a subclass of linear block codes. An ad-
c(X).
vantage of cyclic codes over most other types of codes is that
they are easy to encode.
A binary code is said to be a cyclic code if it exhibits two 7.3.2 Parity-Check Polynomial
fundamental properties: An (n, k) cyclic code is uniquely specified by its generator
1. Linearity property: The sum of any two code words in the polynomial g(X) of order (nk). Such a code is also uniquely
code is also a code word. specified by another polynomial of degree k, which is called
the parity-check polynomial, defined by
2. Cyclic property: Any cyclic shift of a code word in the
k1
code is also a code word. X
h(X) = 1 + hi X i + X k
We define the code polynomial i=1
2
c(X) = c0 + c1 X + c2 X + . . . + cn1 X n1 where the coefficients hi are 0 or 1.
We find that the matrix relation HGT = 0 for linear block
where X is an indeterminante. Naturally, for binary codes, codes corresponds to the relationship
the coefficients ci are 1s and 0s. Each power of X in the
polynomial c(X) represents a one-bit shift in time. g(X)h(X) mod (X n + 1) = 0
Accordingly, we may state that the generator polynomial
If c(X) is a code polynomial, then the polynomial g(X) and the parity-check polynomial h(X) are factors of
c(i) (X) = X i c(X) mod (X n + 1) the polynomial X n + 1, as shown by
41
Communication Systems Summary CONFIDENTIAL Raphael Seebacher
42
Communication Systems Summary CONFIDENTIAL Raphael Seebacher
The probability of generating k frames in a given time inter- 8.2.2 Nonpersistent CSMA
val D is:
(gD)k egD Gk eG If the channel is busy, the station waits a random
P r[k|D] = = amount of time and then checks the channel again.
k! k!
G = g D represents the average number of generated frames
per frame duration. Throughput:
43
Communication Systems Summary CONFIDENTIAL Raphael Seebacher
Bipolar return-to-zero (BRZ) signaling It follows therefore that the Fourier transform G(f ) of g(t)
is given by
Positive and negative pulses of equal amplitude Aare used G(f ) = j sgn(f )G(f )
alternately for symbol 1, with each pulse having half-symbol
width; no pulse is always used for symbol 0.
+ Power spectrum has no DC component.
+ Relatively insifgnificant low-frequency components
if 0 and 1 are equiprobable.
Also called alternate mark inversion (AMI) signaling.
44
Communication Systems Summary CONFIDENTIAL Raphael Seebacher
where G(0) is the value of G(f ) at frequency f = 0. In general, g(t) is a complex-valued quantity
For a given signal g(t) we may determine its pre-envelope g(t) = gI (t) + jgQ (t)
g+ (t) in one of two equivalent ways:
where gI (t) and gQ (t) are both real-valued low-pass func-
1. We determine the Hilbert ransform g(t) of the signal g(t), tions; their low-pass property is inherited from the complex
and then use g+ (t) = g(t) + jg(t) to compute the pre envelope g(t).
envelope. The original band-pass signal may then be expressed in the
canonical, or standard, form:
2. We determine the Fourier transform G(f ) of the signal
g(t), and use G+ (f ) = G(f ) + sgn(f )G(f ) to determine g(t) = gI (t) cos(2fc t) gQ (t) sin(2fc t)
G+ (f ) and then evaluate the inverse Fourier transform of
G+ (f ) to obtain We refer to gI (t) as the in-phase component of the band-pass
Z signal g(t) and to gQ (t) as the quadrature component of the
g+ (t) = 2 G(f ) exp(j2f t) df signal.
0 Since both gI (t) and gQ (t) are low-pass signals limited to
the band W f W , they may be derived from the
Symmetrically we may define the pre-envelope for negative
band-pass signal g(t) (cf Figure below).
frequencies as
g (t) = g(t) jg(t)
The complex envelope g(t) can be expressed in the polar
The two pre-envelopes g+ (t) and g (t) are simply the com- form
plex conjugate of each other, as shown by g(t) = a(t) exp j(t)
where a(t) and (t) are both real-valued low-pass functions.
g (t) = g+ (t)
Based on this polar representation, the original band-pass
C.0.5 Canonical Representation: Band-Pass Signal signal g(t) is defined by
45
Index
A Linear Functional . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3
linear receiver . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5
Additive White Gaussian Noise Channel . . . . . . . . . . . . . . . . 9 Link Margin . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 26
Antipodal Signals . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14 Log-Likelihood Function . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
Autocorrelation Function . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1 AWGN Channel . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
Properties . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
Autocovariance Function . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1 M
Average Probability of Symbol Error . . . . . . . . . . . . . . . . . . . . 6
Matched Filter . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5
Binary Symmetric Channel . . . . . . . . . . . . . . . . . . . . . . . . . . 6
Matched Filters
AWGN Channel . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9
Properties . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5
B Mean . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
Modified Duobinary Signaling . . . . . . . . . . . . . . . . . . . . . . . . . . . 8
Baseband M-ary PAM Transmission . . . . . . . . . . . . . . . . . . . . 8
Baseband Pulse Transmission . . . . . . . . . . . . . . . . . . . . . . . . . . . 5 N
Binary PCM . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5 Noise . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3
Binary Symmetric Channel . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6 White . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3
Bit Duration . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5 Nyquist Bandwidth . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
Bit Rate . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7 Nyquist Rate . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
Nyquists Criterion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6
C
Distortionless Baseband Transmission . . . . . . . . . . . . . . . . 7
Central Limit Theorem . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3
P
Class I Partial Response . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
Class IV Partial Response . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8 Peak Pulse Signal-to-Noise Ratio . . . . . . . . . . . . . . . . . . . . . . . 5
Complementary Error Function . . . . . . . . . . . . . . . . . . . . . . . . . 6 Power Spectral Density . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2
Correlation Matrix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2 Properties . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2
Correlative-Level Coding . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7 Pulse Code Modulation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5
Cross-Correlation Function . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
Q
D
quadrature component . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5
Discrete Memoryless Source . . . . . . . . . . . . . . . . . . . . . . . . . . . 31
R
Discrete Pulse-Amplitude Modulation . . . . . . . . . . . . . . . . . . . 6
Duobinary Signaling . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7 Raised Cosine Spectrum . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
Full-Cosine Rolloff . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
E
Rolloff Factor . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
Einstein-Wiener-Khintchine relations . . . . . . . . . . . . . . . . . . . 2 Transmission Bandwidth . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
Entropy . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31 Random Process . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
Equivalent Noise Temperature . . . . . . . . . . . . . . . . . . . . . . . . . . 3 S
Ergodic Process . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2
Sample Point . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
G Sample Space . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
Signal Energy-to-Noise Spectral Density Ratio . . . . . . . . . . 5
Gaussian Distribution . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3
Source-Coding Theorem . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32
Gaussian Process . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3
Strict Stationarity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
Properties . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3
Geometric Representation of Signals . . . . . . . . . . . . . . . . . . . . 9 T
I Theorem of Irrelevance . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
Threshold . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5
Ideal Nyquist Channel . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7 Optimum . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6
in-phase component . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5 Transmitted Signal Energy per Bit . . . . . . . . . . . . . . . . . . . . . . 6
Information . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31
Intersymbol Interference . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6 V
J Variable-Length Code . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32
L Weakly Stationary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
White Noise . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3
Likelihood Function . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10 Wide-Sense Stationary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
46