Title: EC1351 DIGITAL COMMUNICATION
1EC1351 DIGITAL COMMUNICATION UNIT I Pulse
Modulation
5
2Figure (a) Spectrum of a signal. (b) Spectrum of
an undersampled version of the signal exhibiting
the aliasing phenomenon.
Figure (a) Anti-alias filtered spectrum of an
information-bearing signal. (b) Spectrum of
instantaneously sampled version of the signal,
assuming the use of a sampling rate greater than
the Nyquist rate. (c) Magnitude response of
reconstruction filter.
6
33.3 Pulse-Amplitude Modulation-Amplitude of
carrier pulse train is varied in accordance with
m(t).
8
4 Other Forms of Pulse Modulation
- In pulse width modulation (PWM), the width of
each pulse is made directly proportional to the
amplitude of the information signal. - In pulse position modulation, constant-width
pulses are used, and the position or time of
occurrence of each pulse from some reference time
is made directly proportional to the amplitude of
the information signal.
11
5Quantization Process
12
6Illustration of the quantization process
Figure 3.10 Two types of quantization (a)
midtread and (b) midrise.
Pulse Code Modulation (PCM)
- Pulse code modulation (PCM) is produced by
analog-to-digital conversion process. - As in the case of other pulse modulation
techniques, the rate at which samples are taken
and encoded must conform to the Nyquist sampling
rate. - The sampling rate must be greater than, twice the
highest frequency in the analog signal, - fs gt 2fA(max)
13
7- Time-Division Multiplexing
-
-
- Digital Multiplexers
Advantages of PCM 1. Robustness to noise and
interference 2. Efficient regeneration
3. Efficient SNR and bandwidth trade-off 4.
Uniform format 5. Ease add and drop 6.
Secure
83.12 Delta Modulation (DM) (Simplicity)
9The modulator consists of a comparator, a
quantizer, and an accumulator The output
of the accumulator is
Two types of quantization errors Slope overload
distortion and granular noise
10- 3.13 Linear Prediction (to reduce the sampling
rate) - Consider a finite-duration impulse response (FIR)
- discrete-time filter which consists of three
blocks - 1. Set of p ( p prediction order) unit-delay
elements (z-1) - 2. Set of multipliers with coefficients w1,w2,wp
- 3. Set of adders ( ? )
-
113.14 Differential Pulse-Code Modulation
(DPCM) Usually PCM has the sampling rate higher
than the Nyquist rate .The encode signal contains
redundant information. DPCM can efficiently
remove this redundancy.
Figure 3.28 DPCM system. (a) Transmitter. (b)
Receiver.
12- 3.15 Adaptive Differential Pulse-Code Modulation
(ADPCM) - Need for coding speech at low bit rates , we
have two aims in mind - 1. Remove redundancies from the speech
signal as far as possible. - 2. Assign the available bits in a
perceptually efficient manner. -
-
-
Figure 3.29 Adaptive quantization with backward
estimation (AQB).
Figure 3.30 Adaptive prediction with backward
estimation (APB).
13UNIT II BASEBAND PULSE TRANSMISSION
Correlative-Level Coding
- Correlative-level coding (partial response
signaling) - adding ISI to the transmitted signal in a
controlled manner - Since ISI introduced into the transmitted signal
is known, its effect can be interpreted at the
receiver - A practical method of achieving the theoretical
maximum signaling rate of 2W symbol per second in
a bandwidth of W Hertz - Using realizable and perturbation-tolerant filters
14Correlative-Level Coding
Duobinary Signaling
- Dou doubling of the transmission capacity of a
straight binary system - Binary input sequence bk uncorrelated binary
symbol 1, 0
15Correlative-Level Coding
Duobinary Signaling
- The tails of hI(t) decay as 1/t2, which is a
faster rate of decay than 1/t encountered in
the ideal Nyquist channel. - Let represent the estimate of the original
pulse ak as conceived by the receiver at time
tkTb - Decision feedback technique of using a stored
estimate of the previous symbol - Propagate drawback, once error are made, they
tend to propagate through the output - Precoding practical means of avoiding the error
propagation phenomenon before the duobinary coding
16Correlative-Level Coding
Duobinary Signaling
- dk is applied to a pulse-amplitude modulator,
producing a corresponding two-level sequence of
short pulse ak, where 1 or 1 as before
- ck1 random guess in favor of symbol 1 or 0
17Correlative-Level Coding
Modified Duobinary Signaling
- Nonzero at the origin undesirable
- Subtracting amplitude-modulated pulses spaced 2Tb
second
18Baseband M-ary PAM Trans.
- Produce one of M possible amplitude level
- T symbol duration
- 1/T signaling rate, symbol per second, bauds
- Equal to log2M bit per second
- Tb bit duration of equivalent binary PAM
- To realize the same average probability of symbol
error, transmitted power must be increased by a
factor of M2/log2M compared to binary PAM
19Adaptive Equalizer
- Adaptive equalization
- Adjust itself by operating on the the input
signal - Training sequence
- Precall equalization
- Channel changes little during an average data
call - Prechannel equalization
- Require the feedback channel
- Postchannel equalization
- synchronous
- Tap spacing is the same as the symbol duration of
transmitted signal
- Training mode
- Known sequence is transmitted and synchorunized
version is generated in the receiver - Use the training sequence, so called
pseudo-noise(PN) sequence - Decision-directed mode
- After training sequence is completed
- Track relatively slow variation in channel
characteristic - Large ? fast tracking, excess mean square error
20Eye Pattern
- Experimental tool for such an evaluation in an
insightful manner - Synchronized superposition of all the signal of
interest viewed within a particular signaling
interval - Eye opening interior region of the eye pattern
- In the case of an M-ary system, the eye pattern
contains (M-1) eye opening, where M is the number
of discreteamplitude levels
21III.PASS BAND DATA TRANSMISSION
22(No Transcript)
23Binary amplitude shift keying, Bandwidth
- d 0 ? related to the condition of the line
B (1d) x S (1d) x N x 1/r
24Implementation of binary ASK
25Frequency Shift Keying
- One frequency encodes a 0 while another frequency
encodes a 1 (a form of frequency modulation) - Represent each logical value with another
frequency (like FM)
26DBPSK, QPSK
- Differential BPSK
- 0 same phase as last signal element
- 1 180º shift from last signal element
- Four Level QPSK
27Concept of a constellation diagram
28MPSK
- Using multiple phase angles with each angle
having more than one amplitude, multiple signals
elements can be achieved - D modulation rate, baud
- R data rate, bps
- M number of different signal elements 2L
- L number of bits per signal element
29Generation and Detection of Coherent BPSK
Figure 6.26 Block diagrams for (a) binary FSK
transmitter and (b) coherent binary FSK receiver.
30(No Transcript)
31IV.Error Control Techniques
- Error detection in a block of data
- Can then request a retransmission, known as
automatic repeat request (ARQ) for sensitive data - Appropriate for
- Low delay channels
- Channels with a return path
- Not appropriate for delay sensitive data, e.g.,
real time speech and data - Forward Error Correction (FEC)
- Coding designed so that errors can be corrected
at the receiver - Appropriate for delay sensitive and one-way
transmission (e.g., broadcast TV) of data - Two main types, namely block codes and
convolutional codes. We will only look at block
codes - BLOCK CODE
- We will consider only binary data
- Data is grouped into blocks of length k bits
(dataword) - Each dataword is coded into blocks of length n
bits (codeword), where in general ngtk - This is known as an (n,k) block code
- A vector notation is used for the datawords and
codewords, - Dataword d (d1 d2.dk)
- Codeword c (c1 c2..cn)
- The redundancy introduced by the code is
quantified by the code rate,
32Hamming Distance
- The maximum number of detectable errors is
- That is the maximum number of correctable errors
is given by, - where dmin is the minimum Hamming distance
between 2 codewords and means the smallest
integer
33Linear Block Codes
- As seen from the second Parity Code example, it
is possible to use a table to hold all the
codewords for a code and to look-up the
appropriate codeword based on the supplied
dataword - Alternatively, it is possible to create codewords
by addition of other codewords. This has the
advantage that there is now no longer the need to
held every possible codeword in the table.
- If there are k data bits, all that is required is
to hold k linearly independent codewords, i.e., a
set of k codewords none of which can be produced
by linear combinations of 2 or more codewords in
the set. - The easiest way to find k linearly independent
codewords is to choose those which have 1 in
just one of the first k positions and 0 in the
other k-1 of the first k positions. - For example for a (7,4) code, only four codewords
are required, e.g., - So, to obtain the codeword for dataword 1011, the
first, third and fourth codewords in the list are
added together, giving 1011010 - This process will now be described in more detail
34Linear Block Codes
- An (n,k) block code has code vectors
- d(d1 d2.dk) and
- c(c1 c2..cn)
- The block coding process can be written as cdG
- where G is the Generator Matrix
35Parity Check Matrix
- So H is used to check if a codeword is valid,
- So the dimension of H is n-k ( R) and all
vectors in the null space are orthogonal to all
the vectors of the code - Since the rows of H, namely the vectors bi are
members of the null space they are orthogonal to
any code vector - So a vector y is a codeword only if yHT0
- Note that a linear block code can be specified by
either G or H
R n - k
- The rows of H, namely, bi, are chosen to be
orthogonal to rows of G, namely ai - Consequently the dot product of any valid
codeword with any bi is zero
36Error Syndrome
- For error correcting codes we need a method to
compute the required correction - To do this we use the Error Syndrome, s of a
received codeword, cr - s crHT
- If cr is corrupted by the addition of an error
vector, e, then - cr c e
- and
- s (c e) HT cHT eHT
- s 0 eHT
- Syndrome depends only on the error
- For systematic linear block codes, H is
constructed as follows, - G I P and so H -PT I
- where I is the kk identity for G and the RR
identity for H - Example, (7,4) code, dmin 3
37Hamming Codes
- Double errors will always result in wrong bit
being corrected, since - A double error is the sum of 2 single errors
- The resulting syndrome will be the sum of the
corresponding 2 single error syndromes - This syndrome will correspond with a third single
bit error - Consequently the corrected codeword will now
contain 3 bit errors, i.e., the original double
bit error plus the incorrectly corrected bit! - Convolutional Code Introduction
- Convolutional codes map information to code bits
sequentially by convolving a sequence of
information bits with generator sequences - A convolutional encoder encodes K information
bits to NgtK code bits at one time step - Convolutional codes can be regarded as block
codes for which the encoder has a certain
structure such that we can express the encoding
operation as convolution
38Example
- Convolutional encoder, k 1, n 2, L2
- Convolutional encoder is a finite state machine
(FSM) processing information bits in a serial
manner - Thus the generated code is a function of input
and the state of the FSM - In this (n,k,L) (2,1,2) encoder each message
bit influences a span of C n(L1)6 successive
output bits constraint length C - Thus, for generation of n-bit output, we require
n shift registers in k 1 convolutional encoders
39Representing convolutional codes Code tree
(n,k,L) (2,1,2) encoder
This tells how one input bit is transformed into
two output bits (initially register is all zero)
40Representing convolutional codes compactly code
trellis and state diagram
Input state 1 indicated by dashed line
State diagram
Code trellis
Shift register states
41The maximum likelihood path
Smaller accumulated metric selected
After register length L13 branch pattern begins
to repeat
(Branch Hamming distances in parenthesis)
First depth with two entries to the node
The decoded ML code sequence is 11 10 10 11 00 00
00 whose Hamming distance to the received
sequence is 4 and the respective decoded
sequence is 1 1 0 0 0 0 0 (why?). Note that this
is the minimum distance path. (Black circles
denote the deleted branches, dashed lines '1'
was applied)
42Turbo Codes
- Backgound
- Turbo codes were proposed by Berrou and Glavieux
in the 1993 International Conference in
Communications. - Performance within 0.5 dB of the channel capacity
limit for BPSK was demonstrated. - Features of turbo codes
- Parallel concatenated coding
- Recursive convolutional encoders
- Pseudo-random interleaving
- Iterative decoding
- Turbo code advantages
- Remarkable power efficiency in AWGN and
flat-fading channels for moderately low BER. - Deign tradeoffs suitable for delivery of
multimedia services. - Turbo code disadvantages
- Long latency.
- Poor performance at very low BER.
- Because turbo codes operate at very low SNR,
channel estimation and tracking is a critical
issue. - The principle of iterative or turbo processing
can be applied to other problems. - Turbo-multiuser detection can improve performance
of coded multiple-access systems.
43UNIT V Spread Spectrum MODULATION
- Analog or digital data
- Analog signal
- Spread data over wide bandwidth
- Makes jamming and interception harder
- Frequency hoping
- Signal broadcast over seemingly random series of
frequencies - Direct Sequence
- Each bit is represented by multiple bits in
transmitted signal - Chipping code
- GAINS
- Immunity from various noise and multipath
distortion - Including jamming
- Can hide/encrypt signals
- Only receiver who knows spreading code can
retrieve signal - Several users can share same higher bandwidth
with little interference - Cellular telephones
- Code division multiplexing (CDM)
- Code division multiple access (CDMA)
- Input fed into channel encoder
44Pseudorandom Numbers
- Generated by algorithm using initial seed
- Deterministic algorithm
- Not actually random
- If algorithm good, results pass reasonable tests
of randomness - Need to know algorithm and seed to predict
sequence - Frequency Hopping Spread Spectrum (FHSS)
- Signal broadcast over seemingly random series of
frequencies - Receiver hops between frequencies in sync with
transmitter - Eavesdroppers hear unintelligible blips
- Jamming on one frequency affects only a few bits
- SLOW AND FAST FREQ HOP
- Frequency shifted every Tc seconds
- Duration of signal element is Ts seconds
- Slow FHSS has Tc ? Ts
- Fast FHSS has Tc lt Ts
- Generally fast FHSS gives improved performance in
noise (or jamming) - Frequency Hopping Example
45Direct Sequence Spread Spectrum (DSSS)
- Each bit represented by multiple bits using
spreading code - Spreading code spreads signal across wider
frequency band - In proportion to number of bits used
- 10 bit spreading code spreads signal across 10
times bandwidth of 1 bit code - One method
- Combine input with spreading code using XOR
- Input bit 1 inverts spreading code bit
- Input zero bit doesnt alter spreading code bit
- Data rate equal to original spreading code
- Performance similar to FHSS
46Direct Sequence Spread Spectrum Transmitter