DPCM encodes signal samples by quantizing and encoding the difference between the current sample and a prediction of that sample based on previous samples, rather than quantizing each sample directly. This exploits the high correlation between adjacent samples to remove redundancy. The prediction is generated using a linear predictor with coefficients chosen to minimize the mean-square error between the actual and predicted samples. Adaptive DPCM updates the predictor coefficients over time to account for non-stationarities in speech and other signals. Delta modulation is a simplified version of DPCM that uses a 1-bit quantizer, providing a staircase approximation of oversampled signals.
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0 ratings0% found this document useful (0 votes)
88 views13 pages
DPCM
DPCM encodes signal samples by quantizing and encoding the difference between the current sample and a prediction of that sample based on previous samples, rather than quantizing each sample directly. This exploits the high correlation between adjacent samples to remove redundancy. The prediction is generated using a linear predictor with coefficients chosen to minimize the mean-square error between the actual and predicted samples. Adaptive DPCM updates the predictor coefficients over time to account for non-stationarities in speech and other signals. Delta modulation is a simplified version of DPCM that uses a 1-bit quantizer, providing a staircase approximation of oversampled signals.
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 13
Differential Pulse Code Modulation (DPCM)
• In the PCM system, each sample is quantized and encoded independently.
Previous sample values have no effect on the quantization of the new samples. • Sampling a message signal (random process) at the rate Nyquist rate results in • Sampled values that are highly correlated • This means that the previous samples give some information about the next sample • For instance, if the previous sample values were small, it is highly probable that the next sample value will also be small as well • Encoding these highly correlated samples results in a lot of redundant bits • Correlation information can be utilized to remove the redundancies • Better waveform coding (lesser number of bits) DPCM DPCM (contd.) • Consider the simplest form of DPCM) • Take the difference between two adjacent samples (current and previous samples) and quantized and encode it. 𝑒 𝑛 = 𝑚 𝑛 − 𝑚[𝑛 − 1] (1) • Because two adjacent samples are highly correlated, their difference has small variations (dynamic range) as compared to original sample values Since, step size = DR/L (quantization level) For the same quality (), small DR requires lesser L Less number of bits per sample At the receiver 𝑚 𝑛 = 𝑚 𝑛 − 1 + 𝑒[𝑛] (2) DPCM (contd.) • Since the samples are correlated, therefore current sample 𝑚[𝑛] can be predicted from the previous samples • Quantized and encode the difference of the sample 𝑚[𝑛] and its predicted value 𝑚[𝑛] ෝ 𝑒 𝑛 = 𝑚 𝑛 − 𝑚[𝑛] ෝ (3) • Linear predictor of the form 𝑝 𝑚 σ ෝ 𝑛 = 𝑘=1 𝜔𝑘 𝑚[𝑛 − 𝑘] Convolution sum (4) Where 𝜔𝑘 is predictor coefficients (weights) and is chosen such that the mean square error between the sample and its predicted value is minimized. 𝐸 𝑒 2 [𝑛] = E 𝑚 𝑛 − 𝑚[𝑛] ෝ 2 = 𝐸 𝑚2 [𝑛] − 2𝐸 𝑚 𝑛 𝑚 ෝ 2 [𝑛] (5) ෝ 𝑛 +𝐸 𝑚 Let the sample value 𝑚 𝑛 be the sample function of stationary random process M(t) with zero mean DPCM (contd.) Let us denote 𝑅𝑀 𝑘 = autocorrelation of the process M(t) for a lag of k 𝑅𝑀 𝑘 = 𝐸 𝑚 𝑛 𝑚[𝑛 − 𝑘] (6) 𝑅𝑀 0 = 𝐸 𝑚2 [𝑛] = variance of the process M(t) = average signal power Putting (4) in (5), we have 𝑝 𝑝 𝑝
𝑘=1 𝑗=1 𝑘=1 𝑝 𝑝 = 𝑅𝑀 0 − 2 σ𝑘=1 𝜔𝑘 𝑅𝑀 𝑘 + σ𝑘=1 𝜔𝑗 𝜔𝑘 𝑅𝑀 𝑗 − 𝑘 j= 1,2 , 3, ……… (7) Differentiating (7) w.r.t 𝜔𝑘 and setting equal to zero, we get σ𝑝𝑗=1 𝜔𝑗 𝑅𝑀 𝑗 − 𝑘 = 𝑅𝑀 𝑘 = 𝑅𝑀 −𝑘 1 k p (8) • Solving the above set of equations (usually referred to as Wiener-Hopf equations for linear prediction), one can find the optimal set of predictor coefficients 𝜔𝑘 . DPCM (contd.) • Eqn. (8) can be written in matrix form as 𝑹𝑋 𝝎𝑜 = 𝒓𝑋 (9) where 𝝎𝑜 = 𝑝1 optimum coefficient vector = 𝜔1 , 𝜔2 , … … … … . , 𝜔𝑝 𝒓𝑋 = 𝑝1 autocorrelation vector = 𝑅𝑋 1 , 𝑅𝑋 2 , … … … … . , 𝑅𝑋 [𝑝] 𝑹𝑋 = 𝑝𝑝 autocorrelation matrix … 𝑅𝑋 0 𝑅𝑋 1 𝑅𝑋 𝑝 − 1 … … 𝑅 𝑝−2 𝑅𝑋 1 𝑅 0 𝑹𝑋 = …𝑋 . 𝑋 … … … 𝑅𝑋 𝑝 − 1 𝑅𝑋 𝑝 − 2 … 𝑅𝑋 0 Assume that inverse of 𝑹𝑋 exist, then 𝝎𝑜 = 𝒓𝑋 𝑹𝑋−1 p coefficients can be uniquely determined by knowing the p+1 autocorrelations. DPCM (contd.) • Linear predictor is implemented as a tapped delay line filter DPCM (contd.) Adaptive DPCM • Speech is a non-stationary process • Correlation values 𝑅𝑋 [𝑘] changes with time • Predictor coefficients should be made adaptive Delta Modulation • DM is a simplified version of DPCM • If a message signal is sampled at a rate much higher than the Nyquist rate • Adjacent samples are highly correlated • Only a two-level (1 bit) quantizer can be used. • It provides staircase approximation of the oversampled signal as shown in the adjacent Fig. DM (contd.) DM (contd.) • Slope overload distortion • Occur in the region of the steep slope of 𝑚(𝑡) • To avoid it ∆ 𝑑𝑚(𝑡) ≥ 𝑚𝑎𝑥 𝑇𝑠 𝑑𝑡
• Occur in the region of low slope
Ans. 1.08 V Ans. (a) Step size ∆ ≥ 0.126 𝑉 (b) (SNR)o = 475 = 26.8 dB
Name of The Experiment-Design & Simulation of Differentiator Amplifier Using 741 Op-Amp IC. Instrument/ Components Required - Proteus Simulator. Theory