Information Rate Speech Information Rates speech... · 5 Instantaneous Quantization • separating...

13
1 Digital Speech Processing Digital Speech Processing— Lecture 15 Lecture 15 1 Speech Coding Methods Based Speech Coding Methods Based on Speech Waveform on Speech Waveform Representations and Speech Representations and Speech Models Models—Uniform and Non Uniform and Non- Uniform Coding Methods Uniform Coding Methods Analog Analog-to to-Digital Conversion Digital Conversion (Sampling and Quantization) (Sampling and Quantization) 2 Class of “waveform coders” can be represented in this manner Information Rate Information Rate , () Waveform coder information rate, of the digital representation of the signal, , defined as: w a I x t i 3 / 1 , where is the number of bits used to represent each sample and is the numbe w S S I BF BT B F T = = = r of samples/second Speech Information Rates Speech Information Rates Production level: – 10-15 phonemes/second for continuous speech 32-64 phonemes per language => 6 bits/phoneme Information Rate=60-90 bps at the source Waveform level 4 Waveform level speech bandwidth from 4 – 10 kHz => sampling rate from 8 – 20 kHz need 12-16 bit quantization for high quality digital coding Information Rate=96-320 Kbps => more than 3 orders of magnitude difference in Information Rates between the production and waveform levels Speech Analysis/Synthesis Systems Speech Analysis/Synthesis Systems 5 • Second class of digital speech coding systems: o analysis/synthesis systems o model-based systems o hybrid coders o vocoder (voice coder) systems • Detailed waveform properties generally not preserved o coder estimates parameters of a model for speech production o coder tries to preserve intelligibility and quality of reproduction from the digital representation Speech Coder Comparisons Speech Coder Comparisons • Speech parameters (the chosen representation) are encoded for transmission or storage • analysis and encoding gives a data parameter vector data parameter vector computed at a sampling rate 6 data parameter vector computed at a sampling rate much lower than the signal sampling rate • denote the “frame rate” of the analysis as F fr • total information rate for model-based coders is: • where B c is the total number of bits required to represent the parameter vector m c fr I B F =

Transcript of Information Rate Speech Information Rates speech... · 5 Instantaneous Quantization • separating...

Page 1: Information Rate Speech Information Rates speech... · 5 Instantaneous Quantization • separating the processes of sampling and quantization • assume x(n) obtained by sampling

1

Digital Speech ProcessingDigital Speech Processing——Lecture 15Lecture 15

1

Speech Coding Methods Based Speech Coding Methods Based on Speech Waveform on Speech Waveform

Representations and Speech Representations and Speech ModelsModels——Uniform and NonUniform and Non--Uniform Coding MethodsUniform Coding Methods

AnalogAnalog--toto--Digital ConversionDigital Conversion(Sampling and Quantization)(Sampling and Quantization)

2Class of “waveform coders” can be

represented in this manner

Information RateInformation Rate

,

( )

Waveform coder information rate, of the digital representation of the signal,

, defined as:

w

a

I

x t

i

3

/

1 ,

where is the number of bits used to

represent each sample and is the

numbe

w S

S

I B F B TB

FT

= ⋅ =

=

r of samples/second

Speech Information RatesSpeech Information Rates• Production level:

– 10-15 phonemes/second for continuous speech– 32-64 phonemes per language => 6 bits/phoneme– Information Rate=60-90 bps at the source

• Waveform level

4

Waveform level– speech bandwidth from 4 – 10 kHz => sampling rate

from 8 – 20 kHz– need 12-16 bit quantization for high quality digital

coding– Information Rate=96-320 Kbps => more than 3

orders of magnitude difference in Information Rates between the production and waveform levels

Speech Analysis/Synthesis SystemsSpeech Analysis/Synthesis Systems

5

• Second class of digital speech coding systems:o analysis/synthesis systemso model-based systemso hybrid coderso vocoder (voice coder) systems

• Detailed waveform properties generally not preservedo coder estimates parameters of a model for speech productiono coder tries to preserve intelligibility and quality of reproduction from the digital representation

Speech Coder ComparisonsSpeech Coder Comparisons

• Speech parameters (the chosen representation) are encoded for transmission or storage

• analysis and encoding gives a data parameter vector• data parameter vector computed at a sampling rate

6

• data parameter vector computed at a sampling rate much lower than the signal sampling rate• denote the “frame rate” of the analysis as Ffr• total information rate for model-based coders is:

• where Bc is the total number of bits required to represent the parameter vector

m c frI B F= ⋅

Page 2: Information Rate Speech Information Rates speech... · 5 Instantaneous Quantization • separating the processes of sampling and quantization • assume x(n) obtained by sampling

2

Speech Coder ComparisonsSpeech Coder Comparisons

7

• waveform coders characterized by:• high bit rates (24 Kbps – 1 Mbps)• low complexity• low flexibility

• analysis/synthesis systems characterized by:• low bit rates (10 Kbps – 600 bps)• high complexity• great flexibility (e.g., time expansion/compression)

Introduction to Waveform CodingIntroduction to Waveform Coding

8

1 2

1 2

( ) ( )

( ) ( )

( ) ( )ω

π

ω πω

∞Ω

=−∞

=−∞

=

= Ω +

= Ω ⇒ = +

a

j Ta

k

ja

k

x n x nTj kX e X j

T Tj j kT X e X

T T T

Introduction to Waveform CodingIntroduction to Waveform CodingT=1/ Fs

9

• to perfectly recover xa(t) (or equivalently a lowpass filtered version of it) from the set of digital samples (as yet unquantized) we require that Fs = 1/T > twice the highest frequency in the input signal

• this implies that xa(t) must first be lowpass filtered since speech is not inherently lowpass

for telephone bandwidth the frequency range of interest is 200-3200 Hz (filtering range) => Fs = 6400 Hz, 8000 Hz

for wideband speech the frequency range of interest is 100-7000 Hz (filtering range) => Fs = 16000 Hz

Sampling Speech SoundsSampling Speech Sounds

10

• notice high frequency components of vowels and fricatives (up to 10 kHz) => need Fs > 20 kHz without lowpass filtering

• need only about 4 kHz to estimate formant frequencies

• need only about 3.2 kHz for telephone speech coding

Telephone Channel ResponseTelephone Channel Response

11

it is clear that 4 kHz bandwidth is sufficient for most applications using telephone speech because of inherent

channel band limitations from the transmission path

Statistical Model of SpeechStatistical Model of Speech• assume xa(t) is a sample function of a continuous-time

random process• then x(n) (derived from xa(t) by sampling) is also a

sample sequence of a discrete-time random process• x (t) has first order probability density p(x) with

12

• xa(t) has first order probability density, p(x), with autocorrelation and power spectrum of the form

[ ]( ) ( ) ( )

( ) ( ) τ

φ τ τ

φ τ τ∞

− Ω

−∞

= +

Φ Ω = ∫

a a a

ja a

E x t x t

e d

Page 3: Information Rate Speech Information Rates speech... · 5 Instantaneous Quantization • separating the processes of sampling and quantization • assume x(n) obtained by sampling

3

Statistical Model of SpeechStatistical Model of Speech

[ ][ ]

( ) has autocorrelation and power spectrum of the form: ( ) ( ) ( )

( ) ( ) ( )

( ) ( )

φ

φ

φ∞

Ω − Ω

=−∞

= +

= + =

Φ = ∑

a a a

j T j Tm

m

x nm E x n x n m

E x nT x nT mT mT

e m e

13

1 2 ( )

since ( ) is a samp

π

φ

=−∞

= Φ Ω +

∑ ak

kT T

m

2

led version of ( ), its transform is an infinite sum

of the power spectrum, shifted by aliased version of the power

spectrum of the original analog signal

φ τ

π=>

a

kT

Speech Probability Density FunctionSpeech Probability Density Function• probability density function for x(n) is the same as for xa(t) since

x(n)= xa(nT) => the mean and variance are the same for both x(n)and xa(t)

• need to estimate probability density and power spectrum from speech waveforms– probability density estimated from long term histogram of amplitudes– good approximation is a gamma distribution, of the form:

14

– simpler approximation is Laplacian density, of the form:

1 2 323 0

8

/

( ) ( )| |

σ

πσ

−⎡ ⎤= = ∞⎢ ⎥⎢ ⎥⎣ ⎦

x

x

x

p x e px

21 102 2

| |

( ) ( )σ

σ σ

= =x

x

x x

p x e p

Measured Speech DensitiesMeasured Speech Densities• distribution normalized so mean is 0 and variance is 1 (x=0, σx=1)

• gamma density more closely approximates measured distribution for

15

measured distribution for speech than Laplacian

• Laplacian is still a good model and is used in analytical studies

• small amplitudes much more likely than large amplitudes—by 100:1 ratio

LongLong--Time Autocorrelation and Power Time Autocorrelation and Power SpectrumSpectrum

{ }

{ } { }1 2

analog signal

( ) ( ) ( ) ( ) ( )

discrete-time signal ( ) ( ) ( ) ( ) ( ) ( )

τφ τ τ φ τ τ

φ φ

π

∞− Ω

−∞

∞ ∞Ω Ω

= + ⇔ Φ Ω =

= + = + =

∑ ∑

ja a a a a

a a a

j T j T

E x t x t e d

m E x n x n m E x nT x nT mT mT

k

16

1 2( ) ( ) ( )

estimated correlation and power

πφΩ − Ω

=−∞ =−∞

⇔ Φ = = Φ Ω +

∑ ∑j T j Tma

m k

ke m eT T

1

0

1 0 1

spectrum

ˆ ( ) ( ) ( ), | | ,

is large integer

ˆˆ ( ) ( ) ( )

φ

φ

− −

=

Ω − Ω

=−

= + ≤ ≤ −

Φ =

L m

n

Mj T j Tm

m M

m x n x n m m LL

L

e w m m e

Estimates ≠ Expectations

• tapering with m due to finite windows

• need window on estimated correlation for smoothing because of discontinuity at ±M

Speech AC and Power SpectrumSpeech AC and Power Spectrum

1ˆ[ ]ˆ[ ] ˆφρ− ≤ =

mm

• can estimate long term autocorrelation and power spectrum using time-series analysis methods

Lowpass filtered speech

Bandpass filtered speech

17

1

01

2

0

10

1

11

0 1

[ ] ˆ[ ]

( ) ( ),

( )

, is window length

ρφ− −

=− −

=

+= ≤

≤ ≤ −

L m

nL m

n

m

x n x n mL

x nL

m L L

• 8 kHz sampled speech for several speakers

• high correlation between adjacent samples

• lowpass speech more highly correlated than bandpass speech

Speech Power SpectrumSpeech Power Spectrum

• power spectrum estimate derived from one minute of speech

• peaks at 250-500 Hz (region of maximal spectral

18

(region of maximal spectral information)

• spectrum falls at about 8-10 db/octave

• computed from set of bandpass filters

Page 4: Information Rate Speech Information Rates speech... · 5 Instantaneous Quantization • separating the processes of sampling and quantization • assume x(n) obtained by sampling

4

Alternative Power Spectrum Alternative Power Spectrum EstimateEstimate

2 2/ /

ˆ estimate long term correlation, ( ), using sampled speech, thenweight and transform, giving:

ˆˆ ) ( ) ( )π π

φ

φ −

Φ = ∑M

j k N j km N

M

m

(e w m m e

19

ˆ this lets us use ( ) to get a power ˆspectrum estimate

φ=−

Φ

m M

m(e 2 / )

via the weighting window, ( )

πj k N

w m

Contrast linear versus logarithmic scale for power

spectrum plots

Estimating Power Spectrum via Estimating Power Spectrum via Method of Averaged PeriodogramsMethod of Averaged Periodograms

21

0

1

Periodogram defined as:

( ) ( ) ( )

where [ ] is an -point window (e.g., Hamming window), and

ω ω−

=

= ∑

i

i

Lj j n

nP e x n w n e

LUw n L

20

( )1

2

0

1 ( )

is a normalizing constant that compensates for window

=

= ∑i

L

nU w n

LU

212 2

0

1 0 1( )/ ( / )

tapering use the DFT to compute the periodogram as:

( ) ( ) ( ) , ,

is size of DFT

π π−

=

= ≤ ≤ −∑

iL

j k N j N kn

nP e x n w n e k N

LUN

Averaging (Short) PeriodogramsAveraging (Short) Periodograms variability of spectral estimates can be reduced by averaging

short periodograms, computed over a long segment of speech using an -point window, a short periodogram is the same as

the STFT of the weL

i

i

12 / 2 /[ ] ( ) [ ] [ ] , 0 1

ighted speech interval, namely:rR L

j k N j km Nr rR

m rRX k X e x m w rR m e k Nπ π

+ −−

=

= = − ≤ ≤ −∑

21

where is the DFT transform size (number of frequency estimates) Consider using samples of speech (where S

NN

ii

212 / 2 /

0

1( ) ( ) , 0 1

is large); theaveraged periodogram is defined as:

where is the number of windowed segments in samples, is the window length, and is the windo

S

Kj k N j k N

rRr

S

N

e X e k NKLU

K NL U

π π−

=

Φ = ≤ ≤ −∑i

/ 2w normalization factor

use (shift window by half the window duration betweenadjacent periodogram estimates)

R L=i

88,00088,000--Point FFT Point FFT –– Female Female SpeakerSpeaker

22Single spectral estimate using all signal samples

Long Time Average SpectrumLong Time Average Spectrum

23

Long Time Average SpectrumLong Time Average Spectrum

Female speaker

24

Male speaker

Page 5: Information Rate Speech Information Rates speech... · 5 Instantaneous Quantization • separating the processes of sampling and quantization • assume x(n) obtained by sampling

5

Instantaneous QuantizationInstantaneous Quantization• separating the processes of sampling and

quantization• assume x(n) obtained by sampling a

bandlimited signal at a rate at or above the

25

gNyquist rate

• assume x(n) is known to infinite precision in amplitude

• need to quantize x(n) in some suitable manner

Quantization and CodingQuantization and CodingCoding is a two-stage process 1. quantization process:

ˆ( ) ( ) 2. encoding process:

ˆ( ) ( )- where is the (assumed fixed)

→Δ

x n x n

x n c n

26

quantization step sizeDecoding is a single-stage process 1. decoding process:

ˆ( ) ( )if ( ) ( ), (no errors in

ˆ ˆ transmission) then ( ) ( )ˆ ( ) ( ) coding and

quantization loses information

′ ′→′• =

′ =′ ≠ ⇒

c n x nc n c n

x n x nx n x nassume Δ’=Δ

BB--bit Quantizationbit Quantization• use B-bit binary numbers to represent the quantized

samples => 2B quantization levels• Information Rate of Coder: I=B FS= total bit rate in

bits/second– B=16, FS= 8 kHz => I=128 Kbps

27

– B=8, FS= 8 kHz => I=64 Kbps– B=4, FS= 8 kHz => I=32 Kbps

• goal of waveform coding is to get the highest quality at a fixed value of I (Kbps), or equivalently to get the lowest value of I for a fixed quality

• since FS is fixed, need most efficient quantizationmethods to minimize I

Quantization BasicsQuantization Basics

• assume |x(n)| ≤ Xmax (possibly ∞)– for Laplacian density (where Xmax = ∞),

can show that 0.35% of the samples fall

28

poutside the range -4σx ≤ x(n) ≤ 4σx => large quantization errors for 0.35% of the samples

– can safely assume that Xmax is proportional to σx

Quantization ProcessQuantization Process• quantization => dividing amplitude

range into a finite set of ranges, and assigning the same bin to all samples in a given range

0 1 1

1 2 2

0 100101

ˆ( ) ( )ˆ( ) ( )

= < ≤ ⇒

< ≤ ⇒

x x n x xx x n x x

3-bit quantizer => 8 levels

29

2 3 3

3 4

1 0 1

2 1 2

3 2 3

3 4

110111

0 011010001000

ˆ( ) ( )ˆ( ) ( )

ˆ( ) ( )ˆ( ) ( )ˆ( ) ( )ˆ( ) ( )

range level

− −

− − −

− − −

− −

< ≤ ⇒

< < ∞ ⇒

< ≤ = ⇒

< ≤ ⇒

< ≤ ⇒

−∞ < ≤ ⇒

x x n x xx x n x

x x n x xx x n x xx x n x x

x n x xcodeword

• codewords are arbitrary!! => there are good choices that can be made (and bad choices)

Uniform QuantizationUniform Quantization• choice of quantization ranges and levels so that

signal can easily be processed digitally

mid-riser mid-tread

30

1

1ˆ ˆquantization step size

− = Δ

− = Δ

Δ =

i i

i i

x xx x

Page 6: Information Rate Speech Information Rates speech... · 5 Instantaneous Quantization • separating the processes of sampling and quantization • assume x(n) obtained by sampling

6

MidMid--Riser and MidRiser and Mid--Tread QuantizersTread Quantizers• mid-riser

– origin (x=0) in middle of rising part of the staircase– same number of positive and negative levels– symmetrical around origin

• mid-tread– origin (x=0) in middle of quantization level– one more negative level than positive– one quantization level of 0 (where a lot of activity occurs)

31

• code words have direct numerical significance (sign-magnitude representation for mid-riser, two’s complement for mid-tread)

[ ]

[ ]2

1 0

1 1

- for mid-riser quantizer:

ˆ( ) ( ) ( )

where ( ) if first bit of ( ) if first bit of ( )

- for mid-tread quantizer code words are 3-bittwo's complement representation, givin

Δ= + Δ

= + =

= − =

x n sign c n c n

sign c n c nc n

gˆ ( ) ( ) = Δx n c n

AA--toto--D and DD and D--toto--A ConversionA Conversionˆ x [n]

x[n]

32

ˆ x B[n]

Quantization errorˆ[ ] [ ] [ ]= −e n x n x n

Quantization of a Sine WaveQuantization of a Sine WaveUnquantizedsinewave

3-bit quantizationwaveform

33

3-bitquantizationerror

8-bitquantizationerror

Δ

Δ

Quantization of Complex SignalQuantization of Complex Signal[ ] sin(0.1 ) 0.3cos(0.3 )x n n n= +

(a) Unquantized signal

(b) 3-bit quantized x[n]

34

(b) 3-bit quantized x[n]

(c) 3-bit quantization error, e[n]

(d) 8-bit quantization error, e[n]

Uniform QuantizersUniform Quantizers• Uniform Quantizers characterized by:

– number of levels—2B (B bits)– quantization step size-Δ

• if |x(n)|≤Xmax and x(n) is a symmetric density, thenΔ2B = 2 Xmax

Δ= 2 X / 2B

35

Δ 2 Xmax/ 2• if we let

• with x(n) the unquantized speech sample, and e(n) the quantization error (noise), then

ˆ( ) ( ) ( )x n x n e n= +

2 2( )Δ Δ

− ≤ ≤e n(except for last quantization level which can exceed Xmax

and thus the error can exceed Δ/2)

Quantization Noise ModelQuantization Noise Model1. quantization noise is a zero-mean, stationary

white noise process

2 quantization noise is uncorrelated with the

[ ] 2 0

0

( ) ( ) ,σ+ = =

=eE e n e n m m

otherwise

36

2. quantization noise is uncorrelated with the input signal

3. distribution of quantization errors is uniformover each quantization interval

[ ] 0( ) ( )E x n e n m m+ = ∀

1 2 20

= Δ − Δ ≤ ≤ Δ=

ep e eotherwise

( ) / / /2

2012

, σ Δ⇒ = =ee

Page 7: Information Rate Speech Information Rates speech... · 5 Instantaneous Quantization • separating the processes of sampling and quantization • assume x(n) obtained by sampling

7

Quantization ExamplesQuantization Examples

37

how good is the model of the

previous slide?

Speech Signal 3-bit Quantization Error

8-bit Quantization Error (scaled)

Typical Amplitude DistributionsTypical Amplitude Distributions

Laplaciandistribution isoften used asa model for speech signals

38

A 3-bit quantized signal has only8 different values

33--Bit Speech QuantizationBit Speech Quantization

Input toquantizer

39

3-bit quantizationwaveform

33--Bit Speech Quantization ErrorBit Speech Quantization Error

Input toquantizer

40

3-bitquantizationerror

55--Bit Speech QuantizationBit Speech Quantization

Input toquantizer

41

5-bit quantizationwaveform

55--Bit Speech Quantization ErrorBit Speech Quantization Error

Input toquantizer

42

5-bitquantizationerror

Page 8: Information Rate Speech Information Rates speech... · 5 Instantaneous Quantization • separating the processes of sampling and quantization • assume x(n) obtained by sampling

8

Histograms of Quantization NoiseHistograms of Quantization Noise

3-bit quantizationhistogram

43

8-bit quantizationhistogram

Spectra of Quantization NoiseSpectra of Quantization Noise

44

≈ 12 dB

Correlation of Quantization ErrorCorrelation of Quantization Error

45assumptions look good at 8-bit quantization, not as good at 3-bit levels

(however only 6 db variation in power spectrum level)

3-bit Quantization

8-bit Quantization

Sound DemoSound Demo• “Original” speech sampled at 16kHz, 16 bits/sample

• Quantized to 10 bits/sample

• Quantization error (x32) for 10 bits/sample

46

• Quantization error (x32) for 10 bits/sample

• Quantized to 5 bits/sample

• Quantization error for 5 bits/sample

SNR for QuantizationSNR for Quantization• can determine SNR for quantized speech as

( )( )

222

2 22

( )( )

( )( )σσ

= = =∑∑

x n

en

x nE x nSNR

e nE e n

47

222

2

22

12 2

0

12 3 2

max

max

(uniform quantizer step size)

assume ( ) (uniform distribution)

( )

σ

Δ =

Δ Δ• = − ≤ ≤

Δ=

Δ= =

B

e B

X

p e e

otherwiseX

SNR for QuantizationSNR for Quantization• can determine SNR for quantized speech as

( )( )

222

2 22

222 max

( )( )

( )( )σσ

σ

= = =

Δ= =

∑∑

x n

en

x nE x nSNR

e nE e n

X

48

2

22

10 102 2

12 3 2

3 2 10 6 4 77 20

4 6 7 216

max

max

max

( )

( ) ; ( ) log . log

if we choose , then . ,

σ

σσ σ

σ

σ

= =

⎡ ⎤ ⎡ ⎤= = = + −⎢ ⎥ ⎢ ⎥⎡ ⎤ ⎣ ⎦⎣ ⎦⎢ ⎥⎣ ⎦

• = = −

=

e B

Bx

e x

x

x

XSNR SNR dB BX

X SNR BB S 88 8

8 40 83 10 8

. , . , .

== == =

NR dBB SNR dBB SNR dB

The term Xmax/σx tells how big a signal can be

accurately represented

Page 9: Information Rate Speech Information Rates speech... · 5 Instantaneous Quantization • separating the processes of sampling and quantization • assume x(n) obtained by sampling

9

Variation of SNR with Signal LevelVariation of SNR with Signal Level

Overloadfrom

clipping

SNR improves 6 dB/bit,but it decreases 6 dB for each halving of theinput signal amplitude

49

12 dB

Clipping StatisticsClipping Statistics

50

ReviewReview----Linear Noise ModelLinear Noise Model

• assume speech is stationary random signal.

2 2{( [ ]) } σ= xE x n

51

• error is uncorrelated with the input.

• error is uniformly distributed over the interval

• error is stationary white noise, (i.e. flat spectrum)2

2

12( ) , | |ω σ ω πΔ

= = ≤e eP

2 2( / ) [ ] ( / ).− Δ < ≤ Δe n

0{ [ ] [ ]} { [ ]} { [ ]}= =E x n e n E x n E e n

Review of Quantization Review of Quantization AssumptionsAssumptions

1. input signal fluctuates in a complicated manner so a statistical model is valid

2. quantization step size is small enough to remove any signal correlated patterns in quantization error

52

quantization error3. range of quantizer matches peak-to-peak

range of signal, utilizing full quantizer range with essentially no clipping

4. for a uniform quantizer with a peak-to-peak range of ±4 σx, the resulting SNR(dB) is 6B-7.2

Uniform Quantizer SNR IssuesUniform Quantizer SNR IssuesSNR=6B-7.2

• to get an SNR of at least 30 dB, need at least B≥6 bits (assuming Xmax=4 σx)– this assumption is weak across speakers and different

transmission environments since σx varies so much (order of 40 dB) across sounds speakers and input conditions

53

dB) across sounds, speakers, and input conditions– SNR(dB) predictions can be off by significant amounts if full

quantizer range is not used; e.g., for unvoiced segments => need more than 6 bits for real communication systems, more like 11-12 bits

– need a quantizing system where the SNR is independent of the signal level => constant percentage error rather than constant variance error => need non-uniform quantization

Instantaneous CompandingInstantaneous Companding• to order to get constant percentage error (rather than constant

variance error), need logarithmically spaced quantization levels– quantize logarithm of input signal rather than input signal itself

54

Page 10: Information Rate Speech Information Rates speech... · 5 Instantaneous Quantization • separating the processes of sampling and quantization • assume x(n) obtained by sampling

10

μμ--Law CompandingLaw Companding

[ ] [ ] ( ) ln | ( ) | ( ) exp ( ) ( )

=

= ⋅

y n x nx n y n sign x n

μ−LawCompander Quantizer

y[n] ˆ y [n]x[n]

55

[ ] [ ][ ]

[ ]

1 0

1 0

( ) p ( ) ( )

where ( ) ( )( )

the quantized log magnitude isˆ log

log | ( ) | ( ) new error signalε

• = + ≥

= − <•

=

= +

y g

sign x n x nx n

y(n) Q | x(n) |x n n

μμ--Law CompandingLaw Companding

[ ][ ] [ ]

[ ][ ] 1

assume that ( ) is independent of log | ( ) | . The inverse isˆ ˆ exp ( )

| ( ) | ( ) exp ( )

( ) exp ( )

assume ( ) is small, then exp ( ) ( ) ...

ε

ε

ε

ε ε ε

= ⋅⎡ ⎤⎣ ⎦= ⋅

= ⋅

• ≈ + +

n x nx(n) y(n) sign x n

x n sign x n n

x n n

n n n

56

[ ]1ˆ ( )= +x(n) x n [ ]

2 2 2

2

2 2

2

1

( ) ( ) ( ) ( ) ( ) ( ) since we assume ( ) and ( ) are independent, then

is independent of --it depends only on step size

ε

ε

ε ε

ε

σ σ σ

σσ σ

σ

= + = +

= ⋅

= =

f x

x

f

x

n x n n x n x n f nx n n

SNR

SNR

PseudoPseudo--Logarithmic CompressionLogarithmic Compression

• unfortunately true logarithmic compression is not practical, since the dynamic range (ratio between the largest and smallest values) is infinite => need an infinite number of quantization levels

• need an approximation to logarithmic compression => μ-law/A-law compression

57

μμ--law Compressionlaw Compression[ ]

[ ]1

1max

max

| ( ) |log

log( )

μ

μ

=

⎡ ⎤+⎢ ⎥

⎣ ⎦= ⋅+

y(n) F x(n)

x nX

X sign x(n)

58

0 00

max

max

when ( ) ( ) when , ( ) ( ) linear compression when is large, and for large | ( ) |

| ( ) | | ( ) | loglog

μμ

μμ

• = ⇒ =• = = ⇒•

⎡ ⎤≈ ⋅ ⎢ ⎥

⎣ ⎦

x n y ny n x n

x n

X x ny nX

Histogram for Histogram for μμ --Law CompandingLaw Companding

Speechwaveform

59

Output ofμ-Lawcompander

μμ--law approximation to loglaw approximation to log law encoding gives a good approximation to constant percentage

error (| ( ) | log | ( ) |)

μ• −

∝y n x n

60

Page 11: Information Rate Speech Information Rates speech... · 5 Instantaneous Quantization • separating the processes of sampling and quantization • assume x(n) obtained by sampling

11

SNR for SNR for μμ--law Quantizerlaw Quantizer

[ ]2

10 106 4 77 20 1 10 1 2

6

max max( ) . log ln( ) log

dependence on

μμσ μσ

⎡ ⎤⎛ ⎞ ⎛ ⎞⎢ ⎥= + − + − + +⎜ ⎟ ⎜ ⎟⎢ ⎥⎝ ⎠ ⎝ ⎠⎣ ⎦

• ⇒

x x

X XSNR dB B

B B good

61

6

max

max

dependence on

much less dependence on

for large , is less sensitive to changes in

σ

μσ

• ⇒

• ⇒

• ⇒

x

x

B BX

XSNR

good

good

good

- -law quantizer used in wireline telephony for more than 40 yearsμ

μμ--Law CompandingLaw Companding

Mu-law compressed signal utilizes almost the

62

full dynamic range (±1) much more

effectively than the original speech

signal

μμ--Law Quantization ErrorLaw Quantization Error(a) Input speech

signal

63

(b) Histogram of μ=255 compressed speech

(c) Histogram of 8-bit companded quantization error

Comparison of Linear and Comparison of Linear and μμ--law Quantizerslaw Quantizers

Dashed lineDashed line –linear (uniform) quantizers with 6, 7, 8 and 12 bit quantizers

Solid lineSolid line – μ-

64

μlaw quantizers with 6, 7 and 8 bit quantizers (μ=100)

• can see in these plots that Xmax characterizes the quantizer (it specifies the ‘overload’ amplitude), and σx characterizes the signal (it specifies the signal amplitude), with the ratio (Xmax/σx) showing how the signal is matched to the quantizer

Comparison of Linear and Comparison of Linear and μμ--law law QuantizersQuantizers

Dashed lineDashed line –linear (uniform) quantizers with 6, 7, 8 and 13 bit

ti

65

quantizers

Solid lineSolid line – μ-law quantizers with 6, 7 and 8 bit quantizers (μ=255)

Analysis of Analysis of μμ--Law PerformanceLaw Performance• curves show that μ-law quantization can maintain

roughly the same SNR over a wide range of Xmax/σx, for reasonably large values of μ– for μ=100, SNR stays within 2 dB for

8< Xmax/σx<30– for μ=500, SNR stays within 2 dB for

66

μ , y8< Xmax/σx<150

– loss in SNR in going from μ=100 to μ=500 is about 2.6 dB => rather small sacrifice for much greater dynamic range

– B=7 gives SNR=34 dB for μ=100 => this is 7-bit μ-law PCM-the standard for toll quality speech coding in the PSTN => would need about 11 bits to achieve this dynamic range and SNR using a linear quantizer

Page 12: Information Rate Speech Information Rates speech... · 5 Instantaneous Quantization • separating the processes of sampling and quantization • assume x(n) obtained by sampling

12

CCITT G.711 StandardCCITT G.711 Standard

• Mu-law characteristic is approximated by 15 linear segments with uniform quantization within a segment.

uses a mid tread quantizer +0 and 0 are

67

– uses a mid-tread quantizer. +0 and –0 are defined.

– decision and reconstruction levels defined to be integers

• A-law characteristic is approximated by 13 linear segments– uses a mid-riser quantizer

Summary of Uniform and Summary of Uniform and μμ--Law PCMLaw PCM

• quantization of sample values is unavoidable in DSP applications and in digital transmission and storage of speech

• we can analyze quantization error using a random noise model

• the more bits in the number representation the lower the

68

the more bits in the number representation, the lower the noise. The ‘fundamental theorem’ of uniform quantization is that “the signal-to-noise ratio increases 6 dB with each added bit in the quantizer”; however, if the signal level decreases while keeping the quantizer step-size the same, it is equivalent to throwing away one bit for each halving of the input signal amplitude

• μ-law compression can maintain constant SNR over a wide dynamic range, thereby reducing the dependency on signal level remaining constant

Quantization for Optimum SNR Quantization for Optimum SNR (MMSE)(MMSE)

• goal is to match quantizer to actual signal density to achieve optimum SNR– μ-law tries to achieve constant SNR over wide

range of signal variances => some sacrifice

69

range of signal variances > some sacrifice over SNR performance when quantizer step size is matched to signal variance

– if σx is known, you can choose quantizer levels to minimize quantization error variance and maximize SNR

Quantization for MMSEQuantization for MMSE• if we know the size of the signal (i.e., we know

the signal variance, σx2) then we can design a

quantizer to minimize the mean-squared quantization error.– the basic idea is to quantize the most probable

l ith l d l t b bl ith hi h

70

samples with low error and least probable with higher error.

– this would maximize the SNR– general quantizer is defined by defining M

reconstruction levels and a set of M decision levels defining the quantization “slots”.

– this problem was first studied by J. Max, “Quantizing for Minimum Distortion,” IRE Trans. Info. Theory, March,1960.

Quantizer Levels for Maximum Quantizer Levels for Maximum SNRSNR

( )[ ]

22 2

2 2 1 1 1 2( / ) ( / ) ( / )

variance of quantization noise is:

ˆ ( ) ( ) ( )

ˆ with ( ) ( ) . Assume quantization levels

ˆ ˆ ˆ ˆ ˆ , ,..., , ,...,

associating quant

σ

− − + −

⎡ ⎤⎡ ⎤= = −⎣ ⎦ ⎣ ⎦• =

⎡ ⎤⎣ ⎦•

e

M M M

E e n E x n x n

x n Q x n M

x x x x x

ization level with signal intervals as:

71

1ˆ quantization level for interval ,

for symmetric, zero-mean distributions, with large amplitudes ( )it makes sense to define the boundary

−⎡ ⎤= ⎣ ⎦• ∞

j j jx x x

0 2

2 2

0 /

points: (central boundary point), the error variance is thus

( )σ

±

−∞

= = ±∞

= ∫

M

e e

x x

e p e de

Optimum Quantization LevelsOptimum Quantization Levels

22 2

ˆ/

/

ˆ by definition, ; thus we can make a change of variables to give:ˆ ˆ ( ) ( ) ( / ) ( )

giving

ˆ ( ) ( )σ

• = −= − = =

= −∑ ∫i

e e xx x

xM

e i x

e x xp e p x x p x x p x

x x p x dx

72

12 1/

assuming ( ) ( ) so that th−

=− +

• = −

∑ ∫i

e i xi M x

p x p x

1

22 2

1

2

2/

e optimum quantizer is antisymmetric, thenˆ ˆ and

thus we can write the error variance as

ˆ ( ) ( )

goal is to minimize through choices of

σ

σ−

− −

=

= − = −

= −

∑ ∫i

i

i i i i

xM

e i xi x

e

x x x x

x x p x dx

ˆ { } and { }i ix x

Page 13: Information Rate Speech Information Rates speech... · 5 Instantaneous Quantization • separating the processes of sampling and quantization • assume x(n) obtained by sampling

13

Solution for Optimum LevelsSolution for Optimum Levels

1

2

2 0 1 2 2 1

12

ˆ to solve for optimum values for { } and { }, we differentiate wrtthe parameters, set the derivative to 0, and solve numerically:

ˆ ( ) ( ) , ,..., / ( )

(

σ

− = =

=

∫i

i

i i e

x

i xx

i

x x

x x p x dx i M

x ( )1 1 2 2 1 2ˆ ˆ ) , ,..., /++ = −i ix x i M

73

2i ( )

( )1

0 20 3/ with boundary conditions of , can also constrain quantizer to be uniform and solve for value of Δ

that maximizes SNR- optimum boundary points lie halfway betw

+

±• = = ±∞

i i

Mx x

1

2een / quantizer levelsˆ - optimum location of quantization level is at the centroid of the probability

density over the interval to - solve the above set of equations (1,2,3) iterative

i

i i

Mx

x x

{ } { }ˆly for ,i ix x

Optimum Quantizers for Laplace DensityOptimum Quantizers for Laplace Density

1Assumes

=xσ

74

M. D. Paez and T. H. Glisson, “Minimum Mean-Squared Error QuantizationIn Speech,” IEEE Trans. Comm., April 1972.

Optimum Quantizer for 3Optimum Quantizer for 3--bit Laplace Density; bit Laplace Density; Uniform CaseUniform Case

75step size decreases roughly exponentially

with increasing number of bitsquantization levels get further apart as

the probability density decreases

Performance of Optimum QuantizersPerformance of Optimum Quantizers

76

SummarySummary• examined a statistical model of speech showing

probability densities, autocorrelations, and power spectra• studied instantaneous uniform quantization model and

derived SNR as a function of the number of bits in the quantizer, and the ratio of signal peak to signal variance

• studied a companded model of speech that i t d l ith i i d h d th t

77

approximated logarithmic compression and showed that the resulting SNR was a weaker function of the ratio of signal peak to signal variance

• examined a model for deriving quantization levels that were optimum in the sense of matching the quantizer to the actual signal density, thereby achieving optimum SNR for a given number of bits in the quantizer