TRANSFORMED VECTOR QUANTIZATION BASED - Shodhganga

23
CHAPTER 3 Transformed Vector Quantization with Orthogonal Polynomials 3.1. Introduction In the previous chapter, a new integer image coding technique based on orthogonal polynomials for monochrome images was proposed. After the proposed transformation the coefficients are scalar quantized and entropy coded in order to obtain a higher compression ratio. This technique proved to be better in the sense that it gives higher PSNR value. However, the quality of the reconstructed image degrades when the quality factor increases due to the scalar quantization step effect. To overcome this problem, in this chapter, a new vector quantization technique has been proposed in the transformed domain. The rationale behind the introduction of vector quantization is that the vector quantization of signal reduces the coding bit rate significantly with good quality of reconstruction picture. The proposed transformed vector quantization exploits the combined features of energy preservation by the proposed transform coding and high compression ratio of the vector quantization. 3. 1. 1 Vector quantization In the current scenario, Vector Quantization [Alle92] has been found to be an efficient data compression technique for speech and image as it provides many attractive features for image compression. A vector quantizer Q of dimension k and size N is mapped from a point in k-dimensional Euclidean space R k , into a finite set C containing N output or reproduction points that exist in the same Euclidean space as the original point. These reproduction points are known as codewords and these set of codewords are called codebook C with N distinct codewords in the set. Thus, the mapping function Q is defined as, Q: R k → C ... (3.1)

Transcript of TRANSFORMED VECTOR QUANTIZATION BASED - Shodhganga

CHAPTER 3

Transformed Vector Quantization with Orthogonal

Polynomials

3.1. Introduction

In the previous chapter, a new integer image coding technique based on

orthogonal polynomials for monochrome images was proposed. After the

proposed transformation the coefficients are scalar quantized and entropy coded

in order to obtain a higher compression ratio. This technique proved to be better

in the sense that it gives higher PSNR value. However, the quality of the

reconstructed image degrades when the quality factor increases due to the scalar

quantization step effect. To overcome this problem, in this chapter, a new vector

quantization technique has been proposed in the transformed domain. The

rationale behind the introduction of vector quantization is that the vector

quantization of signal reduces the coding bit rate significantly with good quality

of reconstruction picture. The proposed transformed vector quantization exploits

the combined features of energy preservation by the proposed transform coding

and high compression ratio of the vector quantization.

3. 1. 1 Vector quantization

In the current scenario, Vector Quantization [Alle92] has been found to be

an efficient data compression technique for speech and image as it provides

many attractive features for image compression. A vector quantizer Q of

dimension k and size N is mapped from a point in k-dimensional Euclidean space

Rk, into a finite set C containing N output or reproduction points that exist in the

same Euclidean space as the original point. These reproduction points are known

as codewords and these set of codewords are called codebook C with N distinct

codewords in the set. Thus, the mapping function Q is defined as,

Q: Rk → C ... (3.1)

CHAPTER 3. TVQ WITH ORTHOGONAL POLYNOMIALS

36

The rate of the vector quantizer or the number of bits (r) used to express each

quantized vector is given by the relation,

r = log2 N / k … (3.2)

This rate equation is very useful as it gives the amount of compression that can

be expected in a particular VQ coding scheme.

Vector quantization can be considered as a pattern recognizer where an

input pattern is approximated by a predetermined set of standard patterns

[Robe02]. Experiments have shown that vector quantization produces superior

performance over scalar quantization even when the components of the input

vectors are statistically independent. Vector quantization exploits the linear and

nonlinear dependence among vectors. It also provides flexibility in choosing

multi-dimensional quantizer cell shapes and in choosing a desired codebook size.

If scalar quantization is extended to k dimensional vectors using N levels, then

the codebook would contain N x k codewords. In the case of vector quantization

there could be arbitrary partitions with integer number of codewords N. Another

advantage of vector quantization over scalar quantization is the fractional value

of resolution that is achievable. This is very important for low bit rate

applications where low resolution is sufficient.

The number of codewords in the codebook decides the quality of the

reconstructed vectors. If the number of code words is large, the output vectors

would be close to the input vectors. The dimension (i.e the number of elements

present in each vector) of the input vectors and code words also play a crucial

role in quality of reconstruction. Ideally, the compression performance improves

as the vector dimension increases but the tradeoff is the increased coding

complexity. Besides dimension, the difficult task in any VQ scheme is the

generation of codewords that best represent the input vectors [Lind80, Gray84].

The performance of the quantizer is assessed using a suitable statistical distortion

CHAPTER 3. TVQ WITH ORTHOGONAL POLYNOMIALS

37

measure. The generic statistical distortion measure as applied to vectors is

represented as:

))(( , iiixQxdD … (3.3)

where xi is the input vector and Q(xi) is the approximation of xi and

))(,( ii xQxd represents the squared Euclidean Distance between the input vector

ix and its approximation )( ixQ .

3. 2 Literature survey

Vector quantization is a very powerful method for lossy compression of

data such as images and speech. The lossy compression scheme can be analyzed

using rate distortion theory [Alle92]. In this scheme the decompressed data will

not be a replica of the original. Instead, it will be distorted by an amount D.

According to Shannon’s rate distortion theory [Jude76], vector quantization of

signals reduce the coding bit rate significantly when compared to scalar

quantization. Vector Quantization takes M number of multi dimensional vectors

and reduces their representation to k number of code words, where k < M. The

key to Vector Quantization is to construct a good codebook of representative

vectors. The most popular method for designing a codebook was proposed by

Linde, Buzo and Gray in [Lind80, Gray84]. This method is now commonly

referred to as LBG algorithm. In this algorithm, all the training vectors are

clustered, using the minimum distortion principle, around trial code vectors. The

centroids of these clusters then become the new trial code vectors at the next

iteration. This procedure continues until there is no significant change in the total

distortion between cluster members and the code vectors around which they are

clustered. Then the training vectors are compared with codebook that is

generated by LBG algorithm. The result is an index position of codebook with

minimum distortion. This algorithm works directly on the image pixels and it

uses the full search technique in the encoding process. So it takes longer time to

CHAPTER 3. TVQ WITH ORTHOGONAL POLYNOMIALS

38

construct the codebook and each codeword contains every pixel of a block. Due

to the enormous size of the code book the search time to find the best match

vector in the encoding process increases drastically. The methods available in the

literature to alleviate this problem are presented below.

Generation of fast codebooks for Vector Quantization of images, based on

the features of training vectors has been reported by Hsieh. C.H et. al [Hsie91].

This method uses the good energy compaction capability of the Discrete Cosine

Transform and uses certain significant components of the feature space to

construct the binary tree. Design of codebook for vector quantization with the

discrete cosine transform Coefficients as training vector feature has been

reported by Hsieh. C. H. [Hsie92]. In this work, the energy preserving property

of the DCT has been used to reduce the dimension of the feature training vector.

From these works, research activities on design of transformed vector

quantization (TVQ) that combines the features of transform coding and vector

quantization have gained popularity.

Timo Kaukoranta, Pasi Fanti and Olli Nevalanen [Timo00] have reported

a scheme for reducing the number of distance calculations in the LBG algorithm

and are included in several fast LBG variants reducing their running time by over

50% on average. A scheme based on Vector Quantization in transformed domain

has been reported in [Robe02] by Roberts et al. This scheme uses a fast Kohonen

self-organizing neural network algorithm to achieve reduction in codebook

construction time and transformed vector quantization to obtain better

reconstructed images. Hsien-Wen Tseng et al. [Hsie05] have reported a classified

vector quantization (CVQ) in the DCT transform domain. In this method DC

coefficients are coded by difference pulse code modulation and only the most

important AC coefficients are coded using classified vector quantization (CVQ).

These AC coefficients are selected to train the codebook according to the energy

packing region of different block classes. Evaluation of TVQ as low bit rate

CHAPTER 3. TVQ WITH ORTHOGONAL POLYNOMIALS

39

image coding has been reported by Clyde Shamers et al. [Clyd04]. This coding

technique which is based on the combined use of discrete cosine transform and

Vector Quantization eliminates the artifacts generated in JPEG compression.

Use of wavelet transformation in the design of TVQ has been reported in

[Min05]. Here the relationship between the input vector and codeword, as well as

the relationship among code words and characteristics of code words in wavelet

domain are utilized. Another scheme for image compression with transform

vector quantization of the wavelet coefficients has been reported by Momotaz

Begum et al. [Momo03]. This scheme utilizes mean-squared error and variance

based selection for good clustering of data vectors in the training space. The two

major drawbacks of the LBG algorithm namely, the choice of initial codebook

and the huge computational burden have been alleviated by this scheme. Fast

search algorithm for vector quantization with multiple triangle inequalities in the

wavelet domain has been reported by Chaur H.H and Liu. Y.J. [Chau00]. The

multiple triangle inequalities confine a search range using the intersection of

search areas generated from several control vectors. Also a systematic way for

designing the control vectors is reported. The wavelet transform combined with

the partial distance elimination is used to reduce the computational complexity of

the distance calculation of vectors. A fast codeword searching algorithm based

on mean-variance pyramids of codewords [Lu00] and Hadamard Transformation

[Lu00a] have also been found in the literature. Given initial vectors, two design

techniques for adaptive orthogonal block transforms based on VQ codebooks are

presented in [Cagl98]. Both the techniques start from reference vectors that are

adapted to the characteristics of the signal to be coded, while using different

methods to create orthogonal bases. The resulting transforms represent a signal

coding tool that stands between a pure VQ scheme on one hand and signal-

independent, fixed block transformation like discrete cosine transform (DCT) on

the other.

CHAPTER 3. TVQ WITH ORTHOGONAL POLYNOMIALS

40

Review of early works on VQ can be found in [Nasr98]. S. Esakkirajan et

al. [Esak06] have proposed an image coding scheme based on contourlet

transform and multiscale VQ. Recently filter banks approach for VQ has been

developed by Brislawn and Wohlberg [Chri06] to overcome obstructions for a

class of half-sample symmetric filter banks. They employ lattice vector

quantization to ensure symmetry preserving rounding in reversible

implementations. Z. Liu et al. [Liu07] reported the use of biorthogonal wavelet

filter banks (BWFBs) for image coding with lower computational costs. Here a

new class of Integer Wavelet Transforms (IWT) parameterized simply by one

parameter, obtained by introducing a free variable to the lifting based

factorization of a Deslauriers-Dubuc interpolating filter, is introduced. The exact

one-parameter expressions for this class of IWTs are deduced. In this technique,

different IWTs are obtained by adjusting the free parameter and several IWTs

with binary coefficients are constructed.

In this chapter we explore the possibility of introducing a new VQ with

the integer transform coding proposed in chapter 2. This new integer

Transformed Vector Quantization takes the advantage of decorrelation and

energy compaction properties of Orthogonal Polynomials based Transform

coding and the superior rate distortion performance of VQ in the orthogonal

polynomials transformed domain. In this work, the energy preserving property of

the proposed transformation coding scheme is analyzed to truncate higher

frequency components. This truncated sub-image is then subjected to vector

quantization for effective codebook design with less complexity in sample space.

The important steps involved in the work presented in this chapter are

highlighted below.

Analysis of the point spread operator introduced in chapter 2 that defines the

proposed coding and shows its completeness, for the purpose of perfect

reconstruction of the original image by proposing difference operators.

CHAPTER 3. TVQ WITH ORTHOGONAL POLYNOMIALS

41

Formations of training vectors with scale quantized high energy transform

coefficients.

Design of codebook on training vectors as in LBG algorithm.

Identification of index values by performing vector quantization on training

vectors and entropy coding of the identified indices.

Inverse transformation with proposed basis operators after carrying out a

simple look-up process in the codebook.

3.3 Proposed orthogonal polynomials based framework for vector

quantization

3.3.1 Completeness of the proposed transformation

Before presenting the new Vector Quantization in the orthogonal

polynomials based transformation domain, we first prove its completeness and

the same is presented in this subsection.

The point spread operator in equation (2.3) described in chapter 2 that

defines the linear orthogonal transformation for gray scale images is obtained as

|M| |M|, where |M| is computed and the elements are scaled to make them

integers as follows.

222120

121110

020100

xuxuxu

xuxuxu

xuxuxu

M =

111

201

111

... (3.5)

The set of polynomial operators Oijn (0 ≤ i, j ≤ n-1) can be computed as

Oijn = ûi ûj

t

where ûi is the (i + 1)st

column vector of |M| and is the outer product.

In this subsection we prove that the proposed polynomials based

difference operators is complete and hence the reconstruction of the image under

CHAPTER 3. TVQ WITH ORTHOGONAL POLYNOMIALS

42

analysis after the said 2-D transformation is possible in terms of linear

combination of basis operators Oij and the transform coefficients ji .

The following symmetric finite differences for estimating partial

derivatives at (x, y) position of the gray level image I are analogous to the eight

finite difference operators Oijs excluding O00.

1

1

, 1,1,i

yx yixIyixIy

I

1

1

, ,1,1i

yx iyxIiyxIx

I

1

1

,2

2

1,,21,i

yx yixIyixIyixIy

I

1

1

,2

2

,1,2,1i

yx iyxIiyxIiyxIx

I … (3.6)

and so on.

In general,

ijji

ji

Oyx

and

02,0,,

jiandjiIOyx

Ijiijji

ji

… (3.7)

where | | indicates the arrangement in dictionary sequence and ( , ) indicates the

inner product. Hence, Oijs are symmetric finite difference operators. ji s are the

coefficients of the linear transformations and are defined as follows.

ji | = |M|t |I| … (3.8)

where | M | is the 2-D point-spread operator defined as | M | = |M| |M|.

CHAPTER 3. TVQ WITH ORTHOGONAL POLYNOMIALS

43

Now we will show that the orthogonal transformation defined in equation (3.8)

by the orthogonal system | M | is complete. We may obtain an orthogonal system

| H | by normalizing | M | as follows.

|H| = | M | (|M |t | M |)

Consider the following orthonormal transformations

|Z|= | H |t |I| = (|M |

t | M |)

½ | M |

t |I| = (|M)

t |M |)

-½ | |

Since, | H | is unitary,

|I| = |H| |Z| = | M |

2

0

2

0i j

ijij O … (3.9)

where | | = (|M |t | M |)

-1 | |.

As per equation (3.9) the image region | I | can be expressed as a linear

combination of the nine basis operators of which |O00| is the local averaging

operator and the remaining eight are finite difference operators (equation 3.7).

From equation (3.9) we obtain the completeness relation or Bessel's equality as

follows.

2

0

2

0

22

0

2

0

2..,,i j

ij

i j

ij ZIeiZZII ... (3.10)

Thus it is proved that the proposed transformation is complete and hence the

transformed image can be reconstructed perfectly. In the next section, Vector

Quantization on the proposed orthogonal polynomials based transformation

coefficients is presented.

CHAPTER 3. TVQ WITH ORTHOGONAL POLYNOMIALS

44

3. 4. Proposed vector quantization

In this section, a new Transformed Vector Quantization scheme that

facilitates the image coding using Orthogonal Polynomials is proposed. This

proposed scheme combines both transform coding and VQ technique. The

advantage of combining both the proposed transformation and VQ is that, when a

linear transform is applied to the vector signal, the information is compacted into

a subset of the vector components. In the frequency domain, the high energy

components are concentrated in the low frequency region. This means that the

transformed vector components in the high frequency regions have very little

information and so these low energy components can be entirely discarded. The

procedure involved in the proposed transformed vector quantization is presented

hereunder.

3.4.1 Formation of training vector

The proposed TVQ starts by portioning the original image of size (R x C)

into non-overlapping sub-blocks of size (n x n) and mapping them to the

frequency domain by applying the proposed orthogonal polynomials based

transformation as described in section 2.2.2. The resultant transform coefficients

ji {i, j = 0, 1, 2, …, n-1} are subjected to scale quantization using the default

quantization table of JPEG baseline system. The aim of using the scale

quantization is to obtain a suboptimal VQ codebook with reduced reconstruction.

As the proposed orthogonal polynomials based transform and DCT based JPEG

are both unitary, without loss of generosity, the default quantization table of

JPEG is utilized here. The scale quantized transform coefficients ji are then

arranged into 1-D zig-zag sequence to form a k-dimensional transformed input

vector Y and it is mapped into a p-dimensional (p < k) training vector T by

considering only the energy preserving components due to the proposed

CHAPTER 3. TVQ WITH ORTHOGONAL POLYNOMIALS

45

orthogonal polynomials transformation. The energy preservation by the proposed

transform is extracted as follows.

The energy preserving property of the proposed transformation is based

on the estimates of the variances 2

, jiZ s corresponding to the mean squared

amplitude responses of the basis, difference operators Oi,j. These variances are

computed as

jjii

ji

jiWW

Z,,

2

,2

,

… (3.11)

where MMWt

and IMMt

ji

The F-ratio test [Fish87] is then conducted on the variances 2

, jiZ s to

identify the significant responses towards signal compared to a threshold. The

fact that a variance passes the test implies that considerable energy is compacted

in the transform coefficients ji corresponding to that variance. After applying

the F-ratio test on every 2

, jiZ the insignificant responses that do not contain much

energy can be discarded.

The aforementioned process is repeated to form the training vectors of all

the partitioned sub-blocks and the codebook is designed as described in the

following section using the training vectors.

3.4.2 Codebook design

The selection of the initial set of codewords is a very tricky problem in

any VQ design. A variety of techniques are available in the literature for the

initial selection of codewords [Alle92]. The proposed technique uses the splitting

technique for choosing the initial set of codewords. In the splitting technique, the

CHAPTER 3. TVQ WITH ORTHOGONAL POLYNOMIALS

46

initial codeword C0 is chosen by taking the centroid of the entire training vectors

T. Then, this codeword is split into two, namely C0 + ε and C0 - ε, where ε is any

Euclidean norm and indicates the optimization precision. This process of

iteratively splitting each codeword into two continues until the desired number of

codewords of the codebook is obtained. These codewords do not qualify as final

codewords for quantizing the input vectors as they do not satisfy the necessary

conditions of optimality. However these can be used as the initial codewords. To

optimize the codewords, the proposed technique makes use of Linde Buzo Gray

(LBG) Algorithm [Lind80]. Here the training vectors T are clustered by

computing the minimum distortion of the training vectors against the initial

codewords C. The centroids of the clusters thus formed become the new

codewords for the next iteration. This procedure continues until there is no

significant change in the total distortion between cluster members and the

codewords around which they are clustered. The final set of code vectors

obtained constitutes the codebook. The steps involved in the design of codebook

in the proposed transform domain are presented below:

Step 1. Initialize the initial codeword C0 with the mean of the entire set of

training vectors Z and perturbation value ε to a fixed value.

Step 2. Initialize iteration number n to zero and distortion D-1 to ∞.

Step 3. Form the desired number of codewords for the new codebook by

splitting each codeword into two using the binary splitting

operation.

Step 4. Optimize the new codebook formed in step 3 using the centroid

condition.

Step 5. Compute r = (D n-1 - D n)/ Dn. where D n-1 distortion before

optimization and D n is the distortion after optimization.

Step 6. Repeat steps 3 through 5 until r ≤ ε.

CHAPTER 3. TVQ WITH ORTHOGONAL POLYNOMIALS

47

Then the training vectors T are compared with codebook, and index

positions of code vectors that give minimum distortion, are identified. These

index values are entropy coded as in JPEG baseline system and are transmitted to

the receiver.

3.4.3 Reconstruction

The receiver decodes the received bits to get the index values. It then

initiates the lookup process in order to get the p-dimensional transformed

coefficients vector from the codebook which is identical to the one at the sender.

Then (k–p) additional components with value zero are appended to the vectors,

producing the k-dimensional vectors Y and scale dequantization is performed on

the elements of Y to get the transformed coefficients ji . Finally these

coefficients are subjected to inverse transform with the help of basis functions of

the proposed orthogonal polynomials as described in section 2.3 to get back the

decompressed image.

3.4.4 Time minimization

The goal of the proposed TVQ scheme is threefold. First, the proposed

scheme tries to minimize the time taken for construction of the codebook. The

second goal is to reduce the size of the codebook. Thirdly, the proposed scheme

aims to reduce the time consumption for the encoding process. To perceive how

these goals are achieved, let us consider a k-dimensional input vector X with a

resolution of r-bits per component constituting a total bit allocation of r x k bits.

Normally in VQ, the codebook size would be N = 2rk

. With the proposed TVQ,

using the same r-bits the maximum possible codebook size is reduced to N = 2rp

,

which can be of smaller magnitude since p < k. Also, the time taken for

constructing the codebook is reduced drastically, as the proposed scheme uses

only p-dimension vectors instead of k-dimension vectors. In a generic VQ, the

CHAPTER 3. TVQ WITH ORTHOGONAL POLYNOMIALS

48

number of comparisons required during the encoding process is N x k whereas

the proposed scheme requires only N x p comparisons. Hence, it is evident that

the proposed TVQ technique consumes less time and takes less storage for image

coding. Alternatively, for the same time and space, the resolution or codebook

size can be increased to obtain better performance. The steps involved in the

proposed TVQ technique, are presented below.

3. 4. 4. TVQ Algorithm

Input: Gray-level image of size (R x C). [ ] denotes the matrix and the suffix

denotes the elements of the matrix. Let [I] be the (n x n) non-overlapping image

region (block) extracted from the image.

Begin

Step 1. Divide the given input image into number of non-overlapping image

regions [I] of size (n x n).

Step 2. Repeat the steps 3 to 7 for all the image regions.

Step 3. Compute the orthogonal polynomials based transform coefficients

[] as described in section 2.2.2.

Step 4. Apply scale quantization on the transform coefficients [].

Step 5. Arrange the scale quantized [] in 1-D zig-zag sequence.

Step 6. Truncate the low energy components from the scale quantized []

based on the energy preserving property of the proposed

Orthogonal Polynomials based transformation as described in

section 3.4.1.

Step 7. Form a vector T using the truncated low frequency coefficients.

Step 8. Design the codebook with LBG algorithm on the vectors T as

described in section 3.4.2.

CHAPTER 3. TVQ WITH ORTHOGONAL POLYNOMIALS

49

Step 9. Perform VQ operation as explained in section 3.1.1 and obtain the

index values.

Step 10. The index values are subjected to entropy coding as in JPEG and

the coded value is transmitted to the receiver through channel.

Step 11. At the receiving end, decode the index values and form the

truncated code words using index values as a table look-up process.

Step 12. Obtain all the n2 coefficients by substituting zero values in

truncated high frequency coefficients.

Step 13. These n2

coefficients are then subjected to the scale dequantization

to form an approximation to the original transform coefficients [].

Step 14. Reconstruct the input image region [I] using the polynomial basis

functions as discussed section 2.3.

Step 15. Repeat the steps 11 to 14 until all the image regions [I] are

reconstructed.

End.

The above said proposed algorithm is presented in diagrammatic way in Figure

3. 1.

CHAPTER 3. TVQ WITH ORTHOGONAL POLYNOMIALS

50

Figure 3. 1: Flow diagram of the proposed TVQ technique

Index Value

Index Value

1001 …

1001 …

Choose the

Energy

Preserving

Coefficients

Channel

Symbol Decoder

Nearest Neighbor

principle

Symbol Encoder

Table Lookup

Scale

Quantizer

Generate

codebook with

LBG

Add zero values

to the decoded

code words to

compensate 16

coefficients

Scale

Dequantizer

Inverse

Transformation

XXX

XXX

Code book

Proposed

Transformation

Original Image

4 x 4 blocks

Reconstructed

Image

CHAPTER 3. TVQ WITH ORTHOGONAL POLYNOMIALS

51

3. 5 Experiments and results

The proposed orthogonal polynomials based Transformed Vector

Quantization has been experimented with 2000 test images, having different low

level primitives. For illustration two test images viz, Lena and Pepper, both of

size (256 x 256) with gray scale values in the range (0 – 255) are shown in

figures 3.2(a) and 3.2(b) respectively. The input images are partitioned into

various non-overlapping sub-images of size (4 x 4). We then apply the proposed

orthogonal polynomials based transformation on each of these image blocks and

obtain the transform coefficients ji . All these ji s of each block are then

subjected to scale quantization. The resulting coefficients are re-ordered to 1-D

zig-zag sequence and a subset of corresponding ji s due to energy compaction

of the proposed orthogonal polynomials based transformation are extracted as

described in section 3.4.1. These resulting frequency coefficients are treated as a

vector Ti with dimensionality six and the experiment is repeated for all the sub-

images to form a set of vectors T = {Ti , i = 1, 2,…, k} where ‘i’ represents the

sub-images and ‘k’ represents total number of sub-images.

Then the codebook is designed on truncated scale quantized transform

coefficients with LBG algorithm as described in section 3.4.2. The vectors and

the codebook thus generated are subjected to Vector Quantization to obtain the

index value for each vector Ti corresponding to the sub-images under analysis.

These index values are subjected to entropy coding and are transmitted to the

receiver side. In the decompression process, these index values are decoded and

are used to generate the approximated truncated transform coefficients, with the

help of the codebook that were generated in the earlier stage. These 1-D

transform coefficients are reordered to the original 2-D array after compensating

zeros to the truncated high frequency components. Then the decompressed

original image is obtained with the orthogonal polynomials basis functions as

described in section 2.3.

CHAPTER 3. TVQ WITH ORTHOGONAL POLYNOMIALS

52

(a) (b)

Figure 3.2: Original test images considered for proposed TVQ

(a) (b)

Figure 3.3: Results of proposed TVQ when bpp = 0.25

CHAPTER 3. TVQ WITH ORTHOGONAL POLYNOMIALS

53

Table 3.1: PSNR values obtained with proposed TVQ, DCT based TVQ and the

JPEG baseline for different bpps.

Bit rate

(bpp)

Proposed TVQ

(Dimensionality 6)

DCT based TVQ

(Dimensionality 6)

JPEG baseline

system

Lena Pepper Lena Pepper Lena Pepper

0.25

0.20

0.18

0.16

0.14

0.12

33.09

30.14

29.51

29.02

28.67

27.64

33.22

30.90

29.80

29.47

28.92

28.03

32.49

29.71

28.64

28.14

27.42

26.04

32.53

29.88

29.02

28.44

27.49

26.53

31.62

29.47

28.28

27.02

24.99

21.48

31.86

29.81

28.69

27.26

25.29

21.58

The bit per pixel (bpp) scheme is used to estimate the transmission bit

rate. The performance of the proposed TVQ scheme is measured with the

standard measure Peak-Signal-to-Noise-Ratio (PSNR) as described in section 2.5

with the proposed TVQ. We obtain PSNR values of 33.09dB and 33.22dB for a

bit rate of 0.25 for the input images 3.2(a) and 3.2(b) respectively and the

corresponding resulting images are shown in figures 3.3(a) and 3.3(b)

respectively. The experiment is repeated by varying the bpp for all the 2000

images and the results for the Lena and Pepper images are presented in table 3.1.

In order to measure the efficiency of the proposed orthogonal polynomials

based TVQ, we conduct experiments with discrete cosine transform based

transformed vector quantization. For this comparison, the proposed orthogonal

polynomials based transformation is replaced with discrete cosine transformation

and the other steps are kept unaltered. The experiments are conducted for

different bpp and the corresponding PSNR values obtained are incorporated in

the same table 3.1, for both the input images and the corresponding results for

0.25 bpp are shown in figure 3.4(a) and 3.4(b) respectively. The proposed

CHAPTER 3. TVQ WITH ORTHOGONAL POLYNOMIALS

54

transformed vector quantization algorithm is also compared with the

international standard JPEG type compression algorithm where discrete cosine

transformation and scale quantization are used. For this experiment

normalization and quantization arrays are scaled in JPEG algorithm, to adjust the

compression ratio to the desired level. The experiments are carried out for

varying bit rates for different images and the corresponding results are

incorporated in the same table 3.1. The results of JPEG algorithm for the input

images shown in figures 3.2(a) and 3.2(b) when the bpp is 0.25 are shown in

figures 3.5(a) and 3.5(b) respectively. The graphs of PSNR vs. bpp for Lena and

Pepper images are plotted and the same are shown in figure 3.6(a) and 3.6(b)

respectively.

(a) (b)

Figure 3.4: Results of DCT based TVQ when bpp = 0.25

CHAPTER 3. TVQ WITH ORTHOGONAL POLYNOMIALS

55

(a) (b)

Figure 3.5: Results of JPEG when bpp=0.25

Lena

20

22

24

26

28

30

32

34

.12 .14 .16 .18 .2 .25

Bit rate (bpp)

PS

NR

(d

B)

Proposed TVQ

DCT based TVQ

JPEG

(a)

CHAPTER 3. TVQ WITH ORTHOGONAL POLYNOMIALS

56

Pepper

20

22

24

26

28

30

32

34

.12 .14 .16 .18 .2 .25

Bit rate (bpp)

PS

NR

(d

B)

Proposed TVQ

DCT based TVQ

JPEG

(b)

Figure 3.6: Bit rate versus PSNR comparison of the proposed TVQ with DCT

based TVQ and JPEG

From table 3.1 and figures 3.3, 3.4, 3.5 and 3.6, it is evident that the

proposed TVQ outperforms JPEG and discrete cosine transformation based

TVQ. From these outputs, it is clear that the proposed scheme gives higher

PSNR value with a reasonably good reconstruction quality. It can also be

observed that the quality obtained with JPEG image coding is found to have

degradation at low bit rates. But the proposed Transformed Vector Quantization

produces a stable quality of picture even at low bit rates. [Observe that the PSNR

value that ranges from 31.62dB to 21.48dB and 31.86 to 21.58dB for Lena and

Pepper images respectively for JPEG scheme (from table 3.1)].

3.6 Conclusion

In this chapter a new transformed vector quantization based on orthogonal

polynomials has been proposed for 2-D gray scale images. This technique

combines the features of both transform coding and vector quantization. The

proposed transform coding is based on a set of orthogonal polynomials. The code

book is designed with LBG algorithm that utilizes only few transformed

CHAPTER 3. TVQ WITH ORTHOGONAL POLYNOMIALS

57

coefficients due to the proposed transformation. Training vectors are then formed

as a subset from the image data in frequency domain and is compared with the

code book, to result in the index position of the code book and sent to the

decoder after entropy coding. The decoder has the code book identical to the

encoder and decoding mechanism is a simple table look-up process with

additional null values added to the high frequency samples. These coefficients

are subjected to inverse transform with the help of basis functions of the

proposed orthogonal polynomials transformation to get back the decompressed

image. The performance of the proposed scheme is measured with standard

PSNR value and is compared with DCT based TVQ and JPEG type algorithms.

However, the encoder phase of the proposed VQ scheme uses the full search

algorithm for finding the best match vectors and it leads to increase in searching

time. To overcome this problem, a binary tree based codebook design is

presented in the next chapter.