MPEG-7 Homogeneous Texture Descriptor - Vision … MPEG-7 homogeneous texture descriptor that we...

11
MPEG-7 Homogeneous Texture Descriptor Yong Man Ro, Munchurl Kim, Ho Kyung Kang, B.S. Manjunath, and ~ i n w o o n ~ ~ i r n MPEG7 standardization work has started with the aims of providing fundamental tools for describing multimedia contents. MPEG7 defines the syntax and semantics of de- scriptors and description schemes so that they may he used as fundamental tools for multimedia content description. In this paper, we introduce a texture based image descrip- tion and retrieval method, which is adopted as the homo- geneous texture descriptor in the visual part of the MPEG 7 final committee draft. The current MPEG7 homogene- ous texture descriptor consists of the mean, the standard deviation value of an image, energy, and energy deviation values of Fourier transform of the image. These are ex- tracted from partitioned frequency channels based on the human visual sjktem (HVS). For reliable extraction of the texture descriptor, Radon transformation is employed. This is suitable for HVS behavior. We alqo introduce vari- ous matching methods; for example, intensity-invariant, rotation-invariant andlor scale-invariant matching. This technique retrieves relevant texture images when the user gives a querying texture image. In order to show the prom- ising performance of the texture descriptor, we take the ex- perimental results with the MPEG7 test sets. Experimen- tal results show that the MPEG7 texture descriptor gives an efficient and effective retrieval rate. Furthermore, it gives fast feature extraction time for constructing the tex- ture descriptor. Clanuscnlil rcceivcd Octolm 7.Z(XX): rc\isc;cd May 10.20nI. Xiny hlan Ro ntx1 1 In Kynf Kingaw \\it11 Ihc Intinn'ah(111 and Commr~nic:air,nc Unn-cmit): Tacio~i. Kr*c:~. (plni~ic: +X242 Rhh 6120. c-mail. wnffc icu.ac.kl-) hluncliurl Kiln and .linwoong Klm an. ~~itli tlic Hrnsdcastinp k4ctli;i 'Tcrlinolo~ Dcpamnoit. I(TR1. Taclon. Kolra. (plicmc +82 42 Shh 5847. e-mail: n~kimi~rchi.~c.kl-) B.S. Ma1li11n:1111 is \villi tlie Univctsity of Calitiimin. Slnh B:~rIi;a;i. ITS/\. ((~lionc: 1 805 71 12. c-mall: maniff? ccc.ricsh.~lu\ 1. INTRODUCTION Recently. there has been an ovenvhelming increase in the aiiiount of digital multimedia information going over the Inter- net and broadcasting systems. And users need new method to organize. manipulate and tl+ansmit tlie data they want. The culrent technologies for representing multiinedia in tlie fo~ins of texts liave many limitations. Tiey cannot efficiently represent and remcve various types of multimedia contents. Also. intetnatioiial standards such as JPEG. MPEG-I, MPEG- 2 and MPEG-4 have been developed only for compression of data. These standards are crcated for efficient storage and trans- mission, not for tlie repi-esentation of tlie contents. At tlie 36th MPEG Chicago lneeting in September 1996. MPEG tnenibei-s first discussed an "A~idiovisual Content Dc- sciiptioii 11iterface" fol- cficieiit representation of nlultimedia infolination. They wanted to make an inteiiiational standard that became MPEG-7. Its official name is "M~iltiniedia Con- tent Description Inteifacc" of TSOIIEC JTC 1 SC29NG 1 1 and work began at tlic 37th MPEG Maceio meeting in November 1996. MPEG-7 standardi7ation reached the final coiiimittee draft (FCD) level at the MPEG Singapore meeting in March 2001. This is a technically stable stage. MJ?EG-7 standard specification defines the syntax and sc- mantics of dcscribiiig tlic multi~nedia coiitents and consists of 7 Pam: Systcms. Dcsaiption Dcfi~~itioii Lai~giagc (DDL). Vis- ual dcsciiptor; Audio dcscriptoc Multin~ed~a Desciiption sclieines (MDS), Refcrcncc softwva~c. and Conformance tcsting. 111 the visual pait of the MPEG-7 standard. visual descriptors are spccificd as noniiativc descripto~s. basic dcscriptol-s, and dcscripto~x for localizatro~i. Norinative desciiptor describes tlie color. shape. texture and motion features ofvisual data [I], [2]. In this papci-. we inh-oduce a texturehascd image de~ciiptimi ETRl Journal, Volume 23, Number 2, June 2001 Yong Man Ro et a/. 41

Transcript of MPEG-7 Homogeneous Texture Descriptor - Vision … MPEG-7 homogeneous texture descriptor that we...

Page 1: MPEG-7 Homogeneous Texture Descriptor - Vision … MPEG-7 homogeneous texture descriptor that we describe in this paper had been severely tested and conipared with other proposed texture

MPEG-7 Homogeneous Texture Descriptor

Yong Man Ro, Munchurl Kim, Ho Kyung Kang, B.S. Manjunath, and ~ i n w o o n ~ ~ i r n

MPEG7 standardization work has started with the aims of providing fundamental tools for describing multimedia contents. MPEG7 defines the syntax and semantics of de- scriptors and description schemes so that they may he used as fundamental tools for multimedia content description. In this paper, we introduce a texture based image descrip- tion and retrieval method, which is adopted as the homo- geneous texture descriptor in the visual part of the MPEG 7 final committee draft. The current MPEG7 homogene- ous texture descriptor consists of the mean, the standard deviation value of an image, energy, and energy deviation values of Fourier transform of the image. These are ex- tracted from partitioned frequency channels based on the human visual sjktem (HVS). For reliable extraction of the texture descriptor, Radon transformation is employed. This is suitable for HVS behavior. We alqo introduce vari- ous matching methods; for example, intensity-invariant, rotation-invariant andlor scale-invariant matching. This technique retrieves relevant texture images when the user gives a querying texture image. In order to show the prom- ising performance of the texture descriptor, we take the ex- perimental results with the MPEG7 test sets. Experimen- tal results show that the MPEG7 texture descriptor gives an efficient and effective retrieval rate. Furthermore, it gives fast feature extraction time for constructing the tex- ture descriptor.

Clanuscnlil rcceivcd Octolm 7.Z(XX): rc\isc;cd May 10.20nI.

Xiny hlan Ro ntx1 1 In K y n f Kingaw \\it11 Ihc Intinn'ah(111

and Commr~nic:air,nc Unn-cmit): Tacio~i. Kr*c:~. (plni~ic: +X242 Rhh 6120. c-mail. wnffc icu.ac.kl-)

hluncliurl Kiln and .linwoong Klm an. ~ ~ i t l i tlic Hrnsdcastinp

k4ctli;i 'Tcrlinolo~ Dcpamnoit. I(TR1. Taclon. Kolra.

(plicmc +82 42 Shh 5847. e-mail: n~kimi~rchi .~c.kl- )

B.S. Ma1li11n:1111 is \villi tlie Univctsity of Calitiimin. S l n h B:~rIi;a;i. ITS/\. ((~lionc: 1 805 71 12. c-mall: maniff? ccc.ricsh.~lu\

1. INTRODUCTION

Recently. there has been an ovenvhelming increase in the aiiiount of digital multimedia information going over the Inter- net and broadcasting systems. And users need new method to organize. manipulate and tl+ansmit tlie data they want.

The culrent technologies for representing multiinedia in tlie fo~ins of texts liave many limitations. Tiey cannot efficiently represent and remcve various types of multimedia contents. Also. intetnatioiial standards such as JPEG. MPEG-I, MPEG- 2 and MPEG-4 have been developed only for compression of data. These standards are crcated for efficient storage and trans-

mission, not for tlie repi-esentation of tlie contents.

At tlie 36th MPEG Chicago lneeting in September 1996. MPEG tnenibei-s first discussed an "A~idiovisual Content Dc- sciiptioii 11iterface" fol- cficieiit representation of nlultimedia infolination. They wanted to make an inteiiiational standard that became MPEG-7. Its official name is "M~iltiniedia Con-

tent Description Inteifacc" of TSOIIEC JTC 1 SC29NG 1 1 and work began at tlic 37th MPEG Maceio meeting in November 1996. MPEG-7 standardi7ation reached the final coiiimittee draft (FCD) level at the MPEG Singapore meeting in March 2001. This is a technically stable stage.

MJ?EG-7 standard specification defines the syntax and sc- mantics of dcscribiiig tlic multi~nedia coiitents and consists of 7 Pam: Systcms. Dcsaiption Dcfi~~itioii Lai~giagc (DDL). Vis- ual dcsciiptor; Audio dcscriptoc Multin~ed~a Desciiption sclieines (MDS), Refcrcncc softwva~c. and Conformance tcsting.

111 the visual pait of the MPEG-7 standard. visual descriptors are spccificd as noniiativc descripto~s. basic dcscriptol-s, and dcscripto~x for localizatro~i. Norinative desciiptor describes tlie color. shape. texture and motion features ofvisual data [I], [2].

In this papci-. we inh-oduce a texturehascd image de~ciiptimi

ETRl Journal, Volume 23, Number 2, June 2001 Yong Man Ro et a/. 41

Page 2: MPEG-7 Homogeneous Texture Descriptor - Vision … MPEG-7 homogeneous texture descriptor that we describe in this paper had been severely tested and conipared with other proposed texture

and retrieval method which we proposed and adopted as thc kIo~nogeneous texture descriptor in the Vis~lal pa t of the MPEG-7 FCD. (Xu- proposal was adopted.

The texture information of an i~l~age is a fiu~ctamental visual feature, which has been studied during the last decade to ana- lyze inlages in the areas of niedical ~maging mid satellite imag- ing, etc. [3]-[s]. This contains shuctuu-eness, regllarity, direc- tionality and roughness of inlages, which are impo~tant proper- ties of the content-based indexing of the image [6].

Previous works such as probability distribution of pixels [3], directional filtering [3] and Markov widom field have been studied. More recently, spatial Gabor filters and wavelet trans- formation have been studied to extract tcxhu-e inibnnation. In [5], Gabor, Pyramid shuctured Wavelet Tra1sfol-111 (PWT), Tree sh~lctlu-ed Wavelet Transform (TWT), and Multiresolu- tion Simnultaneous A~~toregress~ve Model (MRSAR) methods have been comnpared. In that paper, Gabor and MRSAR metli- ods show good perfomlance of relevant texture image retrieval. However, the methods require high co~nputational complexity to extract tlie texture infonnation. The MPEG-7 homogeneous texhu-e descriptor we ulvented is efficient not only for comput- ing texhre feahu-es but also u1 representing texhu-e infonnation. 'Illrough the core experirnents in tile MPEG-7 Visual group, the MPEG-7 homogeneous texture descriptor that we describe in this paper had been severely tested and conipared with other proposed texture descripto~s in tenns of computation complex- ity and retrieval accuracy, It outpe~lbllned the other by sliowulg fist feahu-e extraction aid compact representation of texture in- fmnation. It provided higher retrieval accuracy for the testing data sets. Therefore, the homoge~ieous texture descriptor de- scribed in tllis paper was selected as the nonilative MPEG-7 Iio~~ioge~~eous texture descriptor in the Visual part of the MPEG-7 final conunittee draft [I 11-[14].

In this papel; we present technical details of the MPEG-7 honiogeneous texhu-e descriptor and its feahu-e extraction method. In addition to the feature exhaction nietliod, silnilarity ~neas~uing criteria are presented for laation-, scale- and inten- sity-invariant mnatchings [I 51-[I 81.

The homogeneous texhu-e descriptor in this paper consists of the mean and stan&asd deviation values of an irilage. It also in- cludes the energy and energy deviation values of the Fourier mtsfonn of the image. In order to explain the texhu-e represen- tation based on energy and energy deviation features, the tex- ture feature extraction is explained in Section 11 . Texture in- dexing and rehieval algorithms are then presented in Section 111. Section III also addresses various image matching criteria for intensity-, rotation- andlor scale-invariant matching in rebieval. In Section IV, the experimental results are provided for the MPEG-7 data set of die texture experiment.

42 Yong Man Ro et a/.

11. TEXTURE DESCRIPTOR EXTRACTION ALGORITHM FOR THE MPEG-7 HOMOGE- NEOUS TEXTURE DESCRIPTOR

I. H L I I I ~ Visual System for the Texture Descriptor

Recently, texture-feahdng and description techniques based on the HVS have been proposed [4]. Texture feat~u-ing based on the HVS corresponds well to sonle results fi-om psycho- physical experirnents. In these experiments, the response of the visual coltex is huned to a band-liliuted portion of the fre- quency donain. The human brain deconlposes the spectra into perceptual ch~u~nels that are bands in spatial frequency [5], [7]. For texhu-e featuring, the best sub-band representation of HVS is a division of tl~e spatial frequency domain in octave-bands (4-5 divisio~ts) along tile radial direction and in equal-width angles along the angular direction. These sub-bands are sym- nletrical with respect to tile origin of the Polar coordinate. In tl~is section, a ti-equency layout is designed. The frequency lay- out allows extracted texture inforniation to be matched with human perception system. The frequency layout consists of sub-bands. In these bands, the texture descriptor components such as energy and energy deviation are extracted.

According to die HVS properties mentioned above, the sub-

bands a-e designed by dividing the frequency domain to com- pute texture feature values. The frequency space from which tlie texture descriptor in the image is extracted is pmtitioned in equal angles of 30 degrees along the angular direction and in octave division along the radial direction. The sub-bands in the frequency domain are called feature channels indicated as C, in Fig. 1. The frequency space is therefore partitioned into 30 feature channels as shown in Fig. I . In the normalized fre- quency space (0 < (0 < 1) , the nonnalized frequency co is

given by w = R/R ,,,.,, . R ,,,,, is the maxinlun~ frequency value of the image. The center frequencies of the feature chan- nels are spaced equally in 30 degrees along the angular direc- tion such as 0, = 30" x r . Here r is an angular index with I . E [0,1,2,3,4,'5} . The angular width of all feature channels is 30 degree. In tlie radial direction, the center frequencies of the feature channels are spaced with octave scale such as o, = p,, . 2 ' , s E {0,1,2,3,4} where s is a radial index and

o , is the highest center frequency specified by 314. The oc- tave bandwidth of the feature channels in the radial direction is written as B, = B,, .2-' , s E {0,1, 2, 3,4} where B, is the largest bandwidth specified by 112.

Figure 1 shows the two-dimensional (2D) frequency layout configured by the above divisibn scheme. As shown in Fig. 1, each partitioned region corresponds to a band-limited portion of the frequency donain that is the response of the visual cor- tex in the HVS. Therefore, the region can be denoted as a

ETRl Journal, Volume 23, Number 2, June 2001

Page 3: MPEG-7 Homogeneous Texture Descriptor - Vision … MPEG-7 homogeneous texture descriptor that we describe in this paper had been severely tested and conipared with other proposed texture

channel to transfer the response of the visual cortex. The chan- nels located in the low frequency areas are of smaller sizes while those of the high frequency areas are of larger sizes. This corresponds to the human vision that is more sensitive to the change of low fi-equency area. Also, note that half of the entire frequency space is used because images are assumed to be of real value.

" c o ~ ~ n l slice theo~.ei~i'' in which the 1 D Fourier transform of a projection of image at angle O equals the slice at angle Q through the 2D Fourier transform of that image (see Fig. 3).

P,J (R) Projection

R

Fig. 1. Frequency region division with HVS filter. I Fig. 2. Radon transform scheme. Image,f(v.l~) is transformed to

p, ( R ) in Radon space (R, 81. 2. Data Sampling in Feature Channel

As shown in Fig. 1, the channel layout in the spatial fre- quency domain is center-symmetrical. Since the partitioned frequency regions are relatively small compared with those in the high frequency regions in the Cartesian coordinate system, the frequency samples are sparse in the low frequency regions where the texture'infomation is insufficient. In order to avoid this, we employ Radon transform for images, which allows Fourier transfoml of image in Cartesian to be leepresented in the Polar coordinate system. Using the Radon transformation, 2D image can be transformed to onedimensional (ID) projection data, i.e.. Cartesian space (x, j,) will be mapped to Radon space (R. 6) as shown in Fig. 2.

The line integral along the line L(R, 8 ) at angle Qin coun- terclockwise direction from y-axis and at a distance R from the origin can be written as

= I: ~m/(x.Y)G(xcosQ+!~sin 8-R)d.ds. ( I )

where f (.r. ?) is an image function, R is projection axis and Go is delta hnction. The hnction p, ( R ) is a projection, since it collapses a 2D image to a ID projection for each angle. The co~nplete collection of line integrals is called the Radon transform of f (x, JJ) and also called the Sinogam. The fre- quency properties in Radon transformation can be explained by

Sinogram 2D Fourier transform

Fig. 3. Relationship bctween sino_grarn and 2-dimcnsional fourier domain.

One-dimensihnal Fourier transform of a projection can be written as

where o = ,/- and 0 = tan '(o, / ID,).

The Radon h-ansfom~ is suitable for the HVS since each cen- tral slice in Fourier domain is fit to the data representation in the HVS frequency layout mentioned previously. Data acquisi- tion in the HVS-based frequency layout is done with polar- oriented sampling scheme.

ETRl Journal, Volume 23, Number 2, June 2001 Yong Man Ro eta/. 43

Page 4: MPEG-7 Homogeneous Texture Descriptor - Vision … MPEG-7 homogeneous texture descriptor that we describe in this paper had been severely tested and conipared with other proposed texture

Figure 4 shows a smplilg grid structure of a Polar fre- quency domain after the Radon trmsform followed by its FOLL- rier tranqform. As shown in the figure, sampling density is dense in the low and middle fi-equency areas while sparse in the high frequency area. This property corresponds to the hu- nlan visual properties such that the human vision is more sensi- tive in the low kquency area than in the high ti-equency area. This property supports that the Radon transfornution of image is suitable to the HVS 1151.

rection and 5 partitions in the radial direction. For the frequency layout shown in Fig. 1, D~ is a constant

value of 15' / in the angular direction. In the radial direction, a,#, is dependent on the octave bandwidth and is written as

Tables 1 and 2 show parameters in the feature channels and the Gabor fhctions.

the pass band edges of the ideal tilters between channels, Ga- bor filter banks are instead used to consh.Llct the ti-equency lay- out. By applying the Gabor filter banks, the channels are over-

( B,, )

lapped so that the channels can affect neighbor channels each other (each other is fedundant) at the boundary areas. 11e Ga- bor function defined for Gabor filter banks is written as

Fig. 3. Sampling grid of a Polar fkequency donuin after the Radon Table 1. Parameters of octave band in the radial direction. transform followed by Fourier transfomi.

- (6 - 0, )? Table 2. Parameters of angular band in the angular direction.

where G,, , (w,6) is Gabor hlction at s-th radial index and r-th angular index. o,,\ and o,, are the standard devia- tions of the Gabor function in the radial direction and the angu- lar duection, respectively. The standard deviations ofthe Gabor function are determined by touching the Gabor function with its neighbor functions at half the maximum (112) in both radial and angular duections. Figure 5 shows the Gabor filters 011 top ofthe frequency layout, which are 6 partitions in the angular di-

44 Yong Man Ro et a/. ETRl Journal, Volume 23, Number 2, June 2001

4

3 - 64

In signal proces~ing pe~spective, the frequency layout in Fig. 1 is actually realized with a set of ideal filter banks that have abmpt channel boundaries. In order to relax the sharpness of

3

3 - 3 2

Radial index (s) Centre

frequency

(0,)

1

3 - 8

0

3 - 4

2

3 - 16

Page 5: MPEG-7 Homogeneous Texture Descriptor - Vision … MPEG-7 homogeneous texture descriptor that we describe in this paper had been severely tested and conipared with other proposed texture

3. MPEG-7 Hoinogeneous Texture Descriptor

To extract the texture feahu-e values, wc take the Radon transfo~in on an image and subsequent ID Fourier b-ansfonn on the data. l i en . we can obtain a central slice of ~ ( ( 0 . 6 ) in 2D fiequaicy domain. The texture descriptor consists of fca- ture values extracted fiom each channel shown in Fig. 1. In tliis paper; the texhire desci-iptor colnponents are tlie fi~st and scc- ond moments of ena-gy in channels. i.e., enapies and energy deviations. The energies and eriel-gy deviations that constitute tlie texture descriptor are written as [el . el ,K . e,,,] and [d, .d2.K .d,,], respectively. Hcre tlie indexes from I to 30 indicate the feahre channel numbcrs.

Bascd on the fiequency layout (partitioned fiequency do- main) and the Gabor fi~nctions. tlie energy e, of tlie i-th fea- ture channel is defined as the log-scaled sum of the squares of Gabor-filtcrcd Fourier transfo~in coefficients of an iriiagc:

at the base layer, and

at the cnha~iccment layer. l i e texture descriptor can be represented at two different

layers: base laya and enhancement layer. The textu11-e descrip- tor only consists of f,,, , f,,, , and 30 energy valucs (c, ) of tlie Fouricr transform of the image. In the enhancement layer. the texture descriptor (additioiially o ~ ) adds 30 (or additional) energy deviation valucs of the Fourier transform of tlie iniape in tlie texture descriptor vector. Tlie layering schenie of tlie tex- ture descriptor provides scalability of representing image tex- ture depending upon applications. For the delivery of limited bandwidths, only texture descriptor components at the base

e, =log[l+p, l , (5' layer may be transmitted. Also. fast matching can be per- fonned at tlie base layer by satisfying retrieval acctu-acy.

where

:OO" 4. Quantization of Textul-e Descriptor

17, = Z[G ,.<,, ( o . ~ ) . lo1 . F(~,.Q)I' . (6) 11-1) + hi this paper, the quantization levels of the texture descriptor

values set to 256. Eight bits are assigned and used for linear

where Iwl is Jacobian term between Cartesian and Polar fie- quanti72tion of eacli descriptor value. n i e linear quanti7ation

. used in the paper is mitten as quency coordinates and can be writtcn as lo/

~ ( o , 0 ) is Fourier transfmn of the image

suliimation is taken over the entire frequency domain except D~,, ,~,~, = D,,,,,,,,r,r,,,, - Dl,ltn x q - Ici~el x I Pll1.1, - P1111" q - lei~el

+ Dmtti tlic DC component. Note that i = 6 x .s + I . + l wlierc s is fill,.,, - b'llllll

(11) , - - , tlic ladial index and I- is the radial index. l i e energy deviation

d of tlie i-tl~ feature channel is defined as the log-scaled stan- whcre D,,,n,,, is a quantized descriptor value, D,,,,,,,n,,, is the - dard deviation of tlie squares of Gabor-filtered Fourier trans- of fl,,,,, Ptll~t~ are the lnaximtlln

arid lninilnu~n values of fcahlr-es in tlie database. Note that fo~m coefficients of the image:

y Y I ~ ~ e l is set to 255.

wvliere

After the quanti~ation, eacli feature has 1 byte in size. With 4 = log[l+ q, I (7' many MPEG-7 colr experiments. I byte was good enougli not

to lose the textire infonilation. Fiirtlicl; an entropy coding is possible to rcduce tlie bits morc. But, tliis can be considered for coding efficiency in fi~ltlier work. Total texture dcsniptor

I Ihll" c ~{[~,~,co.a)-l(0l~co.e,E-,7,) . ' (8, lengjli is therefore reduced to 32 bytes for base layer and 62

l!, = r , n l n II hytcs for the ailiancement layer.

Fu~tlier.

,f/,<. ) and are added

b~ighhiess info~~iiation of texh~re (riiean denotcd by standard deviation ( f,, ) of thc entire iriiagc pixels

111. RETRIEVAL ALGORTTIIM FOR THE TEX-

as tlie texture feahlre valucs m tlie textulr descriptor. TURE DESCRIPTOR

Finally. the image intensity avcrage f,, . standard deviation 1. ~ i ~ ~ ~ i l ~ r i ~ M~~~~~~~~~~~~ f , . energies e, , and energy deviat~ons d, of tlic channels

colistih~tc tlic Iio~nopeneous texture descrrptor (TD) in the or- To rehieve similar texture images for a quely, a matching

der as follows: procedure sliould be pcrfonned. The matching procedure is as

ETRl Journal, Volume 23, Number 2, June 2001 Yong Man Ro et a/. 45

Page 6: MPEG-7 Homogeneous Texture Descriptor - Vision … MPEG-7 homogeneous texture descriptor that we describe in this paper had been severely tested and conipared with other proposed texture

follo\vs. First, Radon tril~l~fo~m of qi~erying image is per- formed so that 1D-projection sigials are obtained. Using "cevr- rtul s1ic.e rlleor-em", fiequcncy data in the polar space are ob- tained. For the texhu-e desctiptol; enerLy and deviation men- tioned in the previous section are calculated. Then, the similar- ity between a quelying image and images in the database is meas~red. The feahre of a quetying image i is denoted by TD, while the teahue of an imagej in the database by TD, . The sinlilillity meas~red by calculating the distance between the two feahre vectols is as follows:

original image. By using the rotational property, we propose rotation invariant similarity nlatching method. We first measure the distance between texture descriptor vectors in the database and a quelying texture descriptor vector by shifting the query- ing texture vector in the angulx direction such as

d(i, j ,m&) = distance (TD, ( k ) I ,,,+, TD, (k)) (15)

where 4 = 30 degrees. Then, for rotation invariant descriptor, distance is calculated as

d(i, j) = distance (TD, , TD, ) d(i, j ) = ~ninilnumn of (tl(i, j , t n $ ) 1 H I = 1 to 6 ) . (16)

ul(k)[TD, (k) - TD, ( k ) ] (12)

5. Layered Texh~re Descriptor

where ~ ~ ( k ) is the weighting factor of ti-th descriptor value. The nolmalization values a (k) are standard deviations of tex- ture descriptor values for a rekrence chtabase. (During MPEG- 7 core experiments, T1 dataset was used as refet-ence database). The weighting pmle t e r 11(k) and the normalization value

a (k ) are calculated in advance so that they are independent or1 the database. These values could be obtained n pt.iori at die begiruing of establishing the database.

2. Intensity-Invariant Matching

For the intensity invari;u~ce that is usually required for 1i10st applications, f ,, is eluninated fi-0111 the feature vector when the similarity measurement is perfomled.

For a11 efficient storage or transmission, 62 features can be assigned with priority. Nanely, with linlited storage or band- width, the texture descriptor can be reduced without degrading retrieval accuracy significantly. Especially, in wireless Intenlet which has poor network environment, only a part of the texture descriptor components could be transmitted to the MPEG-7 database. In this case, the entire texture descriptor components are not used for the content-based indexing. To meet the above requirements, the layered configuration of the texture descrip- tor is helpfil for the better retrieval perfol~nance. The texture descriptor is layered as follows:

3. Scale-Invariant Matchng TD I""'- 1," C! is the texture descriptor at the base layer ,

F~~ a givell queryillg image, querying image is zoollled in which is represented with the first and second moments of the

and out with N difTere11t rooming factolr, The distance ' / ( I , j ) image pixels ellergJ' (el ) between the querying image i i d the inlagc j indexed in data- base is obtained by

T ~ h ~ ~ ) a , - / t l l ~ ~ =[fr)(?f;n?e,,e,,K &;,,I, (I8)

d(i, j , n ) = distance (TD, (k) I,, , TD, (k) ) TD ~ ,ho t ,~ .~ * , twru - hll.~,, is an extended texture descriptor at the base layer to enhance'the retrieval efficiency, that is, it uses fill fea-

d(i, j ) =minimum of { d ( i , j.17) I n = I to N) (14) hre values in the descriptor. It can be written as

where N is the number of scaled (~00111-in and zoonl-out) ver- TDC~;/IL,~IL~,IIIC,U- ~ , , Y v = [&, ~ ~ ~ ~ e ~ , e ~ , ~ -e,,,d,,d?,K ,4.I.

siom of the querying feature. N is usually 3 so for exanple, the original and two scaled versions of the querying image are (19)

30% zoom-in and 30% zoom-out. One can iuse different zoom- in and zoom-out.

IV. MEASUREMENT OF RETFW5VAL 4. Rotation-Invariant Matching PERFORMANCE

Since the frequency space division for the texture descriptor ' To ven@ the performance of the texture descriptor men- is made in the polar donlain as show11 in Fig. 2, the texture de- tioned above, experiments have been performed with tile test scriptor of a rotated image is an angular-shifted version of the data sets for the hon~ogeneous texture descriptor in order to

46 Yong Man Ro et a/. ETRl Journal, Volume 23, Number 2, June 2001

Page 7: MPEG-7 Homogeneous Texture Descriptor - Vision … MPEG-7 homogeneous texture descriptor that we describe in this paper had been severely tested and conipared with other proposed texture

measure feature extraction time and retrieval accuracy. Re- matrix size of 5 12x5 12 is divided into 16 non-overlapped parti- trieval performance of the texture descriptor is measured by re- tions. In the T2 data set, one image has 15 ground truths and triwal rate (RR) which is a ratio between the number of rele- the number of querying images is 52 images. The relevant im- vant images and the number of ground truth image for a given ages are ground truth images which belong to the first 15 im- querying image, Similar images, which are of the same number ages with minimum distance. Figure 7 shows one example of of the ground truth, are selected by measuring distance from the retrieval by a querying image with the T2 data set. the querying image. The relevant images are those belonging to the ground-truth Images among the similar images. The RR can be written as

# of relevant retrieved images (20) RR = &~-,.sbh.&d.,Lc,~m--.~jj

# of ground trzltli

The average retrieval rate for a data set (AVRR) is, therefore, denoted by

[c,,hct " "" ' AVRR = numher of quer?, (21)

The MPEG-7 test data sets for the texture descriptor have 7 different kinds of test data sets, which are TI. T2, T3, T4. T5, T6, and T7 data sets. The following subsections explain MPEG-7 test data sets used in the core experiments of the ho- mogeneous texture descriptor in detail. Fig, 6. An example of retrieved images in TI data set. The upper-

left image is a query image. Tlie other 15 images are retrieved images for the query

I . TI Data Set

TI data set contains texture pattern images which have been used popularly as a test image set for the texture experiments m many Ilteratures. It consists of 1856 images with matrix si7e of P

- I*g 128x128. 1856 images are made from 116 Brodatz images with matrix size of 5 12x5 12. Each Brodatz image with matrix size of 512x512 is divided into 16 non-overlapped partitions, i.e., 16 images with matrix size of 128x128. In the TI data set, one image has 1 5 ground truths since 1 6 images are generated from one Brodatz pattern. Therefore, the relevant images in (20) belong to the ground truth images as well as the first 15 re- hieved images having minimum distance. The querying im- ages for T2 data set are the original patterns. So the number of query in (20) is 11 6. Figure 6 shows an example of the retrieval by a querying image with the TI data set. As shown in the fig- ure, an image at top-left is the query and remaining 15 images are retrieved ones. The right side of the querying image has the Fig. 7. An example of retrieved images in T2 data set. The upper-

smallest distance value. left image is a query image. The other 15 images are retrieved images for tlie query.

2. T2 Data Set 3. T3 Data Set

T2 data set consists of real patterns taken from outdoor and indoor scenes. It consists of 832 images with s i x of 128x128. T3 data set is the rotated version of the TI and T2 data sets. Like the TI data set, 832 images in T2 data set are made from Fifty five original images with matrix size of 512x512 are 52 irnages with matrix size of 512x5 12 such that an irnage of taken from the TI and T2 data sets such that 30 patterns are

ETRI Journal, Volume 23, Number 2, June 2001 Yong Man Ro eta/. 47

Page 8: MPEG-7 Homogeneous Texture Descriptor - Vision … MPEG-7 homogeneous texture descriptor that we describe in this paper had been severely tested and conipared with other proposed texture

from the T1 data set and 25 pattems from the T2 data set. Then, the 55 images are rotated by 10, 15,20,30,40,50, 70, 75, 80, 100, 110, 130, 135 140, 160, and 170 degrees. Finally, the T3 data set is constructed by taking 128x128 size image at arbi- trary position liom the rotated images. So the total number of image in the T3 data set is 880. To evaluate the performance with the T3 data set, RR and A KW are measured. The number of ground truth is 16 and the querying images are 55 images which are rotated by 30 degrees. Figure 8 shows one example of the retrieval by a querying image with the T3 data set.

Fig. 8. An example of retrieved images in T3 data set. The upper- left image is a query image. The other 15 images are retrieved images for the query.

Fig. 9. ,411 example of retrieved images in T4 data set. The upper- left image is a query image. The other 15 images are retrieved images for the query.

ing inlage with the T4j data set.

5. T5 Data Set

T5 data set includes images from CorelO album. The ground truth is selected by t h g similar images with the querying im- age. The querying images are chosen so that they have rela- tively large texture patterns among the data sets. The number of chosen queries is 16. Total 2400 images constitute the data set. Fig. 10 shows an example of the retrieval by a querying image with the T5 data set. The query in the figure has 4 ground truths.

4. T4 Data Set

T4 data set is the scaled version of the T1 data set. One hun- dred and sixteen onginal images with matnx size of 5 12x5 12 are taken from the T1 data set. Then, the 1 16 images are scaled up and down by 5 %, i e ,95%, 100% and 105% scaled images are obmed. Then, 128x 128 sue of images are taken at arbl- tray posihons from those scaled images and composed of a data set whch is called as T4a data set. Next Images w~th scal- ing up and down w~th 10% are added onto the T4a data set and 90%, 95%, loo%, 105% and 110% scaled unages are comn- posed of T4b data set. Above procedure are repeated unhl the scales reach 50% and 150%. Then, 10 data sets are generated fi-om T4a to T4j. The T4j data set is composed of 50% to 150% scaled inlages wth ~ncren~ent of 5% scalmg factor. So the total ntul~ber of mlages for the T4j data set IS 2436.

RR and are lneasured T4a T4j data r e 'Fig. 10. An example of retrieved images in T5 data set. The spectively. 116 images with 1000/0 scaling factor are querying uooer-left image is a ouew image. The other 15 images . . - . - - images. Figure 9 shows an example of the retrieval by a query- are retrieved images for the query.

48 Yong Man Ro et a/ . ETRl Journal, Volume 23, Number 2, June 2001

Page 9: MPEG-7 Homogeneous Texture Descriptor - Vision … MPEG-7 homogeneous texture descriptor that we describe in this paper had been severely tested and conipared with other proposed texture

6. T6 Data Set

T6 data set consists of aerial images with 34,000 images of 128x128. For a query, ground truth is determined by taking similar images. Figure 11 shows an example of the retrieval by a querying image with the T6 data set

Fig. 11. An example of retrieved images in T6 data set. Upper- left image is a query image. The other 15 images are retrieved images for the query.

Fig. 12. An example of retrieved images in T7 data est. Upper-left image is a query image. The other 15 images are retrieved images for the query.

7. T7 Data Set

T7 data set is both scaled and rotated version derived ti-om T1, T2, and T5 data sets. Seventy original images with matrix size of 5 12x5 12 are taken from the T 1. T2 and T5 data sets

such as 30 patterns h m the TI data set, 25 patterns from the T2 data set, and 15 patterns from T5 data set. The 70 images were rotated with 0,8,25,55, 107, 131, and 174 degrees. Then, the rotated images are scaled with 90%, 80%, 70%, 60% and 50%. Finally, the T7 data set is constructed by taking 128x128 size image at arbitrary position from both rotated and scaled images. Total number of images in the T7 data set is 2400. The RR and AVRR are measured with 34 ground truths. The queries are 70 images with no rotation and 70% scaling. Figure 12 shows an example of the retieval by a q u q n g image with the T7 data set.

V. EXPERTMENTAL RESULTS

To verify the performance of the MPEG-7 texture descriptor, experiments were performed with test data sets of the homoge- neous texture descriptor. These are T1, T2, T3, T4, T5, T6, and T7 data sets. The constitution of databases and paformance test procedure are mentioned in the previous sections. Table 3 shows the average retieval rates for the texture descriptor over TI, T2, T3, T4, and T7 data sets. As we can see, more than 75 ?4 of AVRR have been achieved for T1 data set. The per- formance of the proposed methods can be compared to the re- ported results in the literatures. This is because the T1 data set is widely used for texture description-experiments. Our results were the best among participants of the MPEG-7 texture-core expeliments. Furthermore, as shown in T3, T4 and T7 data sets. our proposed method shows good results for rotated andfor scaled images.

In Table 4, experimental results are shown to verifL the effec- tiveness of the scalability of the layered feature descriptor for TI data set. Scalable representation of feature description (the meaning of allow for is not fit here) provides the flexibility for hmsmission bandwidth and database storage. As sliowr~ in Ta- ble 4. 76.39 ?h of AVRR was obtained with the base layer in the T1 data set. Only 32 components of the descriptor were used. 77.32 % of AVRR was the result for the enhancement layer. 62 comp6ncnts of the descriptor were used. For the TI data set, half of the description size codd be saved with only about 1% loss of AVRR.

Furthermore, thc texture descriptor is easy to compute be- cause it is directly exb-acted in the frequency domain. We measured feature extraction time in a PC (Pentiurn I1 system with a 400 MHz CPU and an NT operating system). It takes around 0.1 4 seconds per one image query (1 28x128 image size).

Table 5 shows a comparison of the average rehieval I-ates with those of other texh~re descriptor extraction methods which are available In literatures. The average retrieval rates for other methods are refen-ed to in [5]. There the same experiments

ETRl Journal, Volume 23, Number 2, June 2001 Yong Man Ro et a/. 49

Page 10: MPEG-7 Homogeneous Texture Descriptor - Vision … MPEG-7 homogeneous texture descriptor that we describe in this paper had been severely tested and conipared with other proposed texture

Table 3. AVRR on the unified texture descriptor.

Table 4. AVRR at the first layer and the second layer,

A VRR (%)

77.32

90.67

92.00

86.11

87.50

88.94

87.07

84.91

84.55

83.87

82.60

79.60

76.72

60.46

75.18

78.66

Data set

T 1

T2

T3

were performed &th the same database of the T1 data set. The AVRR was reported as 74.37% in the TI data set using the Ga- bor spatial filtering method. And it was reported as less than 70% in the TI using wavelet related methods. These are the pyramid wavelet transform inethod (PWT) and the tree wave- let transform inethod (TWT) [5 ] , [20]. With this T1 data set, the proposed algorithnl of tile texture descriptor extraction gives 77.32% of A m .

T4

Layer

Base

Enhancement

Table 5. AVRR on Brodaz albun~.

T4a

T4b

Tlc

T4d

T4e

T4f

TJk! T4h

T4i

T4j

MPEG-7 Texture

descriptors

T5

T6

T7

Features used

i;,~. .f.i,,, ?((I). . . . ~ ( 3 0 )

.fbc. ji~, e(O), . . . , c(30/ , ed(1). . . . , ed(30)

Vl. CONCLUSIONS

Texture is one of the salient feahtres representing image con-

A VRR(%)

76.39

77.32

tents. In this paper, we present a texture description method for images. These feature vectors are made up of an image inten- sity mean, a standard deviation 30 energy values and 30 energy deviations. The Polar frequency domain is partitioned based on the human visual system. From the feature channels within this domain, we can extract the energy values and energy devia- tions. We have shown this to be a very effective texture description. For fast and reliable feature extraction, Radon transform is used to obtain Fourier transform of the image in the Polar domain. Radon transform provides dense sampling in low frequency regions and sparser sampling away from the o r i g ~ of the Polar frequency domain. Thls is well suited to the Munan visual system. The Human visual system is more sen- sitive to signal variation in low frequencies and less sensitive in higher frequencies.

Our texture descriptor is compact in representation regardless of inlage size and is shown to be effective in relevant image re- trieval. Furthermore, the intensity-, scale-, and rotation- invariant matching methods provide effective retrieval memcs for various applications.

OLtr proposed texture description method can be utilized to index and retrieve image and video. Some examples of applica- tions are fast video searcl-ling and video parsing. Another exanl- ple is contents-based image retrieval of aerial photos, fabric im- ages, and electronic photo albums. The texhue descriptor is a very effective way to describe object segmentation and image and video contents.

REFERENCES

[I] ISOIIEC JTC I SC29 WG I 1 (MPEG), MPEG-7 fissllul pr~rt of' fiperimmtation Model Version 4.0, m3068, Maui, Decenlber 1999.

[2] ISOEC JTCl SC29 WGll (MF'EG), MPEG-7 Phial part of'

XM and WD, N3335, Noordwijkerhout, May 2000. [3] 0. D. Faugeras and W. K. Partt, "Decorrelation methods of texture

feature extraction," IEEE Trans. Pattern Anal. and Machine bl- fell., Vo1.2, .fdY 1980, pp. 323-332.

[4] 1. Fog1 and D. Sagi, "Gabor filters as texture discriminator," Bio- logii~.(~I Cvhernetics, Vo1.6 1, 1989, pp. 103- 1 13.

[5] B. S. Manjunath and W. Y. Ma, 'Texture Features for Browsing and Retrieval of Image Data," IEEE Pansactions on Pattern Ana!r?~k rrnd Muchine Intelligence, Vol. 18, No. 8, August 1996.

[6] Y. S. Kim, Y. S. Kim, W. Y. Kin1 and M. J. Kin\ "Development of Content-Based Trademark Retieval System on the World Wide Web", ETRI Jo~~rnal, Vol. 21, No. I, March 1999, pp. 4@54.

[7] R. Chellappa, “Two-dimensional discrete Gaussian Markov ~m- doln field ~nodels for unage processing," Prrrtetn Recognifioti, Vol. 2, 1985, pp.79-112.

[8] P. Wu, W. Y. Ma, B. S. Manjimath, H. D. Shin and Y. L. Choi, "A I

texture descriptor for MF'EG-7," ISOnEC JTCl SC29 FVGII 1 I

50 Yong Man Ro et a/. ETRI Journal, Volume 23, Number 2 , June 2001

Page 11: MPEG-7 Homogeneous Texture Descriptor - Vision … MPEG-7 homogeneous texture descriptor that we describe in this paper had been severely tested and conipared with other proposed texture

(MPEG). P77. Lancaster. 1999. [9] Y. M. Ro. "Matcliing Putwit : Co~itents feah~ting for hiiage Index-

ing," Pruceedir?g~ nfSPIE. \bl. 3527. 1998. pp. 89- 100. [lo] J. R. Ohm and E Bu~i.ja~nin. "Descr-iptor for tcxhlre ill wavclet

domain," ISO/IECJTCI SC29 WGII (MPEG). P566. Lancastcl: 1999.

[I 11 A. Saadane. H. Scnane and D. Barba. "On the Dcsiyi of Psy- cliovisual Quantizels for a Visual Subband l~nagc Coding." SmE, Vol. 2308. 1994. pp. 1446.

[I21 A. Saadane. F1. Saiane and D. Rarba, "An Entirely Psychovisual based Subband Image Coding Sclie~nc." PIE. Vol. 250 1 . 1995. pp. 1 702.

[I31 J. G. Daugman. "Higli Confidcncc Visual Rccogriition of Pasons by a Tcst of Statistical Tndcpcndcnce." IEEE fi.cr11.s. P C I I ~ C I ~ ~ Arln(1:~i.v m d 114nchi11c h~tcllige~lc~. Vol. 15, No. l I . Nowrc~iibcr. 1993.1711 1148-1161.

[I41 C. .I. La~nhrcclit. "A Working Spatio - Tcmporal Modcl of Human Vision Systcni for Imagc Restomtion and Quality Asscssmcnt Applications." IEEE Irl/cr-rlntiorlnl Conl?rc.r7tu~ o r 1 .4SLYP. New York, NY, USA. Vol. 4. 1996. pp. 229 1-4.

[I 51 Y. M. Ro, S.Y. 101n. K.\V Yoo. M. Kim arid J. Kim. "Tcxtu~r dc- scriptor using atoms for matching pursuit." /SO,TEC.TTCI SC2Y II'GII (IIdPEG). P6 12, Lancaster 1999.

[I 61 Y. M. Ro. K.W. Yoo. M. Kiln and J. Kini. "Tcxh~rc Description using Radon hansfonii," I;YO/IEC ./TCI SC29 M'GII (,k/lPEG). m4703. Vancouvc~; 1 999.

[I 71 Y. M. Ro. K.W. Yoo. M. Kim and J. Kim, "Texh~rc dcscriptio~i 11s- ing Radon ~ansform and expailncntal results on CT-5 corc cx- peri~ncnt using atoms for ~natcliing pu~suit." I<YO/IEC./TCI SC2Y Ii'GII (AdPEG). 1115 152. Mclbo~umc. 1999.

[ I 81 Y. M. Ro. K.W. Yoo, M. Kim. J. Kiln. B. S. Ma~ijunath. D. G. Sim. H. K. Kiln and J. R. Ohm. "An unified tcxturc dcsc~iptor," ISO//EC./TCI SC29 WGII CZdPEG). 1n5490. Maui, 1999.

[I 91 1. G. Daugiian. "Complctc discrctc 2-D Gabor transforms by neu- 1x1 nctworks for image analysis and comprcssion." IEEE Ti.trrls. .4SSP, Vol.36. .luly 1988. pp. 1 160-1 179.

[20] T. Cliang and C. S. KLIO, "Tcxh~re analysis and classification with Tree-Sh-uch~rcd Wavelet T~ansfo~ius." IEEE Pnrls. Irim!,.e Prnc<- e.~,~ir;q. Vo1.2. No.4, Oct. 1996.pp. 429-441.

-- q Yong Man Ro rcccivcd tlie B.S. fin~n Yonsei University. Seoul, Korca in 1981 and the M.S. and P1i.D. depecs h n i tlic Korca Advanced

I 6 - - I - Institute in Science and Technology (KAIST). ,!,! .. .. - / in 1987 and 1992. rcspectivcly. In 1987, lic wvas

A a staff associate at Colu~nbia Univc~sity. and fin~n 1992 to 1995. lic was a visitinp rcscarclicr

A12 in Unive~sity of California at I~vinc and KAIST. 111 1996. lic was a trsearcli fcllow at depa~tnicnt

of clcchical cnginccring and computcr scicnccs in University of Cali- fornia at Berkeley. In 1997. Iiejoincd Info~iiiation and Co~nmunication Univcrsity. Korca wlicrc lic is cumc~itly associate pnfcssor and director of Image Video Systcni Lab. I-lis rcscar-ch interests includc im- agc/vidco proccssitig. MPEG-7. feature rccogiition. iniagclvideo in- dexing. and spectral analysis of illlap sigial. He ~rccivcd the Young Invcsti_cator Finalist Award in ISMRM in 1992. He is a scnior nie~nher

of IEEE and member of SPTE and ISMRM.

Munchurl Kim lias I-eceived tlie R.E. dcgrc in electronics fiam Kyungook National IJniwr- sity. Korca in 1989. and M.E. and P1i.D. d c y ~ e s in elcctr-ical arid computcr engineering h ~ n Uniwnsity of Flo~ida Gaincs\ille. USA. in 1992 and 1996. respectively. After his graduation. lic joined Elcchnnics and Tclecomniunications Re-

.+ -4- search Institute (ETRI) where lic had wvorkcd ill

tlic MPEG-4 standardization ~rlatcd research arcas. Sincc 1998. lie lias been inw~olw,cd in MPEG-7 stmidardi7ation wvorks. In tlic cou~sc of MPEG standardization. lic lias hen conhibut- ing niorc than 30 l7lnposals in tlic areas of automatic/scmi-a~~to~iiatic scgiicntation of movin~ ol?iccts. MPEG-7 visual descriptors and Mul- timcdia Description Sclicmcs. and sc~vcd as tlie teal11 leader on cwralua- tion of Vidco Description Sclic~ne proposals in MPEG-7 in Lancaster- U.K.. 1999. In 2001, Iic joined. as assistant professor in school of aigi- ncc~ing. tlic Infomiation and Communications Unive~sity (ICU) in Tacjon, Korea. His rcscarcli areas of intel-est iricludc multimedia coni- putiny. communications and broadcasting. and ~nulti~ncdia interactive sc~~.iccs.

Ha K p n g Kang reccivcd tlie R.S. degree in elcctranic engineering finm Korea Uniw~ersity, Ko~ra. in 1998 and M.S. degree in image Ixnc- essing frorii Infoniiation and Communicatio~i

-. . IJnivc~sity. Korea. in 2000. Since 2000, He has :I

us hecn a P1i.D. candidate in the salnc universie. I . Ifis rcscalrli intcrcsts include contents-based

nir~lti~ncdia infomiation lrhieval. \vatcniiarlting and imagcl vidco processing.

dinlvoong Kim ~rccivcd tlie R.S. and tlie M.S. dcg~res tinm Seoul National U~iivcrsity. Seoul. Korea. in 198 1 and 1983. ~rspcctively, and tlie P1i.D. degree in the delm~tmcnt of elcchical cu-

-.a

I gincc~ing from Texas A&M U~i~vc~sity. Unitcd Statcs in 1993. Since 1983. lie lias been a IT- sca~rli staff in Elcchonics and Tclecoln~ntuiica- tions Research Instih~tc(ETRI). Korea. Hc is c~~~lncntlv a director in tlie broadcast ~ncdia tcch-

nolog dcpa~-hiic~it. He lias bccn cngagcd in tlic devcloplnc~it of TDX digital switching systc~n. MPEG-2 vidco cncoda: ITDTV cncoder sys- tcni. and MPEG-7 tcchnolo_sy. His resea~rh intel-ests includc digital signal p~occssing in tlic ficld of vidco comniunications. ~nultiriicdia systems. and intc~activc broadcast systems.

ETRI Journal, Volume 23, Number 2, June 2001 Yong Man Ro et a/. 51