Resource Selection & Integration of Data on Deep Web search engines cannot recover content in the...

6
IJSRD - International Journal for Scientific Research & Development| Vol. 3, Issue 03, 2015 | ISSN (online): 2321-0613 All rights reserved by www.ijsrd.com 1485 Resource Selection & Integration of Data on Deep Web Sumit Patel 1 Vinod Azad 2 1,2 Department of Computer Science & Engineering 1,2 Truba College of Engineering and Technology AbstractMost web structures are huge, intricate and users often miss the purpose of their inquest, or get uncertain results when they try to navigate through them. Internet is enormous compilation of multivariate data. Several problems prevent effective and efficient knowledge discovery for required better knowledge management techniques it is important to retrieve accurate and complete data. The hidden web, also known as the invisible web or deep web, has given rise to a novel issue of web mining research. a huge amount documents in the hidden web, as well as pages hidden behind search forms, specialized databases, and dynamically generated web pages, are not accessible by universal web mining application. in this research we proposed a approach is designed that has a robust ability to access these hidden web techniques for better invisible web resources selection and integration system. In this research we using SC technique for invisible web resources selection and integration and its construction for real-world domains based on database schemas clustering, web searching interfaces and improve traditional methods for information retrieve. Applications of our proposed system include invisible web query interface mapping and intelligent user query intension recognition based on our domain knowledge-base. Key words: Invisible Web integration, deep web, web resource, Schema matching, knowledge base. I. INTRODUCTION Now, the internet has emerged a growing number of online databases called web database. According to statistics, the number of web databases is more than 500 million, on this basis constitutes a deep web. The invisible web the majority of the information is stored in the retrieval databases and the greatest division of them is structured data stored in the backend databases, such as mysql, db2, access, oracle, sql server and so on. Traditional search engines create their index by spidering or crawling surface web pages. to be exposed, the page ought to be static and linked to other pages. Traditional search engines cannot recover content in the invisible web those pages do not live until they are created dynamically as the result of a precise search. Because traditional search engine crawlers cannot probe beneath the surface, to make easy the users to retrieve the invisible web databases, an amount of invisible web sites have done a lot of subsidiary work, such as classifying the invisible web databases manually by constructing a summary database, and providing users with a unified query interface, the user can retrieve the information by comparing the query results of similar topic to conclude which one can answer their needs better. However, bodily classification efficiency is extremely low and cannot convene user’s information needs. in this paper, an automatic classification approach of invisible web sources based on schema matching data analysis techniques technique according to query interface characteristics is presented. Domain specific search sources focus on documents in confined domains such as documents concerning an association or in a exact subject area. Most of the domain specific search sources consist of organizations, libraries, businesses, universities and government agencies. In our daily life we are provided with several kinds of database directories to store critical records. Likewise to position an exacting site in the ocean of internet there have been efforts to systematize static web content in the form of web directories i.e. bing. The procedure adopted is both manual and automatic. Likewise to organize myriad invisible web databases, we need a impressive database to store information about all the online invisible web databases. a few aspects which make the task of automatic organization of invisible web sources indispensable are: the understanding of the semantic web can be made possible. This paper surveys the automatic invisible web organization techniques proposed and we discuss the section II related works section IIISC approach, and lst section IV conclusion so far in the light of their effectiveness and coverage for web intelligent integration and information retrieval. The key characteristics significant to both exploring and integrating invisible web sources are thoroughly discussed. II. RELATED WORK In the history few years, in order to extract data records from web pages, many semi-automatic and Automatic approaches to integration data records have been reported in the literature, e.g., [2] [3] [4]. These existing works adopts many techniques to solve the problem of web integration, including data source selection and schema-matching. We briefly discuss some of the works below. W. su in [11] proposed a hierarchical classification method that classifies structured deep web data sources into a predefined topic hierarchy automatically using a combination of machine learning and query probing techniques. Not much effective in dealing with structured data sources. For convergence in results of deep web classification domains. H.le et al. [12] investigates the problem of identifying suitable feature set among all the features extracted from the deep web search interfaces. Such features remove divergence of domain in the retrieved results. d.w.embley, et al in [5] describes a heuristic approach to discover record boundaries in web documents. in thisapproach, it captures the structure of a document as a tree of nested html tags, locates the sub tree containing the records of interest, identifies candidate separator tags within the sub tree using five independent heuristics and selects a consensus separator tag based on a combined heuristic. k. lerman, et al in [4] describes a technique for extracting data from lists and tables and grouping it by rows and columns. the authors developed a suite of unsupervised

Transcript of Resource Selection & Integration of Data on Deep Web search engines cannot recover content in the...

IJSRD - International Journal for Scientific Research & Development| Vol. 3, Issue 03, 2015 | ISSN (online): 2321-0613

All rights reserved by www.ijsrd.com 1485

Resource Selection & Integration of Data on Deep Web

Sumit Patel1 Vinod Azad

2

1,2Department of Computer Science & Engineering

1,2Truba College of Engineering and Technology

Abstract— Most web structures are huge, intricate and users

often miss the purpose of their inquest, or get uncertain

results when they try to navigate through them. Internet is

enormous compilation of multivariate data. Several

problems prevent effective and efficient knowledge

discovery for required better knowledge management

techniques it is important to retrieve accurate and complete

data. The hidden web, also known as the invisible web or

deep web, has given rise to a novel issue of web mining

research. a huge amount documents in the hidden web, as

well as pages hidden behind search forms, specialized

databases, and dynamically generated web pages, are not

accessible by universal web mining application. in this

research we proposed a approach is designed that has a

robust ability to access these hidden web techniques for

better invisible web resources selection and integration

system. In this research we using SC technique for invisible

web resources selection and integration and its construction

for real-world domains based on database schemas

clustering, web searching interfaces and improve traditional

methods for information retrieve. Applications of our

proposed system include invisible web query interface

mapping and intelligent user query intension recognition

based on our domain knowledge-base.

Key words: Invisible Web integration, deep web, web

resource, Schema matching, knowledge base.

I. INTRODUCTION

Now, the internet has emerged a growing number of online

databases called web database. According to statistics, the

number of web databases is more than 500 million, on this

basis constitutes a deep web. The invisible web the majority

of the information is stored in the retrieval databases and the

greatest division of them is structured data stored in the

backend databases, such as mysql, db2, access, oracle, sql

server and so on. Traditional search engines create their

index by spidering or crawling surface web pages. to be

exposed, the page ought to be static and linked to other

pages. Traditional search engines cannot recover content in

the invisible web those pages do not live until they are

created dynamically as the result of a precise search.

Because

traditional search engine crawlers cannot probe

beneath the surface, to make easy the users to retrieve

the invisible web databases, an amount of invisible web sites

have done a lot of subsidiary work, such as

classifying the invisible web databases manually by

constructing a summary database, and providing users with

a unified query interface, the user can retrieve the

information by comparing the query results of similar topic

to conclude which one can answer their needs better.

However, bodily classification efficiency is extremely low

and cannot convene user’s information needs. in this paper,

an automatic classification approach of invisible web

sources based on schema matching data analysis techniques

technique according to query interface characteristics is

presented. Domain specific search sources focus on

documents in confined domains such as documents

concerning an association or in a exact subject area. Most of

the domain specific search sources consist of organizations,

libraries, businesses, universities and government agencies.

In our daily life we are provided with several kinds of

database directories to store critical records. Likewise to

position an exacting site in the ocean of internet there have

been efforts to systematize static web content in the form of

web directories i.e. bing. The procedure adopted is both

manual and automatic. Likewise to organize myriad

invisible web databases, we need a impressive database to

store information about all the online invisible web

databases. a few aspects which make the task of automatic

organization of invisible web sources indispensable are: the

understanding of the semantic web can be made possible.

This paper surveys the automatic invisible web

organization techniques proposed and we discuss the section

II related works section IIISC approach, and lst section IV

conclusion so far in the light of their effectiveness and

coverage for web intelligent integration and information

retrieval. The key characteristics significant to both

exploring and integrating invisible web sources are

thoroughly discussed.

II. RELATED WORK

In the history few years, in order to extract data records

from web pages, many semi-automatic and Automatic

approaches to integration data records have been reported in

the literature, e.g., [2] [3] [4]. These existing works adopts

many techniques to solve the problem of web integration,

including data source selection and schema-matching. We

briefly discuss some of the works below.

W. su in [11] proposed a hierarchical classification

method that classifies structured deep web data sources into

a predefined topic hierarchy automatically using a

combination of machine learning and query probing

techniques. Not much effective in dealing with structured

data sources. For convergence in results of deep web

classification domains.

H.le et al. [12] investigates the problem of

identifying suitable feature set among all the features

extracted from the deep web search interfaces. Such features

remove divergence of domain in the retrieved results.

d.w.embley, et al in [5] describes a heuristic

approach to discover record boundaries in web documents.

in thisapproach, it captures the structure of a document as a

tree of nested html tags, locates the sub tree containing the

records of interest, identifies candidate separator tags within

the sub tree using five independent heuristics and selects a

consensus separator tag based on a combined heuristic.

k. lerman, et al in [4] describes a technique for

extracting data from lists and tables and grouping it by rows

and columns. the authors developed a suite of unsupervised

Resource Selection & Integration of Data on Deep Web

(IJSRD/Vol. 3/Issue 03/2015/365)

All rights reserved by www.ijsrd.com 1486

learning algorithms that induce the structure of lists by

exploiting the regularities both in the format of the pages

and the data contained.

bingliu, et al in [1] proposes an algorithm called

mdr to mine data records in a web page automatically. the

algorithm works in three steps, e.g. building the html tag

tree, mining data region, identifying data records.

shirenyingwang, et al in [13]importing focused crawling

technology makes the identification of deep web query

interface locate in a specific domain and capture relative

pages to a given topic instead of pursuing high overlay

ratios. This method has dramatically reduced the quantity of

pages for the crawler to identify deep web query interfaces.

Jia-lingkoh, et al in [14] theyhave beendiscuss about a

multi-level hierarchical index structure to group similar tag

sets. Not only the algorithms of similarity searches of tag

sets, but also the algorithms of deletion and updating of tag

sets by using the constructed index structure are

provided.for the purpose of this study, following features are

considered essential to the invisible web intelligent

integration and information retrieval. a huge amount

documents in the hidden Web, as well as pages hidden

behind search forms, specialized databases, A lot of work

has been carried out in these fields by researchers. This part

of study enlightens briefly on some of work done by those

researchers., [16] introduces the semantic Deep Web, and

some high complexity of schema extraction, which do not

achieve the practical standards, so these methods are not

adapt to extract schema of query interface automatically.

Therefore, that how to extract the meaningful information

from query interfaces and merge them into attributes plays

an important role for the interface integration in deep web.

III. OUR APPROACH AND TECHNIQUES

Our Approach is based on four following points:-

A. Effectiveness Improvement Model:

The Effectiveness Improvement is based on

estimating the Effectiveness of the web database bringing to

a given status of invisible web integration system by

integrating it. In this section, we describe how the

Effectiveness of web database is estimated.

Suppose we are given an integration system I and a

set of candidate web databases K= {k1, k2... kn} Everything

is a double-edged sword, given a candidate web database ki,

if the system integrates ki, the

Integration system I would be affected by the positive and

negative utility of ki. In this research, the positive and

negative utility of kibringing to I by integrating kiare

respectively denoted by

and hence, the utility of kibringing to I by integrating kican

be expressed as the following difference:

( )

* +

In next two subsection, we show how we measure and

respectively.

1) Positive Utility

In this research, can be expressed by the volume ofnew

data that add to the integration system by integrating ki,

denoted by and the importance of new data,denoted

by. In this research, the importance of newdata is

expressed by correlation of the degree of thesenew data with

greater importance query. So the morevolume of new data

and they are involved in morequeries with greater

importance, kibring more positiveutility to a given status of

integration system I.

Thus, is defined as:

* +

Estimating In this research

can be defined as follows

Definition 1 ( ): Given a candidate invisible web

database kiand the status of integration system I, the is

expressed by the volume of new data that add to the

integration system by integrating i s. Simply speaking, is

the number of data that contained in ki, but not in I. The

can be expressed by the following equation.

| | | |

Where| | is the volume of data of unions of web databases

in I , not counting duplicate data in I .

| |is the volume of data after I integrate ki

Broadly speaking, can be measured by analyzing allthe

data in Iand ki. The analysis of all data makes asolution that

requires fetching all the data from webdatabases

prohibitively expensive. Hence, in nextsubsection we show

how we approximate . Ourexperimental evaluation shows

that despite our

approximations, our approach is to select and

integrate web databases effectively. As discussed above, we

cannot possibly analyze all the data in I and ki. So we

estimate approximate by analyzing partial data which are

obtained by sampling small amount of data from I and ki

randomly with query-based sampling.

B. Queries and Workloads:

Queries are the primary mechanism for retrieving

information from web database. Given a query q , when

querying web database ki, We denote the result set of q over

kiby q(ki ) . In this research, a query workload Q is a set of

random queries: Q= {q1, q2..., qn} as the result set is

retrieved by random queries, query-based results indicate

the objective content of the web database.

To estimate approximate , we analyze the result

set of the query workload Q over i s and D representing all

data in ki and I. In what follows, we show how to estimate

approximate . The approximate

can be expressed by

the following equation.

| ( ) ( )| | ( )|

| ( )| ( )

Where size (ki) s is the amount of data in

ki,| ( )|and | ( )| is separately the size of the result set of

the query workload Q over kiand I .In this research , Q(ki )

s is defined as the union of the result set for the queries in

the workload Q on ki.

( ) ⋃( ( ))

| |

With Q (ki) similar, Q (I) is the union of the result set for the

queries in the workload Q on the integration

Resource Selection & Integration of Data on Deep Web

(IJSRD/Vol. 3/Issue 03/2015/365)

All rights reserved by www.ijsrd.com 1487

System I. Different from query on single web database,

when querying the integration system I, the query processor

utilizes all the integrated web databases. Merging result

from all the integrated web databases into result set,

eliminating all duplication of data at the same time, we

denote result set by Q (I). The high cost work to obtain Q

(I), in the next section; we will introduce an efficiency

approach to obtain Q (I).

C. Centralized sample database and Duplicate detection:

Web databases, as we know, are heterogeneous in

the web. In this research, in order to obtain Q(I)and | ( ) ( )| ,we build a centralized sample database with

consolidated single mediated schema that is set by the

domain expert. We mapped the result set for the queries in

the workload Q on each ki in S to centralized sample

database. Q(I)and | ( ) ( )|can easily be obtained,

and duplication of data can also be detected by using a

probabilistic approach [16] in centralized sample database.

A probabilistic approach is proposed for solving the

duplicate detection problem in [16], it can be used to match

records with multiple fields in the database.

1) Detecting Duplicate Records:

In the previous section, we described methods that can be

used to match individual fields of a record. In most real-life

situations, however, the records consist of multiple fields,

making the duplicate detection problem much more

complicated. In this section, we review methods that are

used for matching records with multiple fields. The

presented methods can be broadly divided into two

categories:

Approaches that rely on training data to "learn"

how to match the records. This category includes

(some) probabilistic approaches

Approaches that rely on domain knowledge or on

generic distance metrics to match records. This

category includes approaches that use declarative

languages for matching and approaches that devise

distance metrics appropriate for the duplicate

detection task.

2) Notation:

We use and to denote the tables that we want to

match, and we assume, without loss of generality, that

and have comparable fields. In the duplicate

detection problem, each tuple pair

is assigned to one of the two

classes and . The class contains the record

pairs that represent the same entity (" match") and the

class contains the record pairs that represent two

different entities (" nonmatch").

We represent each tuple pair as a random

vector with components

that correspond to the comparable fields of

and . Each shows the level of agreement of

the field for the records and . Many

approaches use binary values for the s and

set if field agrees and let if

field disagrees.

3) Probabilistic Matching Model:

Newcombeet were the first to recognize duplicate detection

as a Bayesian inference problem. Then, Fellegi and Sunter

formalized the intuition of Newcombe et al. and introduced

the notation that we use, which is also commonly used in

duplicate detection literature. The comparison vector is

the input to a decision rule that assigns to or

to . The main assumption is that is a random

vector whose density function is different for each of the

two classes. Then, if the density function for each class is

known, the duplicate detection problem becomes a Bayesian

inference problem. In the following sections, we will discuss

various techniques that have been developed for addressing

this (general) decision problem.

The Bayes Decision Rule for Minimum

Error Let be a comparison vector, randomly drawn from

the comparison space that corresponds to the record

pair . The goal is to determine

whether or . A

decision rule, based simply on probabilities, can be written

as follows:

This decision rule indicates that, if the probability of the

match class , given the comparison vector , is

larger than the probability of the non match class ,

then is classified to , and vice versa. By using the

Bayes theorem, the previous decision rule may be expressed

as:

Theration

is called the likelihood ratio. The ratio denotes the

threshold value of the likelihood ratio for the decision. We

Resource Selection & Integration of Data on Deep Web

(IJSRD/Vol. 3/Issue 03/2015/365)

All rights reserved by www.ijsrd.com 1488

refer to the decision rule in (3) as the Bayes

test for minimum error. It can be easily shown [60] that the

Bayes test results in the smallest probability of error and it

is, in that respect, an optimal classifier. Of course, this holds

only when the distributions of , and

the priors and are known; this,

unfortunately, is very rarely the case.One common

approach, usually called Naive Bayes, to computing the

distributions of and is to make a

conditional independence assumption and postulate that the

probabilities and are

independent if . (Similarly, for

and .) In that case, we have

The values of and can be

computed using a training set of pre labeled record pairs.

However, the probabilistic model can also be used without

using training data. Jaro [61] used a binary model for the

values of (i.e., if the field "matches" ,

else ) and suggested using an expectation

maximization (EM) algorithm [62] to compute the

probabilities . The

probabilities can be estimated by taking

random pairs of records (which are with high probability

in ).

When the conditional independence is not a reasonable

assumption, then Winkler [63] suggested using the general

expectation maximization algorithm to

estimate , . In [64], Winkler claims

that the general, unsupervised EM algorithm works well

under five conditions:

1. The data contain a relatively large percentage of matches

(more than 5 percent),

2. The matching pairs are "well-separated" from the other

classes,

3. The rate of typographical errors is low,

4. There are sufficiently many redundant identifiers to

overcome errors in other fields of the record, and

5. The estimates computed under the conditional

independence assumption result in good classification

performance.

Winkler [64] shows how to relax the assumptions

above (including the conditional independence assumption)

and still get good matching results. Winkler shows that a

semi-supervised model, which combines labeled and

unlabeled data (similar to Nigam et al. [65]), performs better

than purely unsupervised approaches. When no training data

is available, unsupervised EM works well, even when a

limited number of interactions is allowed between the

variables. Interestingly, the results under the independence

assumption are not considerably worse compared to the case

in which the EM model allows variable interactions.

Du Bois [66] pointed out the importance of the fact that,

many times, fields have missing (null) values and proposed

a different method to correct mismatches that occur due to

missing values. Du Bois suggested using a new comparison

vector with dimension instead of

the comparison vector such

that

Where

(6)Using this representation, mismatches that occur due to

missing data are typically discounted, resulting in improved

duplicate detection performance. Du Bois proposed using an

independence model to learn the distributions

of and by using a set of pre

labeled training record pairs.

4) Estimate size of database:

Based on equation 3, to compute approximate , we need

to be able tocompute the amount of data in web database ki.

The difficulty is computing the amount of data in web

database, because (1) many sources do not allow

unrestricted access to their data, and (2) even if the sources

did allow access to the data, the sheer amount of data at the

sources makes a solution that requires fetching all the data

from the sources prohibitively expensive [15]. Thus, we

need a way to estimate the amount of database in web

database with a few accessing the data. Ling et al [17]

propose based on the word frequency an approach to assess

the size of web database. In this research , we could use it to

assess the size of web database. For instance, for ki, size(ki)

refers to the size of web database ki.

2) Estimating In this subsection, we mainly focus on the

importance of new data that add to the integration system by

integrating a web database.

Definition 2 given a candidate invisible

webdatabase i s and the status of integration system I,

the is expressed by correlation of the degree of thesenew

data with greater importance query.So we generate a set of

queries with weight to estimate importance of new data. A

Resource Selection & Integration of Data on Deep Web

(IJSRD/Vol. 3/Issue 03/2015/365)

All rights reserved by www.ijsrd.com 1489

query workload QW is a set of pairs of the form {q, w},

where q is a query and w is a weight attributed to the query

denoting its relative importance. Typically, the weight

assigned to a query is proportional to its frequency in the

workload, but it can also be proportional to other measures

of importance, such as the monetary value associated with

answering it, or in relation to a particular set of queries for

which the system is to be optimized [18]. In this research,

we use a query generator to generate a set of queries. Each

generated query refers to a single term and is representative

of the set of queries that refer that term. For simplicity, the

generator only produces keyword queries. The generator

assigns a weight w to each query using a distribution to

represent the frequency of queries on this term. Since the

distribution of query-termfrequencies on web search engines

typically follows a long-tailed distribution [19], for w in our

experiments we use values selected from a Pareto

distribution [20].

In what follows, we show how to estimate

approximate . The approximate

is defined as the

weightedsum of the volume of new data for each the queries

in theworkload QW.

( )

∑ ( )

(| ( ) ( )| | ( )|

| ( )|)

a) . Negative Utility

There are networking and processing costs associated with

integrating a web database in the integration system. These

are the costs to retrieve data from the database while

executing queries, map this data to the global mediated

schema and so on. Those cost aspects are the negative utility

of web database and they may be just as important to users.

In this research, we mainly consider time-cost as negative

utility. Time-cost is expressed by response time that the time

starts from user sending a query to the web database or

integration system and ends at time they return the final

result set of this query. Response time contains time-cost to

retrieve data from the database while executing queries, map

this data to the global mediated schema, and resolve any

inconsistencies with data retrieved from all sources and so

on.

In what follows, we use a random query workload Q and a

query with weighted workload QW that are usedin above

subsection to estimate approximate . Theapproximate

can be expressed by the followingequation.

* + ( ) is the increased average response time of a random query q

in Q over after I integrating ki , ( ) is the

increased average response time of a query q in QW over

after D integrating ki

. ( ) Can be expressed by the following equation

∑ ( ( )

( ))| |

| |

∑ ( ( )

( )) | |

| |

∑ ( ( ))

| |

| |

∑ ( ( ))

| |

| |

D. Resource Selection & Integration using EIM:

In this section, we describe how to use the effectiveness

maximization model, which optimizes the resource selection

problems for invisible web data integration. The goal of the

resource selection algorithm is to build an integration

system contains m web databases (e.g., 20 databases) that

contain as high utility as possible, which can be formally

defined as an optimization problem:

Given a candidate source set: K= {k1, k2... kn} the

status of integration system I, find In order to actually

compute the utility of a web database as defined in Equation

1,2. we Standardize

,

and . One which have the range, 0–1.The

database selection decision is made based on

theapproximate utility of the web database.Our approach is

to select and integrate web databasesin an iterative manner,

where web databases are integrated incrementally. We select

a maximal utility webdatabase i s to integrate from S each

time. This approach takes advantage of the fact that some

web databases provide more utility to the status of

integration system than others: they are involved in more

queries with greater importance or are associated with more

data. Similarly, some data sources may never be of interest,

and therefore spending any effort on them is unnecessary.

The selection and integration algorithm using the utility

maximization model as follow:

Proposed algorithm

Algorithm to invisible web database selection and

integration for knowledge management

Integration Algorithm (I=φ;K: Set of candidates

Web databases; m is the maximum number of sources that

The user is willing to select (m ≤| K |)

Count=0;

While (Count≤ m) do

( ( ))

;//select a maximal

Utility of kiform K

I=integrate (I, k); //integrate (I, k) is integrate s

Resource Selection & Integration of Data on Deep Web

(IJSRD/Vol. 3/Issue 03/2015/365)

All rights reserved by www.ijsrd.com 1490

Into I, the status of integration system I is updated

K = K - k; //Set of candidates web databases K is

updated

Count++;

End while

Return I;

IV. CONCLUSION

In order to create knowledge for makingaccurate and

appropriate decisions we need to integrate data fromthese

heterogeneous deep web sources. In this research a

detailedsurvey of automatic deep web Integration techniques

ispresented, which is key to the realization of the data

integration from heterogeneous datasources. At web scale,it

is infeasible to cluster data sources into domains

manually.We deal with this problem and propose a schema

clusteringapproach that leverages techniques from document

clustering.We use a selection and Integration approach to

handle the uncertaintyin assigning schemas to domains,

which fits with previouswork on data integration with

uncertainty.

REFERENCES

[1] XiaoJun Cui,ZhongShengRen, HongYuXiao,LeXu,

Automatic Structured Web Databases Classification -

IEEE-2010 .

[2] Tantan Liu Fan Wang GaganAgrawal,Instance

Discovery and Schema Matching WithApplications to

Biological Deep Web Data Integration, International

Conference on Bioinformatics and Bioengineering-

IEEE-2010.

[3] BaohuaQiang, Chunming Wu, Long Zhang,Entities

Identification on the Deep Web Using Neural Network ,

International Conference on Cyber-Enabled Distributed

Computing and Knowledge Discovery-2010

[4] KishanDharavath, Sri KhetwatSaritha,Organizing

Extracted Data: Using Topic Maps,Eighth International

Conference on Information Technology: New

Generations-2011.

[5] Bao-huaQiang, Jian-qing Xi, Bao-huaQiang, Long

Zhang,An Effective Schema Extraction Algorithm on

the Deep Web-IEEE- 2008

[6] Bin He, Tao Tao, and Kevin Chen Chang, “Organizing

structured web sources by query schemas: a clustering

approach[R]”. Computer Science Department:

CIKM,2004.

[7] L. Barbosa and J. Freire.Searching for hidden-web

databases.In WebDB, 2005.

[8] B. Liu, R. Grossman, and Y. Zhai. "Mining Data

Records in Web Pages", SIGKDD, USA, August 2003,

pp. 601-606.

[9] B. He and K. Chang. Statistical schema matching across

Web query interfaces. In Proc. of SIGMOD, 2003.

[10] W. Wu, C. T. Yu, A. Doan, and W. Meng. “An

interactive clustering-based approach to integrating

source query interfaces on the Deep Web.” In

Proceedings of ACMSIGMOD International

Conference on Management ofData, pp.95-106, ACM

Press, Paris ,2004.

[11] W. Su, J. Wang, F. Lochovsky: Automatic Hierarchical

Classification of Structured Deep Web Databases.

WISE 2006, LNCS 4255, pp 210-221.

[12] HieuQuang Le, Stefan Conrad: Classifying Structured

Web Sources Using Support Vector Machine and

Aggressive Feature Selection. Lecture Notes in

Business Information Processing, 2010, Volume 45, IV,

270-282.

[13] Ying Wang, WanliZuo, Tao Peng, Fengling

He“Domain-Specific Deep Web Sources Discovery”

978-0-7695-3304-9 Fourth International Conference on

Natural Computation 2008.

[14] D’Souza, J. Zobel, and J. Thom. “Is CORI Effective for

Collection Selection an Exploration of parameters,

queries, and data.” In Proceedings of Australian

Document Computing Symposium, pp.41-46,

Melbourne,Australia ,2004.

[15] H. He, W. Meng, C. Yu, and Z. Wu. Wise-integrator: an

automatic integrator of web search interfaces for

ecommerce. In VLDB, 2003.

[16] W. Wu, C.T.Yu, A. Doan, and W.Meng. An interactive

clustering-based approach to integrating source query

interfaces on the deep web. In Sigmod, 2004.

[17] Jia-Ling Koh and NonhlanhlaShongwe.” A Multi-level

Hierarchical Index Structure for Supporting Efficient

Similarity Search on Tag Sets”978-1-4577-1938-7-

IEEE- 2011.

[18] He H, Meng WY, Lu YY, et al, “Towards Deeper

Understanding of the Search Interfaces of the Deep

Web,” WWW2007, 10 (2):133-155.

[19] Zhang Z., He B., Chang K.C., “Understanding Web

Query Interfaces: Best-effort Parsing with Hidden

Syntax,” In: Proceedings of the 23th ACM SIGMOD

International Conference on Management of Data,

2004, pp107-118.

[20] Yoo JUNG AN, JAMES GELLER, YI-TA WU, SOON

AE CHUN, “Semantic Deep Web: Automatic Attribute

Extraction from the Deep Web Data Sources,” ACM,

2007, pp1667-1672