Evaluating High Accuracy Retrieval Techniques

27
1 Evaluating High Accuracy Retrieval Techniques Chirag Shah ,W. Bruce Croft Center for Intelligent Information Retrie val Department of Computer Science University of Massachusetts Presented By Yi-Ting Chen

description

Evaluating High Accuracy Retrieval Techniques. Chirag Shah ,W. Bruce Croft Center for Intelligent Information Retrieval Department of Computer Science University of Massachusetts Presented By Yi-Ting Chen. Outline. Introduction Analysis of the problem The approaches for modifying queries - PowerPoint PPT Presentation

Transcript of Evaluating High Accuracy Retrieval Techniques

Page 1: Evaluating High Accuracy Retrieval Techniques

1

Evaluating High Accuracy Retrieval Techniques

Chirag Shah ,W. Bruce CroftCenter for Intelligent Information Retrieval

Department of Computer ScienceUniversity of MassachusettsPresented By Yi-Ting Chen

Page 2: Evaluating High Accuracy Retrieval Techniques

2

Outline

Introduction Analysis of the problem The approaches for modifying queries

Method1: giving more weight to the headwords. Method2: Using clarity scores as weights. Method3: Using clarity scores to find terms to exp

and with WordNet. Experiments and analysis Conclusion and future work

Page 3: Evaluating High Accuracy Retrieval Techniques

3

Introduction

Ad-hoc retrieval and question answering: two major research streams in the present IR.

Different goals- Ad-hoc retrieval is mainly about retrieving a set of documents with good precision and recall, whereas QA focuses on getting one correct answer or a small set of answers with high accuracy.

In this paper, tried to link these two achieving high accuracy with respect to the most relevant results.

Page 4: Evaluating High Accuracy Retrieval Techniques

4

Introduction

HARD: TREC introduced a new track called High Accuracy Retrieval from Documents in 2003.

This paper task also deals with the problem of getting high accuracy in retrieval, but with contrast to HARD, do not make use of any additional information.

Studying how QA techniques can help in getting high accuracy for ad-hoc retrieval and propose a different measure for evaluation (MRR) instead of recall and precision.

Page 5: Evaluating High Accuracy Retrieval Techniques

5

Analysis of the problem

To facilitate the evaluation of a system with high results in the rank list, QA system use mean reciprocal rank (MRR) as the measure.

MRR id defined as the inverse of the rank of the retrieved result.

Investigates the problem of achieving high accuracy with this perspective and analyzes the reasons behind low accuracy.

Page 6: Evaluating High Accuracy Retrieval Techniques

6

Analysis of the problem

Baseline TREC’s Tipster vol. I and II as datasets and topics 51-200

(total 150) as queries (both title and description). Used the language modeling framework for retrieval.

Used the Lemur toolkit for implementing our retrieval system.

The results of our baseline runs along with those of various QA systems of TREC-8 [26], TREC-9 [27], TREC-10 [28], and TREC-11 [29] are given in table 1.

Page 7: Evaluating High Accuracy Retrieval Techniques

7

Page 8: Evaluating High Accuracy Retrieval Techniques

8

Analysis of the problem

We can see that our baselines have a correct document in rank 1 about 40%.

Look at the distributions of queries with respect to the rank.

The average MRR could be increased by moving up relevant documents form lower ranks in the case of poorly performing queries, and by moving some of the big group of relevant document at rank 2 to rank 1.

Page 9: Evaluating High Accuracy Retrieval Techniques

9

Page 10: Evaluating High Accuracy Retrieval Techniques

10

Page 11: Evaluating High Accuracy Retrieval Techniques

11

Analysis of the problem

Analyzing some bad queries: We could improve the MRR value in all the cases

by improving the query. There were three problems that we could identify:

ambiguous words in the query. mixtures of words of different importance in the query. Query-document information mis-match.

Not all the words are equally important in a query. Not all kinds of expansion can help.

Page 12: Evaluating High Accuracy Retrieval Techniques

12

Page 13: Evaluating High Accuracy Retrieval Techniques

13

The approaches for modifying queries Method1: giving more weight to the headwords

Find the headword in a given query and giving it more weight than other words of the query. Parse the query for part-of speech (POS) tagging. Find the first noun phrase using POS information. Consider the last noun of this noun phrase as the h

eadword. Give this headword more weight.

Page 14: Evaluating High Accuracy Retrieval Techniques

14

The approaches for modifying queries Method2: Using clarity scores as weights

The reason for poor retrieval is often the use of ambiguous words in the query.

To implement this idea, He used Cornen-Townsend etal.’s technique of finding query clarity scores.

Computing the relative entropy between a query language model and the corresponding collection language model.

Page 15: Evaluating High Accuracy Retrieval Techniques

15

The approaches for modifying queries Method2: Using clarity scores as weights

Find query clarity scores based on the technique. Construct weighted queries with clarity score of each word

as its weight as we want to give more weight to words that have high clarity scores.

Method3: Using clarity scores to find terms to expand with WordNet. Query-dataset mismatch is another factor that affects the a

ccuracy of the retrieval. It is not useful to expand every word of the given query.

Page 16: Evaluating High Accuracy Retrieval Techniques

16

The approaches for modifying queries Method3: Using clarity scores to find terms to

expand with WordNet. Find query clarity scores. Divide all the terms into the following tree categori

es: Terms with high clarity scores should not be touched. Terms with very low clarity scores are likely to be very a

mbiguous. These words are ignored. Expand the terms whose scores are between the two lim

its of clarity scores using wordNet synonyms.

Page 17: Evaluating High Accuracy Retrieval Techniques

17

Experiments and analysis

More than 700,000 documents from Tipster collection and taking more than 2GB of disk space.

Extracted from topics 51-200 making total 150 queries.

The experiments were conducted on both title queries as well as description queries.

Page 18: Evaluating High Accuracy Retrieval Techniques

18

Experiments and analysis

Title queries:

Page 19: Evaluating High Accuracy Retrieval Techniques

19

Experiments and analysis

Page 20: Evaluating High Accuracy Retrieval Techniques

20

Experiments and analysis

Page 21: Evaluating High Accuracy Retrieval Techniques

21

Experiments and analysis

Page 22: Evaluating High Accuracy Retrieval Techniques

22

Experiments and analysis

Description query

Page 23: Evaluating High Accuracy Retrieval Techniques

23

Experiments and analysis

Page 24: Evaluating High Accuracy Retrieval Techniques

24

Experiments and analysis

Page 25: Evaluating High Accuracy Retrieval Techniques

25

Experiments and analysis

Page 26: Evaluating High Accuracy Retrieval Techniques

26

Experiments and analysis

We can see in the runs for title queries that we pushed more queries to rank1, we also got more queries at ranks higher than 100. This means that while trying to improve the queries, we also hurt some queries.

Runs for description queries be significantly better.

Page 27: Evaluating High Accuracy Retrieval Techniques

27

Conclusion and future work

Our focus in the presented work was to improve the MRR of the first relevant document, the proposed techniques also helped in improving overall precision in many case.

As one of our next steps in this research, we carried out experiments with relevance models.

We noticed that while bringing some queries up in the rank list, the model also drove some other down in the list.