* Front page |
Mustru: EvaluationTRECMustru Version 0.1 was tested on a TREC-8 Question & Answer dataset published in 1999. It consists of about 524K articles from various sources including the Financial Times, LA Times, and FBIS. The article text was segmented into passages with the maximum and minimum sizes of passages limited to 250 and 50 bytes respectively. 3.5 M passages were created with an average of 6.7 passages per article. The size of the text content alone excluding all tags was about 1.5 Gbytes. A small set of development questions were provided and Q&A systems were tested on a set of 198 questions. An answer was judged correct if it matched a regular expression generated for the particular question AND if the answer sentence originated in a document considered relevant for the question. A correct answer was awarded points based on the position in the hit list returned by the search engine. Points were scored only for the top ranking answer from the hit list.
Results
The final precision count (aka Mean Reciprocal Rank) for Mustru was 0.58 (144.81 / 198). A question was converted to a search engine query with five components. Each of the five components added to the overall precision of the answer with different contributions.
Not suprisingly, bigrams appear to be the largest contributor to the overall precision followed by unigrams. When used in queries, entities (hypernyms) and transformations appear to provide a marginal improvement in precision. Version 0.2In version 0.2, the entity extractor was replaced by a simple table lookup to speed up indexing and reduce memory requirements. Instead of indexing sentences and documents twice as in Version 0.1, a document is indexed just once. The most likely document to answer a question is first retrieved followed by a search for the top two passages that may answer the question. In version 0.1, a search query was generated for the best passage and the associated document was not fetched. The results in version 0.2 have lower precision , but are reasonable. As before the top 5 hits were used to judge if the search engine found an answer. Even though the document that contains the answer was fetched, passage retrieval extracted the sentence containing the answer in 122 out of 151 (80%) questions.
Copyright © 2007 Mustru Search Services. All rights reserved. |