|
Research
Abstracts - 2006
|
Open Domain Question Answering on Newspaper TextBoris Katz, Gregory Marton, Sue Felshin, Daniel Loreto & Federico MoraThe ProblemNIST's annual TREC Question Answering Track supports the community by providing shared evaluation on a number of important tasks. Factoid and list questions test systems on their ability to answer factual questions exactly, based on a body of English newspaper text. The definition and relationship tasks evaluate systems on their ability to locate the most important facts about a subject, or about the complex relationships between two subjects. MotivationInformation systems aimed at humans often return documents, paragraphs, snippets of documents, or sentences, and the user must find the actual answer in the text returned. While the surrounding context is often useful to humans, knowing the exact answer can help automatic systems answer more complex questions. For example, if a system can find that "the fifth largest country in Africa" is "Chad", then it becomes possible to answer followup questions, "How many people live there?", and more complex questions, "How many people live in Africa's fifth largest country?". Previous WorkSince 1992, TREC evaluations [7] have been providing a way to comparatively evaluate systems on a common data set and task. In 1999, a question answering "track" was introduced. A set of "open domain" news articles is provided in advance, and at evaluation time, questions pertaining to the text are given, to be answered automatically. Human assessors then pool the automatic results and decide which answers are correct. Previous innovations fall into a few categories:
Our system [3] can be decomposed into five components: data indexing, question analysis, passage retrieval, candidate generation, and answer selection. In the data indexing and passage retrieval components, we have explored several query expansion and document retrieval methods, but their effect on question answering performance is still unclear. In the question analysis component, we identify an expected answer type, and possible paraphrases of the question. During candidate generation, we look for structures that indicate a possible answer. During answer selection, we look for support for each candidate (e.g., from the Web), and choose the set of answers to return. The question answering track is separated into four kinds of questions:
We have fielded separate systems for each kind of question, and we are working to integrate our various strategies into a comprehensive package, and into the START system. [1] [2] Research SupportThis work is supported in part by the Advanced Research and Development Activity as part of the AQUAINT Phase II research program. References:[1] Boris Katz. Using English for Indexing and Retrieving. In Proceedings of the 1st RIAO Conference on User-Oriented Content-Based Text and Image Handling (RIAO '88), 1988. [2] Boris Katz. Annotating the World Wide Web using Natural Language. In Proceedings of the Conference on the Computer-Assisted Searching on the Internet, RIAO97, 1997. [3] Boris Katz, Gregory Marton, Gary Borchardt, Alexis Brownell, Sue Felshin, Daniel Loreto, Jesse Louis-Rosenberg, Ben Lu, Federico Mora, Stephan Stiller, Ozlem Uzuner, and Angela Wilcox. External Knowledge for Question Answering. In Proceedings of the Twelfth Text REtrieval Conference (TREC 2005), November 2005. [4] Jimmy Lin and Boris Katz. Question Answering from the Web using Knowledge Annotation and Knowledge Mining Techniques. In Proceedings of the 12th International Conference on Information and Knowledge Management (CIKM 2003), November 2003. [5] Gregory A. Marton. Nuggeteer: Automatic Nugget-Based Evaluation using Descriptions and Judgements. In Proceedings of NAACL/HLT, 2006. [6] Gregory A. Marton. Nuggeteer: Automatic Nugget-Based Evaluation using Descriptions and Judgements. Technical Report 1721.1/30604, 2006. [7] Ellen Voorhees. Overview of the TREC 2005 question answering track. NIST publication, 2005. |
||||
|