Premium
CONTEXTUAL LANGUAGE MODELS FOR RANKING ANSWERS TO NATURAL LANGUAGE DEFINITION QUESTIONS
Author(s) -
Figueroa Alejandro,
Atkinson John
Publication year - 2012
Publication title -
computational intelligence
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 0.353
H-Index - 52
eISSN - 1467-8640
pISSN - 0824-7935
DOI - 10.1111/j.1467-8640.2012.00426.x
Subject(s) - computer science , dependency (uml) , natural language processing , ranking (information retrieval) , question answering , context (archaeology) , artificial intelligence , natural language , semantics (computer science) , language model , natural language understanding , entity linking , information retrieval , knowledge base , paleontology , biology , programming language
Question–answering systems make good use of knowledge bases (KBs, e.g., Wikipedia ) for responding to definition queries. Typically, systems extract relevant facts from articles regarding the question across KBs, and then they are projected into the candidate answers. However, studies have shown that the performance of this kind of method suddenly drops, whenever KBs supply narrow coverage. This work describes a new approach to deal with this problem by constructing context models for scoring candidate answers, which are, more precisely, statistical n ‐gram language models inferred from lexicalized dependency paths extracted from Wikipedia abstracts. Unlike state‐of‐the‐art approaches, context models are created by capturing the semantics of candidate answers (e.g., “ novel ,”“ singer ,”“ coach ,” and “ city ”). This work is extended by investigating the impact on context models of extra linguistic knowledge such as part‐of‐speech tagging and named‐entity recognition. Results showed the effectiveness of context models as n ‐gram lexicalized dependency paths and promising context indicators for the presence of definitions in natural language texts.