Narrative Question Answering with Cutting-Edge Open-Domain QA Techniques: A Comprehensive Study
Author(s) -
Xiangyang Mou,
Chenghao Yang,
Mo Yu,
Bingsheng Yao,
Xiaoxiao Guo,
Saloni Potdar,
Hui Su
Publication year - 2021
Publication title -
transactions of the association for computational linguistics
Language(s) - English
Resource type - Journals
ISSN - 2307-387X
DOI - 10.1162/tacl_a_00411
Subject(s) - computer science , benchmark (surveying) , task (project management) , question answering , event (particle physics) , domain (mathematical analysis) , narrative , enhanced data rates for gsm evolution , data science , natural language processing , information retrieval , artificial intelligence , linguistics , philosophy , physics , mathematics , management , geodesy , quantum mechanics , economics , geography , mathematical analysis
Recent advancements in open-domain question answering (ODQA), that is, finding answers from large open-domain corpus like Wikipedia, have led to human-level performance on many datasets. However, progress in QA over book stories (Book QA) lags despite its similar task formulation to ODQA. This work provides a comprehensive and quantitative analysis about the difficulty of Book QA: (1) We benchmark the research on the NarrativeQA dataset with extensive experiments with cutting-edge ODQA techniques. This quantifies the challenges Book QA poses, as well as advances the published state-of-the-art with a ∼7% absolute improvement on ROUGE-L. (2) We further analyze the detailed challenges in Book QA through human studies.1 Our findings indicate that the event-centric questions dominate this task, which exemplifies the inability of existing QA models to handle event-oriented scenarios.
Accelerating Research
Robert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom
Address
John Eccles HouseRobert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom