TY - JOUR
T1 - Qed
T2 - A framework and dataset for explanations in question answering
AU - Lamm, Matthew
AU - Palomaki, Jennimaria
AU - Alberti, Chris
AU - Andor, Daniel
AU - Choi, Eunsol
AU - Soares, Livio Baldini
AU - Collins, Michael
N1 - Publisher Copyright:
© 2021 Association for Computational Linguistics.
PY - 2021/8/2
Y1 - 2021/8/2
N2 - A question answering system that in addition to providing an answer provides an explanation of the reasoning that leads to that answer has potential advantages in terms of debuggabil-ity, extensibility, and trust. To this end, we propose QED, a linguistically informed, extensible framework for explanations in question answering. A QED explanation specifies the relationship between a question and answer according to formal semantic notions such as referential equality, sentencehood, and entailment. We describe and publicly release an expert-annotated dataset of QED explanations built upon a subset of the Google Natural Questions dataset, and report baseline models on two tasks—post-hoc explanation generation given an answer, and joint question answering and explanation generation. In the joint setting, a promising result suggests that training on a relatively small amount of QED data can improve question answering. In addition to describing the formal, language-theoretic motivations for the QED approach, we describe a large user study showing that the presence of QED explanations significantly improves the ability of untrained raters to spot errors made by a strong neural QA baseline.
AB - A question answering system that in addition to providing an answer provides an explanation of the reasoning that leads to that answer has potential advantages in terms of debuggabil-ity, extensibility, and trust. To this end, we propose QED, a linguistically informed, extensible framework for explanations in question answering. A QED explanation specifies the relationship between a question and answer according to formal semantic notions such as referential equality, sentencehood, and entailment. We describe and publicly release an expert-annotated dataset of QED explanations built upon a subset of the Google Natural Questions dataset, and report baseline models on two tasks—post-hoc explanation generation given an answer, and joint question answering and explanation generation. In the joint setting, a promising result suggests that training on a relatively small amount of QED data can improve question answering. In addition to describing the formal, language-theoretic motivations for the QED approach, we describe a large user study showing that the presence of QED explanations significantly improves the ability of untrained raters to spot errors made by a strong neural QA baseline.
UR - http://www.scopus.com/inward/record.url?scp=85117998888&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85117998888&partnerID=8YFLogxK
U2 - 10.1162/tacl_a_00398
DO - 10.1162/tacl_a_00398
M3 - Article
AN - SCOPUS:85117998888
SN - 2307-387X
VL - 9
SP - 790
EP - 806
JO - Transactions of the Association for Computational Linguistics
JF - Transactions of the Association for Computational Linguistics
ER -