@inproceedings{10aa6231a2d14ca1b95ba273a87e3665,
title = "Single-Turn Debate Does Not Help Humans Answer Hard Reading-Comprehension Questions",
abstract = "Current QA systems can generate reasonable-sounding yet false answers without explanation or evidence for the generated answer, which is especially problematic when humans cannot readily check the model{\textquoteright}s answers. This presents a challenge for building trust in machine learning systems. We take inspiration from real-world situations where diffcult questions are answered by considering opposing sides (see Irving et al., 2018). For multiple-choice QA examples, we build a dataset of single arguments for both a correct and incorrect answer option in a debate-style set-up as an initial step in training models to produce explanations for two candidate answers. We use long contexts—humans familiar with the context write convincing explanations for preselected correct and incorrect answers, and we test if those explanations allow humans who have not read the full context to more accurately determine the correct answer. We do not fnd that explanations in our set-up improve human accuracy, but a baseline condition shows that providing human-selected text snippets does improve accuracy. We use these fndings to suggest ways of improving the debate set up for future data collection efforts.",
author = "Alicia Parrish and Harsh Trivedi and Ethan Perez and Angelica Chen and Nikita Nangia and Jason Phang and Bowman, {Samuel R.}",
note = "Funding Information: This project has benefted from fnancial support to SB by Eric and Wendy Schmidt (made by recommendation of the Schmidt Futures program), Samsung Research (under the project Improving Deep Learning using Latent Structure) and Apple. This material is based upon work supported by the National Science Foundation under Grant Nos. 1922658 and 2046556. Any opinions, fndings, and conclusions or recommendations expressed in this material are those of the authors and do not necessarily refect the views of the National Science Foundation. Publisher Copyright: {\textcopyright} 2022 Association for Computational Linguistics.; 1st Workshop on Learning with Natural Language Supervision, LNLS 2022 ; Conference date: 26-05-2022",
year = "2022",
language = "English (US)",
series = "LNLS 2022 - 1st Workshop on Learning with Natural Language Supervision, Proceedings of the Workshop",
publisher = "Association for Computational Linguistics (ACL)",
pages = "17--28",
editor = "Jacob Andreas and Karthik Narasimhan and Aida Nematzadeh",
booktitle = "LNLS 2022 - 1st Workshop on Learning with Natural Language Supervision, Proceedings of the Workshop",
}