TY - GEN
T1 - Contrastive Learning to Improve Retrieval for Real-world Fact Checking
AU - Sriram, Aniruddh
AU - Xu, Fangyuan
AU - Choi, Eunsol
AU - Durrett, Greg
N1 - Publisher Copyright:
© 2024 Association for Computational Linguistics.
PY - 2024
Y1 - 2024
N2 - Recent work on fact-checking addresses a realistic setting where models incorporate evidence retrieved from the web to decide the veracity of claims. A bottleneck in this pipeline is in retrieving relevant evidence: traditional methods may surface documents directly related to a claim, but fact-checking complex claims requires more inferences. For instance, a document about how a vaccine was developed is relevant to addressing claims about what it might contain, even if it does not address them directly. We present Contrastive Fact-Checking Reranker (CFR), an improved retriever for this setting. By leveraging the AVeriTeC dataset, which annotates subquestions for claims with human written answers from evidence documents, we fine-tune Contriever with a contrastive objective based on multiple training signals, including distillation from GPT-4, evaluating subquestion answers, and gold labels in the dataset. We evaluate our model on both retrieval and end-to-end veracity judgments about claims. On the AVeriTeC dataset, we find a 6% improvement in veracity classification accuracy. We also show our gains can be transferred to FEVER, ClaimDecomp, HotpotQA, and a synthetic dataset requiring retrievers to make inferences.
AB - Recent work on fact-checking addresses a realistic setting where models incorporate evidence retrieved from the web to decide the veracity of claims. A bottleneck in this pipeline is in retrieving relevant evidence: traditional methods may surface documents directly related to a claim, but fact-checking complex claims requires more inferences. For instance, a document about how a vaccine was developed is relevant to addressing claims about what it might contain, even if it does not address them directly. We present Contrastive Fact-Checking Reranker (CFR), an improved retriever for this setting. By leveraging the AVeriTeC dataset, which annotates subquestions for claims with human written answers from evidence documents, we fine-tune Contriever with a contrastive objective based on multiple training signals, including distillation from GPT-4, evaluating subquestion answers, and gold labels in the dataset. We evaluate our model on both retrieval and end-to-end veracity judgments about claims. On the AVeriTeC dataset, we find a 6% improvement in veracity classification accuracy. We also show our gains can be transferred to FEVER, ClaimDecomp, HotpotQA, and a synthetic dataset requiring retrievers to make inferences.
UR - http://www.scopus.com/inward/record.url?scp=85214147146&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85214147146&partnerID=8YFLogxK
M3 - Conference contribution
AN - SCOPUS:85214147146
T3 - FEVER 2024 - 7th Fact Extraction and VERification Workshop, Proceedings of the Workshop
SP - 264
EP - 279
BT - FEVER 2024 - 7th Fact Extraction and VERification Workshop, Proceedings of the Workshop
A2 - Schlichtkrull, Michael
A2 - Chen, Yulong
A2 - Whitehouse, Chenxi
A2 - Deng, Zhenyun
A2 - Akhtar, Mubashara
A2 - Aly, Rami
A2 - Guo, Zhijiang
A2 - Christodoulopoulos, Christos
A2 - Cocarascu, Oana
A2 - Mittal, Arpit
A2 - Thorne, James
A2 - Vlachos, Andreas
PB - Association for Computational Linguistics (ACL)
T2 - 7th Fact Extraction and VERification Workshop, FEVER 2024
Y2 - 15 November 2024
ER -