Calibration-Tuning: Teaching Large Language Models to Know What They Don’t Know

Sanyam Kapoor, Nate Gruver, Manley Roberts, Arka Pal, Samuel Dooley, Micah Goldblum, Andrew Gordon Wilson

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

Large language models are increasingly deployed for high-stakes decision making, for example in financial and medical applications. In such applications, it is imperative that we be able to estimate our confidence in the answers output by a language model in order to assess risks. Although we can easily compute the probability assigned by a language model to the sequence of tokens that make up an answer, we cannot easily compute the probability of the answer itself, which could be phrased in numerous ways. While other works have engineered ways of assigning such probabilities to LLM outputs, a key problem remains: existing language models are poorly calibrated, often confident when they are wrong or unsure when they are correct. In this work, we devise a protocol called calibration tuning for finetuning LLMs to output calibrated probabilities. Calibration-tuned models demonstrate superior calibration performance compared to existing language models on a variety of question-answering tasks, including open-ended generation, without affecting accuracy. We further show that this ability transfers to new domains outside of the calibration-tuning train set.

Original languageEnglish (US)
Title of host publicationUncertaiNLP 2024 - Workshop on Uncertainty-Aware NLP, Proceedings of the Workshop
EditorsRaul Vazquez, Hande Celikkanat, Dennis Ulmer, Jorg Tiedemann, Swabha Swayamdipta, Wilker Aziz, Barbara Plank, Barbara Plank, Joris Baan, Marie-Catherine de Marneffe
PublisherAssociation for Computational Linguistics (ACL)
Pages1-14
Number of pages14
ISBN (Electronic)9798891760752
StatePublished - 2024
Event1st Workshop on Uncertainty-Aware NLP, UncertaiNLP 2024 - St. Julian's, Malta
Duration: Mar 22 2024 → …

Publication series

NameUncertaiNLP 2024 - Workshop on Uncertainty-Aware NLP, Proceedings of the Workshop

Conference

Conference1st Workshop on Uncertainty-Aware NLP, UncertaiNLP 2024
Country/TerritoryMalta
CitySt. Julian's
Period3/22/24 → …

ASJC Scopus subject areas

  • Computational Theory and Mathematics
  • Software
  • Linguistics and Language

Fingerprint

Dive into the research topics of 'Calibration-Tuning: Teaching Large Language Models to Know What They Don’t Know'. Together they form a unique fingerprint.

Cite this