TY - GEN
T1 - Calibration-Tuning
T2 - 1st Workshop on Uncertainty-Aware NLP, UncertaiNLP 2024
AU - Kapoor, Sanyam
AU - Gruver, Nate
AU - Roberts, Manley
AU - Pal, Arka
AU - Dooley, Samuel
AU - Goldblum, Micah
AU - Wilson, Andrew Gordon
N1 - Publisher Copyright:
© 2024 Association for Computational Linguistics.
PY - 2024
Y1 - 2024
N2 - Large language models are increasingly deployed for high-stakes decision making, for example in financial and medical applications. In such applications, it is imperative that we be able to estimate our confidence in the answers output by a language model in order to assess risks. Although we can easily compute the probability assigned by a language model to the sequence of tokens that make up an answer, we cannot easily compute the probability of the answer itself, which could be phrased in numerous ways. While other works have engineered ways of assigning such probabilities to LLM outputs, a key problem remains: existing language models are poorly calibrated, often confident when they are wrong or unsure when they are correct. In this work, we devise a protocol called calibration tuning for finetuning LLMs to output calibrated probabilities. Calibration-tuned models demonstrate superior calibration performance compared to existing language models on a variety of question-answering tasks, including open-ended generation, without affecting accuracy. We further show that this ability transfers to new domains outside of the calibration-tuning train set.
AB - Large language models are increasingly deployed for high-stakes decision making, for example in financial and medical applications. In such applications, it is imperative that we be able to estimate our confidence in the answers output by a language model in order to assess risks. Although we can easily compute the probability assigned by a language model to the sequence of tokens that make up an answer, we cannot easily compute the probability of the answer itself, which could be phrased in numerous ways. While other works have engineered ways of assigning such probabilities to LLM outputs, a key problem remains: existing language models are poorly calibrated, often confident when they are wrong or unsure when they are correct. In this work, we devise a protocol called calibration tuning for finetuning LLMs to output calibrated probabilities. Calibration-tuned models demonstrate superior calibration performance compared to existing language models on a variety of question-answering tasks, including open-ended generation, without affecting accuracy. We further show that this ability transfers to new domains outside of the calibration-tuning train set.
UR - http://www.scopus.com/inward/record.url?scp=85188720827&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85188720827&partnerID=8YFLogxK
M3 - Conference contribution
AN - SCOPUS:85188720827
T3 - UncertaiNLP 2024 - Workshop on Uncertainty-Aware NLP, Proceedings of the Workshop
SP - 1
EP - 14
BT - UncertaiNLP 2024 - Workshop on Uncertainty-Aware NLP, Proceedings of the Workshop
A2 - Vazquez, Raul
A2 - Celikkanat, Hande
A2 - Ulmer, Dennis
A2 - Tiedemann, Jorg
A2 - Swayamdipta, Swabha
A2 - Aziz, Wilker
A2 - Plank, Barbara
A2 - Plank, Barbara
A2 - Baan, Joris
A2 - de Marneffe, Marie-Catherine
PB - Association for Computational Linguistics (ACL)
Y2 - 22 March 2024
ER -