Exploring the Knowledge Mismatch Hypothesis: Hallucination Propensity in Small Models Fine-tuned on Data from Larger Models

Phil Wee, Riyadh Baghdadi

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

Recently, there has been an explosion of large language models created through fine-tuning with data from larger models. These small models able to produce outputs that appear qualitatively similar to significantly larger models. However, one of the key limitations that have been observed with these models is their propensity to hallucinate significantly more often than larger models. In particular, they have been observed to generate coherent outputs that involve factually incorrect information and spread misinformation, toxicity, and stereotypes. There are many potential causes of hallucination, of which, one hypothesis is that fine-tuning a model on data produced by a larger model leads to a knowledge mismatch which contributes to hallucination. In particular, it is hypothesized that there is a mismatch between the knowledge that is fed to the model to fine-tune it and the knowledge that is already present in the graph. Fine-tuning the model on data that has such mismatch could contribute to an increased propensity to hallucinate. We show that on an unseen test set, a smaller model fine-tuned on data generated from a larger model produced more wrong answers when compared to models fine-tuned on data created by the small model, which confirms the hypothesis.

Original languageEnglish (US)
Title of host publicationProceedings - 2024 IEEE/ACM International Conference on Big Data Computing, Applications and Technologies, BDCAT 2024
PublisherInstitute of Electrical and Electronics Engineers Inc.
Pages258-263
Number of pages6
ISBN (Electronic)9798350367300
DOIs
StatePublished - 2024
Event11th IEEE/ACM International Conference on Big Data Computing, Applications and Technologies, BDCAT 2024 - Sharjah, United Arab Emirates
Duration: Dec 16 2024Dec 19 2024

Publication series

NameProceedings - 2024 IEEE/ACM International Conference on Big Data Computing, Applications and Technologies, BDCAT 2024

Conference

Conference11th IEEE/ACM International Conference on Big Data Computing, Applications and Technologies, BDCAT 2024
Country/TerritoryUnited Arab Emirates
CitySharjah
Period12/16/2412/19/24

Keywords

  • evaluation
  • fine-tuning
  • hallucination
  • large language models

ASJC Scopus subject areas

  • Artificial Intelligence
  • Computer Science Applications
  • Information Systems
  • Information Systems and Management
  • Safety, Risk, Reliability and Quality
  • Modeling and Simulation
  • Health Informatics

Fingerprint

Dive into the research topics of 'Exploring the Knowledge Mismatch Hypothesis: Hallucination Propensity in Small Models Fine-tuned on Data from Larger Models'. Together they form a unique fingerprint.

Cite this