Adam Kalai

TILOS Seminar: Unlearnable Facts Cause Hallucinations in Pretrained Language Models

11am PST | Wednesday, January 29, 2025

Adam Tauman Kalai, OpenAI

Abstract: Pretrained language models (LMs) tend to preserve many qualities present in their training data, such as grammaticality, formatting, and politeness. However, for specific types of factuality, even LMs pretrained on factually correct statements tend to produce falsehoods at high rates. We explain these “hallucinations” by drawing a connection to binary classification, enabling us to leverage insights from supervised learning. We prove that pretrained LMs (which are “calibrated”) fail to mimic criteria that cannot be learned. Our analysis explains why pretrained LMs hallucinate on facts such as people’s birthdays but not on systematic facts such as even vs. odd numbers.

Of course, LM pretraining is only one stage in the development of a chatbot, and thus hallucinations are *not* inevitable in chatbots.

This is joint work with Santosh Vempala.


Adam Tauman Kalai is a Research Scientist at OpenAI working on AI Safety and Ethics. He has worked in Algorithms, Fairness, Machine Learning Theory, Game Theory, and Crowdsourcing. He received his PhD from Carnegie Mellon University. He has served as an Assistant Professor at Georgia Tech and TTIC, and is on the science team of the whale-translation Project CETI. He has co-chaired AI and crowdsourcing conferences and has numerous honors, most notably the Majulook prize.

  • 00

    days

  • 00

    hours

  • 00

    minutes

  • 00

    seconds

Local Time

  • Timezone: America/New_York
  • Date: 29 Jan 2025
  • Time: 14:00 - 15:30

Location

HDSI 123 and Virtual
3234 Matthews Ln, La Jolla, CA 92093

Organizer

TILOS

Other Organizers

Halicioglu Data Science Institute
Website
https://datascience.ucsd.edu/

Speaker