Loading Events

« All Events

  • This event has passed.

TILOS Seminar: Unlearnable Facts Cause Hallucinations in Pretrained Language Models

January 29 @ 11:00 - 12:30

Adam Tauman Kalai, OpenAI

Abstract: Pretrained language models (LMs) tend to preserve many qualities present in their training data, such as grammaticality, formatting, and politeness. However, for specific types of factuality, even LMs pretrained on factually correct statements tend to produce falsehoods at high rates. We explain these “hallucinations” by drawing a connection to binary classification, enabling us to leverage insights from supervised learning. We prove that pretrained LMs (which are “calibrated”) fail to mimic criteria that cannot be learned. Our analysis explains why pretrained LMs hallucinate on facts such as people’s birthdays but not on systematic facts such as even vs. odd numbers.
Of course, LM pretraining is only one stage in the development of a chatbot, and thus hallucinations are *not* inevitable in chatbots.
This is joint work with Santosh Vempala.


Adam Tauman Kalai is a Research Scientist at OpenAI working on AI Safety and Ethics. He has worked in Algorithms, Fairness, Machine Learning Theory, Game Theory, and Crowdsourcing. He received his PhD from Carnegie Mellon University. He has served as an Assistant Professor at Georgia Tech and TTIC, and is on the science team of the whale-translation Project CETI. He has co-chaired AI and crowdsourcing conferences and has numerous honors, most notably the Majulook prize.

Details

Date:
January 29
Time:
11:00 - 12:30
Event Categories:
,

Organizers

TILOS
Halicioglu Data Science Institute

Venue

HDSI 123 and Virtual
3234 Matthews Ln
La Jolla, CA 92093 United States
+ Google Map
View Venue Website