Skip to Content

Sponsors

No results

Keywords

No results

Types

No results

Search Results

Events

No results
Search events using: keywords, sponsors, locations or event type
When / Where
All occurrences of this event have passed.
This listing is displayed for historical purposes.

Presented By: Department of Statistics

Statistics Department Seminar Series: Will Wei Sun, Associate Professor, Department of Quantitative Methods, Department of Statistics (by courtesy), Purdue University

"Aligning Large Language Models with Heterogeneous Human Feedback: When Statistics Meets LLMs"

Will Wei Sun Will Wei Sun
Will Wei Sun
Abstract: Reinforcement learning from human feedback (RLHF) has emerged as the leading approach to aligning large language models (LLMs) with human preferences. Despite its success, two challenges remain fundamental: feedback is costly and heterogeneous across annotators, and the resulting reward models often lack principled measures of uncertainty. This talk presents recent advances that address these challenges by integrating tools from optimal design and statistical inference into the RLHF framework. First, I introduce a dual active learning approach, inspired by optimal design, that adaptively selects both conversations and annotators to maximize information gain, improving the efficiency of limited feedback budgets. Second, I present a framework for uncertainty quantification in reward learning, enabling valid statistical comparisons across LLM models and more reliable best-of-n alignment policies. Together, these results illustrate how statistics can help trustworthy and data-efficient LLM alignment.
Will Wei Sun Will Wei Sun
Will Wei Sun

Explore Similar Events

  •  Loading Similar Events...

Keywords


Back to Main Content