When: Oct 23 2025 @ 10:30 AM
Where: B-17 Hackerman Hall
Categories:
Department of Computer Science Distinguished Lecture Series.

Refreshments are available starting at 10:30 a.m. The seminar will begin at 10:45 a.m.

Abstract

As AI models become increasingly powerful, it is an attractive proposition to use them in important decision-making pipelines in collaboration with human decision-makers. But how should a human being and a machine learning model collaborate to reach decisions that are better than either of them could achieve on their own? If the human and the AI model were perfect Bayesians, operating in a setting with a commonly known and correctly specified prior, Aumann’s classical agreement theorem would give us one answer: They could engage in conversation about the task at hand, and their conversation would be guaranteed to converge to (accuracy-improving) agreement. This classical result, however, would require making many implausible assumptions, both about the knowledge and computational power of both parties. We show how to recover similar (and more general) results using only computationally and statistically tractable assumptions, which substantially relax full Bayesian rationality. In the second part of the talk, we go on to consider a more difficult problem: that the AI model might be acting at least in part to advance the interests of its designer, rather than the interests of its user, which might be in tension. We show how market competition between different AI providers can mitigate this problem assuming only a mild “market alignment” assumption—that the user’s utility function lies in the convex hull of the AI providers’ utility functions—even when no single provider is well aligned. In particular, we show that in all Nash equilibria of the AI providers under this market alignment condition, the user is able to advance her own goals as well as they could have in collaboration with a perfectly aligned AI model.

This talk describes the results of three papers—Tractable Agreement Protocols (2025 ACM Symposium on Theory of Computing), Collaborative Prediction: Tractable Information Aggregation via Agreement (ACM-SIAM Symposium on Discrete Algorithms), and Emergent Alignment from Competition—which are joint works with Natalie Collina, Ira Globus-Harris, Surbhi Goel, Varun Gupta, Emily Ryu, and Mirah Shi.

Speaker Biography

Aaron Roth is the Henry Salvatori Professor of Computer and Cognitive Science in the Department of Computer and Information Science at the University of Pennsylvania, with a secondary appointment in the Wharton Department of Statistics and Data Science. He is affiliated with the Warren Center for Network and Data Sciences and is the co-director of the Networked & Social Systems Engineering program. Roth is also an Amazon Scholar at Amazon Web Services. He is the recipient of the Hans Sigrist Prize, a Presidential Early Career Award for Scientists and Engineers, a Sloan Research Fellowship, an NSF CAREER Award, and research awards from Yahoo, Amazon, and Google. Roth’s research focuses on the algorithmic foundations of data privacy, algorithmic fairness, game theory, learning theory, and machine learning. He has authored two books—The Algorithmic Foundations of Differential Privacy with Cynthia Dwork and The Ethical Algorithm with Michael Kearns.

Zoom link >>