William Overman

PhD Student, Stanford

wpo@stanford.edu

Bio

My research develops approaches to AI safety and alignment [NeurIPS'24, NeurIPS'25], with a particular emphasis on helping humans effectively supervise AI systems as they become increasingly capable [arXiv'25]. I draw on tools from reinforcement learning, uncertainty quantification, causal inference, and game theory to provide rigorous guarantees for safe AI deployment and decision-making.

I’m a Ph.D. student in Operations, Information, and Technology at Stanford Graduate School of Business, advised by Mohsen Bayati. Before Stanford, I was a visiting researcher at the Institue for Basic Science in South Korea and earned an M.Sc. in Computer Science from UC Irvine. I graduated from Caltech in 2020 with a double major in Mathematics and Computer Science.

In addition to my primary focus on AI safety and alignment, my research also explores reinforcement learning [ICLR'22, RLC'25], causal inference [NeurIPS'24, arXiv'25], and AI applications in healthcare [ SSRN'23,arXiv'25]. Throughout my graduate studies, I have interned at Uber, where I have applied my research work in RL and causal inference to problems in the ridesharing and delivery marketplaces.

Publications

indicates equal contribution. indicates equal contribution, sole student.

Causal Effects with Unobserved Unit Types in Interacting Human–AI Systems

W Overman, S Shirani, M Bayati

arXiv'26: arXiv preprint arXiv:2603.01339. 2026.

The Oversight Game: Learning to Cooperatively Balance an AI Agent's Safety and Autonomy

W Overman, M Bayati

Early version in NeurIPS'25 Workshop: ML×OR Workshop. 2025. (Spotlight Presentation)

Conformal Arbitrage: Risk-Controlled Balancing of Competing Objectives in Language Models

W Overman, M Bayati

NeurIPS'25: Neural Information Processing Systems. 2025.

Can We Validate Counterfactual Estimations in the Presence of General Network Interference?

S Shirani, Y Luo, W Overman, R Xiong, M Bayati

arXiv'25: arXiv preprint arXiv:2502.01106. 2025.
Accepted for Oral Presentation at the Conference on Digital Experimentation @ MIT (CODE@MIT), 2025
Accepted for presentation at the MSOM Technology, Innovation, and Entrepreneurship SIG, 2025.

Aligning Model Properties via Conformal Risk Control

W Overman, JJ Vallon, M Bayati

NeurIPS'24: Neural Information Processing Systems. 2024.

Higher-Order Causal Message Passing for Experimentation with Complex Interference

M Bayati, Y Luo, W Overman, S Shirani, R Xiong

NeurIPS'24: Neural Information Processing Systems. 2024.

Global convergence of multi-agent policy gradient in markov potential games

S Leonardos, W Overman, I Panageas, G Piliouras

ICLR'22: International Conference on Learning Representations. 2022.

Independent natural policy gradient always converges in markov potential games

R Fox, SM McAleer, W Overman, I Panageas

AISTATS'22: Artificial Intelligence and Statistics. 2022.

Causal Effects with Unobserved Unit Types in Interacting Human–AI Systems

W Overman, S Shirani, M Bayati

arXiv'26: arXiv preprint arXiv:2603.01339. 2026.

The Oversight Game: Learning to Cooperatively Balance an AI Agent's Safety and Autonomy

W Overman, M Bayati

Early version in NeurIPS'25 Workshop: ML×OR Workshop. 2025. (Spotlight Presentation)

Conformal Arbitrage: Risk-Controlled Balancing of Competing Objectives in Language Models

W Overman, M Bayati

NeurIPS'25: Neural Information Processing Systems. 2025.

Can We Validate Counterfactual Estimations in the Presence of General Network Interference?

S Shirani, Y Luo, W Overman, R Xiong, M Bayati

arXiv'25: arXiv preprint arXiv:2502.01106. 2025.
Accepted for Oral Presentation at the Conference on Digital Experimentation @ MIT (CODE@MIT), 2025
Accepted for presentation at the MSOM Technology, Innovation, and Entrepreneurship SIG, 2025.

On aligning prediction models with clinical experiential learning: A prostate cancer case study

JJ Vallon, W Overman, W Xu, N Panjwani, X Ling, S Vij, HP Bagshaw, ...

arXiv'25: arXiv preprint arXiv:2509.04053. 2025.

Aligning Model Properties via Conformal Risk Control

W Overman, JJ Vallon, M Bayati

NeurIPS'24: Neural Information Processing Systems. 2024.

Higher-Order Causal Message Passing for Experimentation with Complex Interference

M Bayati, Y Luo, W Overman, S Shirani, R Xiong

NeurIPS'24: Neural Information Processing Systems. 2024.

Beating price of anarchy and gradient descent without regret in potential games

I Sakos, S Leonardos, SA Stavroulakis, W Overman, I Panageas, G Piliouras

ICLR'24: International Conference on Learning Representations. 2024.

Improved Regret Bound for Safe Reinforcement Learning via Tighter Cost Pessimism and Reward Optimism

K Yu, D Lee, W Overman, D Lee

RLC 2025 (Reinforcement Learning Conference).
Journal version: Reinforcement Learning Journal (2025).

Occupancy Prediction with Patient Data: Evaluating Time-Series, Patient-Level Aggregation, and Deep Set Models

SH Kim, W Overman, J Pauphilet, WC Cha

Major Revision at Manufacturing & Service Operations Management (MSOM).

Global convergence of multi-agent policy gradient in markov potential games

S Leonardos, W Overman, I Panageas, G Piliouras

ICLR'22: International Conference on Learning Representations. 2022.

Independent natural policy gradient always converges in markov potential games

R Fox, SM McAleer, W Overman, I Panageas

AISTATS'22: Artificial Intelligence and Statistics. 2022.

Some Ordered Ramsey Numbers of Graphs on Four Vertices

W Overman, JF Alm, K Coffey, C Langhoff

Australasian Journal of Combinatorics, Vol 88(3), 266–281. 2024.

Vitæ

Full Resume in PDF.

Website design from Martin Saveski. Code from this GitHub repo.