Postdoctoral Fellow
Computer Science, Stanford University
Computer Science, Stanford University
Ph.D. in Computer Science, 2018-2023
Washington State University
M.Sc. in Electrical Engineering, 2017-2018
University of Idaho
I am a Postdoctoral fellow in Computer Science at Stanford University and Stanford Data Science working with Professor Stefano Ermon and Professor Barbara Engelhardt. I obtained my PhD in Computer Science from Washington State University where I was advised by Professor Jana Doppa. I am working at the intersection of AI, generative models, and scientific discovery. My recent work focuses on scalable methods for LLM alignment, active learning, and diffusion-based optimization, with scientific applications ranging from antibiotic discovery to high-throughput screening. I design algorithms that reduce annotation and computation costs while improving performance in safety-critical, resource-constrained domains. My projects include preference-guided diffusion models, active exploration and efficient sampling strategies for LLMs, and automated multi-objective pipelines for antibiotic discovery validated through wet lab experiments. My broader expertise spans Bayesian optimization, uncertainty quantification, and efficient reasoning algorithms for sequential decision-making under uncertainty.
Collecting preference data for LLM alignment is challenging and costly, especially in scientific or safety-critical domains where expert input is required. We propose a method to pre-select high-impact preference pairs using Sharpe Ratio–based gradient analysis, reducing annotation costs while improving win rates.
[Paper (COLM 2025)][Code][SDS lightning Talk]
We tackle the high cost of human feedback in LLM alignment by framing preference learning as an active contextual dueling bandit problem. Our exploration-based algorithm efficiently selects where to query for preferences, with provable regret bounds and strong empirical gains across several LLMs and real-world datasets, including two new benchmarks we introduce: Jeopardy! and Haikus.
We introduce a preference-guided diffusion model for offline multi-objective optimization that generates diverse, Pareto-optimal solutions beyond the training data. By guiding generation with a dominance-based classifier and explicitly promoting Pareto diversity, our method offers a novel generative, surrogate-free solution to inverse design problems.
We develop deep learning pipelines that combine graph neural networks and diffusion-based generative models to discover antibiotic candidates with novel mechanisms of action. Our models prioritize potency, low toxicity, and structural diversity, enabling both the screening and generation of promising molecules, validated through real-world wet lab experiments.
[Paper (coming soon)]
We propose the first active learning approach tailored to derivative-based global sensitivity analysis (DGSMs), using Gaussian processes to guide costly black-box evaluations. By targeting DGSM uncertainty and information gain, our method substantially boosts sample efficiency in scientific and engineering tasks with limited evaluation budgets.
We introduce an output space entropy (OSE) search framework for multi-objective Bayesian optimization, selecting experiments that maximize information gain per resource spent. Our approach generalizes across single- and multi-fidelity, constrained, and continuous-fidelity settings—delivering more accurate Pareto fronts with fewer expensive evaluations.