Will Schwarzer
Hey there, I’m Will. I’m a PhD student at the University of Massachusetts Amherst, where I work with Philip Thomas, Bruno Castro da Silva and Scott Niekum. My long-term research goal is to create robust and aligned agents that can perform largely autonomous research in AI and the sciences. Right now I’m working on agentic evaluation for LLM agents, RL and post-training for safety and alignment, and robustness in generative models and agent behavior. More broadly, I’m interested in reliable long-horizon agents: systems that can carry out complex projects over many steps, keep honest accounts of partial progress and failure, and drive themselves toward genuinely correct solutions.
Fun fact: this headshot was taken at an excellent studio in Beijing while I was studying Mandarin there. I’m always happy to chat in Chinese. Aside from languages, I also enjoy playing the piano and rock climbing. I particularly love exploring climbing gyms in new countries (tally so far: three in Beijing, one in Bangkok, one in Istanbul, one in Fort William, Scotland, one in London, and a few in Montreal).
Note to prospective undergraduate and master’s collaborators: please read this before reaching out.
news
| Sep 23, 2025 | I’m proud to announce that our preprint Evaluation-Aware Reinforcement Learning has been published on arXiv! |
|---|---|
| Sep 19, 2025 | I’m pleased to have served as a reviewer for the Reliable ML from Unreliable Data workshop at NeurIPS 2025. |
| Sep 08, 2025 | I’ve joined Amazon as an Applied Scientist Intern, working on agentic evaluation frameworks for language model agents. |
| May 09, 2025 | I’m pleased to announce that our paper Reinforcement Learning from Human Feedback with High-Confidence Safety Guarantees has been accepted for publication at RLC 2025! |
| Apr 04, 2025 | I’m pleased to have served as a Senior Reviewer for RLC 2025! (Last year, I served as a Technical Reviewer for RLC 2024.) |
| Mar 16, 2025 | I’m proud to announce that our preprint Are Deep Speech Denoising Models Robust to Adversarial Noise? has been published on arXiv! |
| Feb 25, 2025 | I’m proud to announce that our preprint Supervised Reward Inference has been published on arXiv! |