Vinith Suriyakumar
MIT EECS, LIDS, IMES, Bridgewater Associates
I am a fifth year PhD student at MIT EECS where I am advised by Dr. Ashia Wilson and Dr. Marzyeh Ghassemi. I also collaborate frequently with Dr. Dylan Hadfield-Menell. I’m supported through a fellowship from Bridgewater Associates and AIA Lab led by Dr. Jas Sekhon where I also conduct research part-time.
Broadly, I’m interested in the privacy, security, and safety of machine learning. The goal of my research is to prevent the misuse of foundation models, especially when it comes to open-sourcing these models. I work on questions related to uncovering risks of open-weight models, applying intepretability to understand these risks, and building safeguards to address these risks.

news
Sep 18, 2025 | Our work uncovering and formalizing spurious correlations originating from syntactic shortcuts in LLMs, Learning the Wrong Lessons: Syntatic-Domain Shortcuts in Language Models, was awarded a Spotlight at NeurIPS 2025! This work provides a new perspective on different forms of memorization and uncovers a new type of jailbreak in LLMs. |
---|---|
Jun 3, 2024 | Our work on developing a philosophical framework to better reason about the impact of algorithmic decision-making on equal opportunity, Algorithmic Pluralism: A Structural Approach Towards Equal Opportunity, was awarded Best Paper at FAccT 2024! |
Jan 15, 2024 | Our work on developing an efficient method for estimating privacy with only one training run in federated learning setups, One-Shot Empirical Privacy Estimation for Federated Learning, was awarded an Oral presentation at ICLR 2024! |
Apr 29, 2023 | Our work on formalizing and analyzing how to fairly use group attributes in prediction models, When Personalization Harms: Reconsidering the Use of Group Attributes in Prediction, was awarded an Oral presentation at ICML 2023! |
May 14, 2022 | I’m excited to announce that our work from my Summer 2021 internship at Google on Public Data-Assisted Mirror Descent for Private Model Training was accepted to ICML 2022! |