I am a postdoctoral researcher at Mila - Quebec AI Institute and a postdoctoral fellow at McGill University in MontrΓ©al π¨π¦. Before that, I was a postdoctoral researcher at the Language Science and Technology Department of Saarland University in SaarbrΓΌcken π©πͺ, where I also did my PhD in Computer Science.
The goal of my research is to enable reliable, controllable, and trustworthy Natural Language Processing (NLP) systems, particularly the Large Language Models (LLMs) that millions interact with daily. LLMs require significant adaptation (also known as fine-tuning or post-training) to become specialized, safe, and aligned with specific requirements after pre-training. My research program centers on building a fundamental, scientific understanding of this crucial adaptation stage.
For our work, my collaborators and I have received a Best Paper Award π at COLING 2022, the Best Theme Paper Award π at ACL 2023, and the Most Interesting Paper Award π at the BabyLM Challenge 2023.
Latest News
-
Check out our new preprint on tracing value alignment during LLM post-training π.
-
I will be attending COLM 2025 in Montreal and present our work on Unlearning π¨π¦.
-
We have a new paper on the impact of data frequency on LLM unlearning. Surprise: Not all data is unlearned equally π!
-
Check out our new preprint on our analysis of the reasoning chains of DeepSeek-R1 π.
-
Our workshop on Actionable Interpretability π has been accepted by ICML 2025 π!
-
I attended the Simons Institute workshop on The Future of Language Models and Transformers in Berkeley πΊπΈ.
