About Me
I am currently studying Artificial Intelligence (AI) at the University of Groningen, focusing strongly on AI Alignment. I believe that within a worringly short time frame AI will become more powerful, and currently, by default, more uninterpretable. Combined with the enormous incentives to prioritize capabilities over safety, this creates a serious risk of developing systems we do not fully understand and cannot reliably control.
I am actively exploring how we can ensure this goes well, both at a high level and through technical research. On the high-level side I do this by leading the Neuralignment research & discussion group at the AI Inititative Groningen, where we explore new research avenues. On the technical side I'm currently conducting research in mechanistic interpretability where I investigate how to control frontier models using representation engineering.
My Research Interests
Technical AI Safety
- Mechanistic Interpretability
- Computational Neuroscience
- AI Alignment
Skills
- Deep Learning
- PyTorch
- Research
- Python
My CV
For my complete educational background, work experience, and research projects, please check out my CV:

Location
Groningen, Netherlands
Education
BSc in Artificial Intelligence
University of Groningen
Research Groups
AI Safety Initiative Groningen