About Me

I am currently studying Artificial Intelligence (AI) at the University of Groningen, focusing strongly on AI Alignment. I believe that within a worringly short time frame AI will become more powerful, and currently, by default, more uninterpretable. Combined with the enormous incentives to prioritize capabilities over safety, this creates a serious risk of developing systems we do not fully understand and cannot reliably control.

I am actively exploring how we can ensure this goes well, both at a high level and through technical research. On the high-level side I do this by leading the Neuralignment research & discussion group at the AI Inititative Groningen, where we explore new research avenues. On the technical side I'm currently conducting research in mechanistic interpretability where I investigate how to control frontier models using representation engineering.

My Research Interests

Technical AI Safety

  • Mechanistic Interpretability
  • Computational Neuroscience
  • AI Alignment

Skills

  • Deep Learning
  • PyTorch
  • Research
  • Python

My CV

For my complete educational background, work experience, and research projects, please check out my CV:

Alexander Müller

Location

Groningen, Netherlands

Education

BSc in Artificial Intelligence

University of Groningen

Research Groups

AI Safety Initiative Groningen