Alexander Müller

I'm an AI Alignment researcher and student at the University of Groningen, focusing on mechanistic interpretability and ensuring AI systems remain safe and aligned with human values.

Alexander Müller

Research Interests

Mechanistic Interpretability

Understanding how neural networks work internally is crucial for ensuring control and alignment of increasingly powerful AI systems.

Computational Neuroscience

Exploring how insights from neuroscience can inform safer and more interpretable AI systems through understanding both biological and artificial neural networks.

Involvement

Neuralignment

I lead the Neuralignment research & discussion group at the AI Initiative Groningen, where we explore new research avenues in AI safety and alignment.

Latest Writing