Alexander Müller
I'm an AI Alignment researcher and student at the University of Groningen, focusing on mechanistic interpretability and ensuring AI systems remain safe and aligned with human values.

Research Interests
Mechanistic Interpretability
Understanding how neural networks work internally is crucial for ensuring control and alignment of increasingly powerful AI systems.
Computational Neuroscience
Exploring how insights from neuroscience can inform safer and more interpretable AI systems through understanding both biological and artificial neural networks.
Involvement
Neuralignment
I lead the Neuralignment research & discussion group at the AI Initiative Groningen, where we explore new research avenues in AI safety and alignment.