Alexander Müller
Co-Director of AISIG and MSc AI student at the University of Groningen. I work on making sure AI goes well, through technical research and by building institutions that take AI safety seriously.
What I'm Currently Working On
Game Theory & AI Safety
Exploring how game-theoretic ideas (coordination problems, incentive structures, equilibrium selection) can inform AI alignment. See my post on Moloch.
LLMs & EU Legislation
Researching how well and in what contexts LLM agents follow EU law, through the EU-Agent-Bench project.
Growing AI Safety Initiatives
Co-directing AISIG and pushing for more local AI Safety Initiatives across the Netherlands. Also working towards a national-level organization.
Writing & Outreach
Publishing on my Substack and AISIG's Substack about AI safety, philosophy, books, and more.
Research Interests
Mechanistic Interpretability
Understanding how neural networks work internally. If we can't look inside these systems and understand what they're doing, we can't trust them. See this paper.
Game Theory & Coordination
How game-theoretic frameworks can help us think about multi-agent alignment and cooperation, i.e., how to defeat Moloch.
Computational Neuroscience
What can biological neural networks teach us about artificial ones, and vice versa?
AI Governance & Policy
EU law compliance, governance frameworks, and building institutions that take AI safety seriously. See this paper.
AI Alignment
How do we make sure AI systems actually do what we want? The problem is harder than it sounds. See this post.
AI Safety Initiative Groningen
I co-direct AISIG, an interdisciplinary organization focused on ensuring the interaction of AI with society is just and positive. What we do:
Education
- • AI Safety, Ethics & Society course: 9+ cohorts, 100+ students
- • Governance and Technical tracks
- • Discussion groups on Technical AI Alignment and Governance
Research
- • Research Hub with 4 experienced supervisors (Dr. to Asst. Prof.)
- • Papers at NeurIPS, ICLR, and more
- • Regular hackathons with podium placements
Collaborations
- • Municipality Westerkwartier: AI governance framework
- • Dutch Network for AI Safety
- • Helping AISIs starting up again (e.g., AISI Amsterdam) or through collaboration
Events & Talks
- • Turn.io, EAGx Amsterdam, AiGrunn
- • Samenwerking Noord, TEDx
- • Discussion evenings and social events
Active on LinkedIn, Substack, Instagram, LessWrong, and our website.
Recent Publications
View all →EU-Agent-Bench: Measuring Illegal Behavior of LLM Agents Under EU Law
Lichkovski, Müller, Ibrahim, Mhundwa · arXiv 2025
Uncovering Internal Prediction Mechanisms of Transformer-Based Chemical Foundation Models
Müller, Cardenas-Cartagena, Pollice · ChemRxiv 2025
From Steering Vectors to Conceptors: Compositional Affine Activation Steering for LLMs
Abreu, Postmus, Müller, et al. · 2025
Collective Deliberation for Safer CBRN Decisions: A Multi-Agent LLM Debate Pipeline
Müller, Golicins, Lesnic · AISIG & Apart Research, 2025
Latest Writing
View all →I write on my Substack and AISIG's Substack about AI safety, philosophy, books, and more. Although there are some posts on this website, I don't keep them updated, so please check out the Substacks!
Why Smarter Doesn't Mean Kinder: Orthogonality and Instrumental Convergence
September 23, 2025
Why Care About AI Safety? (AISIG)
August 10, 2025
Quotes from Zen & The Art of Motorcycle Maintenance
July 23, 2025