Alexander Müller

Co-Director of AISIG and MSc AI student at the University of Groningen. I work on making sure AI goes well, through technical research and by building institutions that take AI safety seriously.

Alexander Müller

What I'm Currently Working On

Game Theory & AI Safety

Exploring how game-theoretic ideas (coordination problems, incentive structures, equilibrium selection) can inform AI alignment. See my post on Moloch.

LLMs & EU Legislation

Researching how well and in what contexts LLM agents follow EU law, through the EU-Agent-Bench project.

Growing AI Safety Initiatives

Co-directing AISIG and pushing for more local AI Safety Initiatives across the Netherlands. Also working towards a national-level organization.

Writing & Outreach

Publishing on my Substack and AISIG's Substack about AI safety, philosophy, books, and more.

Research Interests

Mechanistic Interpretability

Understanding how neural networks work internally. If we can't look inside these systems and understand what they're doing, we can't trust them. See this paper.

Game Theory & Coordination

How game-theoretic frameworks can help us think about multi-agent alignment and cooperation, i.e., how to defeat Moloch.

Computational Neuroscience

What can biological neural networks teach us about artificial ones, and vice versa?

AI Governance & Policy

EU law compliance, governance frameworks, and building institutions that take AI safety seriously. See this paper.

AI Alignment

How do we make sure AI systems actually do what we want? The problem is harder than it sounds. See this post.

AI Safety Initiative Groningen

I co-direct AISIG, an interdisciplinary organization focused on ensuring the interaction of AI with society is just and positive. What we do:

Education

  • • AI Safety, Ethics & Society course: 9+ cohorts, 100+ students
  • • Governance and Technical tracks
  • • Discussion groups on Technical AI Alignment and Governance

Research

  • • Research Hub with 4 experienced supervisors (Dr. to Asst. Prof.)
  • • Papers at NeurIPS, ICLR, and more
  • • Regular hackathons with podium placements

Collaborations

  • • Municipality Westerkwartier: AI governance framework
  • • Dutch Network for AI Safety
  • • Helping AISIs starting up again (e.g., AISI Amsterdam) or through collaboration

Events & Talks

Active on LinkedIn, Substack, Instagram, LessWrong, and our website.

Recent Publications

View all →

Latest Writing

View all →

I write on my Substack and AISIG's Substack about AI safety, philosophy, books, and more. Although there are some posts on this website, I don't keep them updated, so please check out the Substacks!