Research GroupAI Safety and Alignment
The new AI Safety and Alignment Group focuses on developing technical solutions to reduce risks from general-purpose AI models.
We are particularly interested in alignment of autonomous LLM agents, which are becoming increasingly capable and pose a variety of emerging risks. We are also interested in rigorous AI evaluations and informing the public about the risks and capabilities of frontier AI models. Finally, we aim to advance our understanding of how AI models generalize from training data, which is crucial for ensuring their steerability and alignment with societal values.
Maksym Andriushchenko is a postdoctoral researcher at EPFL and an ELLIS Member. He has worked on AI safety with leading organizations in the field (OpenAI, Anthropic, UK AI Safety Institute, Center for AI Safety, Gray Swan AI). He obtained a PhD in machine learning from EPFL in 2024 advised by Prof. Nicolas Flammarion. His PhD thesis was awarded with the Patrick Denantes Memorial Prize for the best thesis in the CS department of EPFL and was supported by the Google and Open Phil AI PhD Fellowships. He did his MSc at Saarland University and the University of Tübingen, and interned at Adobe Research.