MATS mentors are advancing the frontiers of AI alignment, transparency, and security

Alan Cooney leads the Autonomous Systems workstream within the UK's AI Safety Institute. His team is responsible for assessing the capabilities and risks of Frontier AI systems released by AI labs such as OpenAI, Google and Anthropic. Prior to working in AI safety, he was an investment consultant and start-up founder, with his company Skyhook being acquired in 2023. He also completed Stanford’s Machine Learning and Alignment Theory Scholars Programme, where he was supervised by Google DeepMind researcher Neel Nanda.

Focus:
Empirical
Control, Monitoring
Programs:
Janet Egan
Center for a New American Security (CNAS)
,
Senior Fellow and Deputy Director, Technology and National Security

Janet Egan is a Senior Fellow with the Technology and National Security Program at the Center for a New American Security (CNAS). Her research focuses on the national security implications of AI and other emerging technologies, including how compute can be leveraged for the governance of advanced AI systems. Janet brings a policy lens to tech issues: translating technical research into insights that are salient with policymakers.

Prior to joining CNAS, Janet was a Director in the Australian Government Department of the Prime Minister and Cabinet. She has applied experience working on policy at the intersection of national security, economics and international relations, on issues spanning 5G security, cyber security, countering foreign interference, foreign investment and trade, and critical infrastructure regulations. Janet has a MPP from the Harvard Kennedy School, and a BA from Monash University in Australia.

Focus:
Technical Governance
Policy and Governance, Compute and Hardware
Alex Mallen
Redwood Research
,
Member of Technical Staff

Alex Mallen is a Member of Technical Staff at Redwood Research. He studied CS at the University of Washington, and previously worked at EleutherAI.

Focus:
Empirical
Control, Model Organisms, Scheming and Deception, Strategy and Forecasting
Programs:

Abram Demski is an AI Safety researcher specializing in Agent Foundations, best known for Embedded Agency (co-written with Scott Garrabrant). His overall approach primarily involves deconfusion research in relation to various concepts related to AI risks, including agency, optimization, trust, meaning, understanding, interpretability, and computational uncertainty (more commonly but less precisely known as bounded rationality). More specifically, his recent work focuses on modeling trust, with the objective of clarifying conditions under which humans can justifiably trust AI.

Focus:
Theory
Agent Foundations
Programs:

Alexis is the co-founder and CEO of Asymmetric Security. He was previously an AI security fellow at RAND and part of the founding team of GovAI. 

Focus:
Empirical
Security, Dangerous Capability Evals
Programs:
Paul Riechers
Simplex
,
Simplex research lead

Paul Riechers is a researcher and scientific leader with deep expertise in the physics of information and the fundamental limits of learning and prediction.  He co-founded Simplex, an AI safety research organization, with Dr. Adam Shai, applying insights from theoretical physics and neuroscience to build foundational understanding of internal representations and emergent behavior in neural networks.  Paul earned a PhD in theoretical physics and an MS in electrical and computer engineering from UC Davis. Prior to founding Simplex, he spent five years as a Research Fellow at Nanyang Technological University in Singapore. He is also co-founder of the Beyond Institute for Theoretical Science (BITS), a former Senior Fellow at UCLA’s Mathematics of Intelligences program at IPAM, and has served as both a MATS scholar and mentor. Paul has co-organized multiple workshops on AI interpretability and alignment, and now co-leads the growing Simplex team with support from the Astera Institute.

Focus:
Empirical
Interpretability
James Lucassen
Redwood Research
,
Member of Technical Staff

James Lucassen is a Member of Technical Staff at Redwood Research. He studied CS at Harvey Mudd College, and previously did AI safety research at MIRI and CMU.

Focus:
Empirical
Control, Model Organisms, Scheming and Deception, Strategy and Forecasting
Programs:
Aryan Bhatt
Redwood Research
,
Member of Technical Staff

Aryan Bhatt is a Member of Technical Staff at Redwood Research. He studied Math and CS at Hunter College, and attended MATS in 2023. He studies AI Control.

Focus:
Empirical
Control, Model Organisms, Scheming and Deception, Strategy and Forecasting
Programs:
Megan Kinniment
METR
,
Member of Technical Staff

I am a researcher at METR. 

I think the development of AI is going to be a confusing time for the world. I want to help provide good evidence and methodologies for tracking AI development and risk, so humanity can make sensible decisions.

I've had different roles at different times, including leading task development and our monitoring stream. I like prototyping new kinds of evaluations. I think it's healthy to read transcripts. I'm interested in what capabilities matter for being a competent agent, and why current AI agents fall short. I feel lucky that I get to spend time building an understanding of the models. 

I've previously spent time at the Centre on Long-Term Risk and FHI. Before that I studied physics at university, where I did malaria diagnostics research. 

Focus:
Empirical
Dangerous Capability Evals
Jack Lindsey
Anthropic
,
Member of Technical Staff
Focus:
Empirical
Control, Model Organisms, Red-Teaming, Scheming and Deception
Programs:
Vivek Hebbar
Redwood Research
,
Member of Technical Staff

Vivek Hebbar is a Member of Technical Staff at Redwood Research. He attended Stanford before attending MATS in 2022, and researches AI control.

Focus:
Empirical
Control, Model Organisms, Scheming and Deception, Strategy and Forecasting
Focus:
Technical Governance
Dangerous Capability Evals, Control, Strategy and Forecasting, Policy and Governance, Scalable Oversight, Agent Foundations
Programs:
Richard Ngo
Independent
,
Independent researcher

Richard previously worked on alignment at DeepMind and governance at OpenAI. He's currently an independent researcher focusing on multi-agent intelligence. He's particular interested in understanding how subagents aggregate to form robust larger-scale agents, and how those larger-scale agents change the values of their subagents.

Focus:
Theory
Agent Foundations
Alek Westover
Redwood Research
,
Member of Technical Staff

Alek Westover is a Member of Technical Staff at Redwood Research. He studied CS and Math at MIT, and previously did research in theoretical computer science.

Focus:
Empirical
Control, Model Organisms, Scheming and Deception, Strategy and Forecasting
Programs:
Ryan Greenblatt
Redwood Research
,
Member of Technical Staff
Focus:
Empirical
Control, Model Organisms, Scheming and Deception, Strategy and Forecasting
Programs:

I am an Assistant Professor of Statistics and EECS at UC Berkeley, where I’m also part of BAIR and CLIMB. I am also Founder & CEO of Transluce, a non-profit research lab building open, scalable technology for understanding frontier AI systems.

Focus:
Empirical
Interpretability, Monitoring, Dangerous Capability Evals

Cristian is a Research Fellow at Artificial Intelligence Underwriting Company (AIUC). Insurers have been known to play the role of private regulators (such as in commercial nuclear power); his work broadly focuses on how we might steer the insurance market for AI toward an effective private governance regime. 

He was previously a Winter Fellow at the Centre for the Governance of AI, and independent researcher at the AI Safety Student Team at Harvard. He has an M.A. in Philosophy from the University of British Columbia.

Focus:
Technical Governance
Strategy and Forecasting, Policy and Governance
Programs:

I'm a research scientist at the UK AI Security Institute, working on AI control red teaming and model organisms of misalignment. I was previously a postdoc with Sam Bowman at NYU, did MATS with Owain Evans, and mentored for the MATS, SPAR and Pivotal fellowships. I got my PhD at the University of Edinburgh, supervised by Iain Murray.

Focus:
Empirical
Adam Kaufman
Redwood Research
,
Member of Technical Staff
Focus:
Empirical
Control, Model Organisms, Scheming and Deception, Strategy and Forecasting
Programs:
Stephen McAleer
Anthropic
,
Member of Technical Staff

Stephen McAleer is a Member of Technical Staff at Anthropic, working on the Alignment Science team. He was previously a postdoc at CMU working with Tuomas Sandholm. Stephen received his PhD in computer science from the University of California, Irvine working with Pierre Baldi. During his PhD, he did research scientist internships at Intel Labs and DeepMind. Before that, Stephen received his bachelor's degree in mathematics and economics from Arizona State University in 2017. Projects he is interested in include:

- Anything related to control/monitoring for coding agents

- Scalable oversight for agent alignment

- Scheming evaluations and mitigations

- Adversarial training for robust monitors / reward models

- Reward hacking / deception in agents

Focus:
Empirical
Control, Model Organisms, Red-Teaming, Scheming and Deception

Frequently asked questions

What is the MATS Program?
Who are the MATS Mentors?
What are the key dates of the MATS Program?
Who is eligible to apply?
How does the application and mentor selection process work?