Technical work: Making safeguards 'run deep', including safeguards and risk management for open-weight models.
Governance work: Critical review of industry self-governance, critical review of national AI governance institutes, open-weight model governance, predicting and mitigating future AI incidents.
Stephen (“Cas”) Casper is a final year Ph.D student at MIT in the Algorithmic Alignment Group advised by Dylan Hadfield-Menell. His work focuses on AI safeguards and technical governance. His research has been featured at NeurIPS, AAAI, Nature, FAccT, EMNLP, SaTML, TMLR, IRAIS, several course curricula, a number of workshops, and over 20 news articles and newsletters. He is also a writer for the International AI Safety Report and the Singapore Consensus. In addition to MATS, he also mentors for ERA and GovAI. In the past, he has worked closely with over 30 mentees on various safety-related research projects.
2-3 meetings per week plus regular messaging and collaborative writing.
Here are some examples of papers related to safeguards and technical AI governance. If you are interested in any of them, you might be interested in this stream:
Green flags include:
This stream will follow an academic collaboration model. Scholars will be free to discuss and collaborate externally. However, scholars should also expect to work in collaboration with others in the stream.
Mentor(s) will talk through project ideas with scholar.