Projects in this stream will be on AI welfare and moral status; more specifically, on what it takes to be a moral patient and how we can determine whether AI systems meet the conditions. I'm looking for applicants who have ideas about these topics and are motivated to explore them in more detail.
(i) Eleos AI are launching a program to conduct empirical research on welfare-relevant properties of current AI. I am looking for scholars to conduct experiments as part of this program.
(ii) I am also willing to supervise philosophical research related to AI consciousness and welfare. This could be research in ethics, such as on conditions for moral patienthood, or in philosophy of mind/cognitive science on the nature of welfare-relevant properties and states, such as sentience and desire.
(iii) Other projects concerning AI welfare; if you have an idea for a project in this area, I will be glad to consider it.
I am a philosopher of mind and a researcher at Eleos AI, where I work on AI consciousness, agency and welfare. Before joining Eleos, I worked at the Future of Humanity Institute and Global Priorities Institute in Oxford. I'm interested in projects including purely philosophical work on the grounds of moral status; research drawing on cognitive science to gain a mechanistic understanding of sentience and agency; and empirical studies that can shed light on welfare-relevant features in AI.
By default, scholars will meet with me online for 1hr/week and I will respond to questions on email/slack.
Scholars should have the following characteristics:
Can independently find collaborators, but not required
I will talk through project ideas with scholar