Open Philanthropy recommended a total of $16,306,737 in funding for projects working with deep learning systems that could help us understand and make progress on AI alignment. We sought applications for this funding here.
Recipients include (organized by research direction, with institution name in parentheses):
Interpretability:
- David Bau (Northeastern University)
- Eugene Belilovsky (Concordia University)
- Ruth Fong (Princeton)
- Surya Ganguli (Stanford University)
- Roger Grosse (University of Toronto)
- Yonatan Belinkov (Technion – Israel Institute of Technology)
Measuring and forecasting risks:
- David McAllester (Toyota Technological Institute at Chicago)
- Michael Wellman (University of Michigan)
Techniques for enhancing human feedback:
- Yoav Artzi (Cornell University)
- Samuel Bowman (New York University)
- Greg Durrett (University of Texas at Austin)
- Faramarz Fekri (Georgia Institute of Technology)
- Daniel Kang (University of Illinois)
- Mohit Iyyer (University of Massachusetts, Amherst)
- Victor Veitch (University of Chicago)
Truthful and honest AI:
- David Blei (Columbia University)
- Peter Clark (Allen Institute of AI)
- Dylan Hadfield-Menell (Massachusetts Institute of Technology)
- Tatsunori Hashimo (Stanford University)
- He He (New York University)
- Dan Klein (University of California, Berkeley)
- David Krueger (University of Cambridge)
- Colin Raffel (University of North Carolina, Chapel Hill)
Other:
- Chelsea Finn (Stanford University)
This falls within our focus area of potential risks from advanced artificial intelligence.
The grant amount was updated in August 2024.