• Focus Areas
    • Cause Selection
    • Global Health & Wellbeing
      • Effective Altruism Community Growth (Global Health and Wellbeing)
      • Farm Animal Welfare
      • Global Aid Policy
      • Global Health & Development
      • Scientific Research
      • South Asian Air Quality
    • Longtermism
      • Biosecurity & Pandemic Preparedness
      • Effective Altruism Community Growth (Longtermism)
      • Potential Risks from Advanced AI
    • Other Areas
      • Criminal Justice Reform
      • History of Philanthropy
      • Immigration Policy
      • Land Use Reform
      • Macroeconomic Stabilization Policy
  • Grants
  • Research & Updates
    • Research Reports
    • Blog Posts
    • Notable Lessons
    • In the News
  • About Us
    • Grantmaking Process
    • How to Apply for Funding
    • Team
    • Contact Us
    • Stay Updated
  • We’re hiring!
  • Focus Areas
    • Cause Selection
    • Global Health & Wellbeing
      • Effective Altruism Community Growth (Global Health and Wellbeing)
      • Farm Animal Welfare
      • Global Aid Policy
      • Global Health & Development
      • Scientific Research
      • South Asian Air Quality
    • Longtermism
      • Biosecurity & Pandemic Preparedness
      • Effective Altruism Community Growth (Longtermism)
      • Potential Risks from Advanced AI
    • Other Areas
      • Criminal Justice Reform
      • History of Philanthropy
      • Immigration Policy
      • Land Use Reform
      • Macroeconomic Stabilization Policy
  • Grants
  • Research & Updates
    • Research Reports
    • Blog Posts
    • Notable Lessons
    • In the News
  • About Us
    • Grantmaking Process
    • How to Apply for Funding
    • Team
    • Contact Us
    • Stay Updated
  • We’re hiring!

Funding for AI Alignment Projects Working With Deep Learning Systems

  • Focus Area: Potential Risks from Advanced AI
  • Amount: $14,459,002

  • Award Date: April 2022

Table of Contents

    Open Philanthropy recommended a total of $14,459,002 in funding for projects working with deep learning systems that could help us understand and make progress on AI alignment. We sought applications for this funding here.

    Recipients include (organized by research direction, with institution name in parentheses if applicable):

    Interpretability:

    • David Bau (Northeastern University)
    • Eugene Belilovsky (Concordia University)
    • Ruth Fong (Princeton)
    • Roger Grosse (University of Toronto)
    • Monte MacDiarmind
    • Kevin Wang

    Measuring and forecasting risks:

    • David McAllester (Toyota Technological Institute at Chicago)
    • Michael Wellman (University of Michigan)

    Techniques for enhancing human feedback:

    • Yoav Artzi (Cornell University)
    • Samuel Bowman (New York University)
    • Greg Durrett (University of Texas at Austin)
    • Faramarz Fekri (Georgia Institute of Technology)
    • Mohit Iyyer (University of Massachusetts, Amherst)
    • Gabriel Recchia
    • Victor Veitch (University of Chicago)

    Truthful and honest AI:

    • David Blei (Columbia University)
    • Peter Clark (Allen Institute of AI)
    • Dylan Hadfield-Menell (Massachusetts Institute of Technology)
    • Tatsunori Hashimo (Stanford University)
    • He He (New York University)
    • Dan Klein (University of California, Berkeley)
    • Colin Raffel (University of North Carolina, Chapel Hill)

    Other:

    • Chelsea Finn (Stanford University)
    • James Payor

    This falls within our focus area of potential risks from advanced artificial intelligence.

    Related Items

    • Potential Risks from Advanced AI

      Wilson Center — AI Policy Training Program (2022)

      Open Philanthropy recommended a grant of $2,023,322 over two years to the Wilson Center to support their AI policy training program, which is aimed at staffers for members...

      Read more
    • Potential Risks from Advanced AI

      Catherine Brewer — OxAI Safety Hub

      Open Philanthropy recommended a grant of £10,540 (approximately $11,622 at the time of conversion) to Catherine Brewer to support the OxAI Safety Hub, which is a new Oxford-based...

      Read more
    • Potential Risks from Advanced AI

      Northeastern University — Large Language Model Interpretability Research

      Open Philanthropy recommended a grant of $562,128 over two years to Northeastern University to support Professor David Bau’s research on interpreting large language models. This follows our April 2022 support for David...

      Read more
    Back to Grants Database
    Open Philanthropy
    Open Philanthropy
    • Careers
    • Press Kit
    • Governance
    • Privacy Policy
    • Stay Updated
    Mailing Address
    182 Howard Street #225
    San Francisco, CA 94105
    Email
    [email protected]
    Media Inquiries
    [email protected]
    Anonymous Feedback
    Feedback Form

    © Open Philanthropy 2022 This work is licensed under a Creative Commons Attribution-Noncommercial-Share-alike 4.0 International License. If you'd like to translate this content into another language, please get in touch!