Home , About , Compare , , Upcoming people
AI Watch
Welcome! This is a website to track people and organizations in the AI safety/alignment/AI existential risk communities.
A position or organization being on AI Watch does not indicate an assessment that that position or organization
is actually making AI safer or that the position or organization is good for the world in any way. It is mostly
a sociological indication that the position or organization is associated with these communities, as well as an
indication that the position or organization claims
to be working on AI safety or alignment . (There are some
plans to eventually introduce such assessments on AI Watch, but for now there are none.)
See the code repository
for the source code and data of this website.
This website is developed by Issa Rice
with data contributions from Sebastian Sanchez ,
Amana Rice , and
Vipul Naik ,
and has been partially
funded by Vipul Naik
and Mati Roy (who in July 2023 paid for the time Issa had spent answering people’s questions about AI Watch up until that point).
Last updated on 2024-11-02; see here for a full list of recent changes.
Table of contents
Agendas
Agenda name
Associated people
Associated organizations
Iterated amplification
Paul Christiano , Buck Shlegeris , Dario Amodei
OpenAI
Embedded agency
Eliezer Yudkowsky , Scott Garrabrant , Abram Demski
Machine Intelligence Research Institute
Comprehensive AI services
Eric Drexler
Future of Humanity Institute
Ambitious value learning
Stuart Armstrong
Future of Humanity Institute
Factored cognition
Andreas Stuhlmüller
Ought
Recursive reward modeling
Jan Leike , David Krueger , Tom Everitt , Miljan Martic , Vishal Maini , Shane Legg
Google DeepMind
Debate
Paul Christiano
OpenAI
Interpretability
Christopher Olah
Inverse reinforcement learning
Preference learning
Cooperative inverse reinforcement learning
Imitation learning
Alignment for advanced machine learning systems
Jessica Taylor , Eliezer Yudkowsky , Patrick LaVictoire , Andrew Critch
Machine Intelligence Research Institute
Learning-theoretic AI alignment
Vanessa Kosoy
Counterfactual reasoning
Jacob Steinhardt
Positions grouped by person
Showing 0 people with positions.
Name
Number of organizations
List of organizations
Positions grouped by organization
Showing 6 organizations.
Organization
Number of people
List of people
GoodAI
32
Karolína H. , Ryan Camilleri , Jose Solorzano , Alex Angelini , Dominik Čech , Sarka Krejcova , Stephanie Wendler , Reham Bukhari , Šimon Šicko , Lucia Šicková , Nicholas Guttenberg , Viktorie Knezkova , Steffen Eichler , Isabeau Premont-Schwarz , Filip Hauptfleisch , Petr Sramek , Jan Štafa , Christine Lee , Michal Dvořák , Will Millership , Lucie Krestova , Marek Havrda , Jan Feyereisl , Olga Afanasjeva , Martin Poliak , Marek Rosa , Simon Andersson , Přemek Paška , Jaroslav Vitku , Joseph Davidson , Shantesh Patil , Petr Hlubuček
Machine Intelligence Research Institute
6
Carson Jones , Kurt Brown , Aaron Silverbook , Jesse Galef , Elizabeth Morningstar , Erica Edelman
Berkeley Existential Risk Initiative
1
Colleen Gleason
EthicsNet
1
Aleksandra Orchowska
Foundational Research Institute
1
Max Daniel
Global Catastrophic Risk Institute
1
Robert de Neufville