Information for David Krueger

Table of contents

Basic information

Item Value
Facebook username 13963931
Intelligent Agent Foundations Forum username 111
Donations List Website (data still preliminary)
Agendas Recursive reward modeling

List of positions (3 positions)

Organization Title Start date End date AI safety relation Subject Employment type Source Notes
Centre for the Study of Existential Risk Research Affiliate position [1]
Future of Humanity Institute Intern 2016-07-01 2016-09-01 [2], [3], [4], [5], [6] April 2020: Long-Term Future Fund Grant Recipient
Center for Human-Compatible AI Affiliate 2022-10-13 [7], [8]

Products (0 products)

Name Creation date Description

Organization documents (0 documents)

Title Publication date Author Publisher Affected organizations Affected people Document scope Cause area Notes

Documents (1 document)

Title Publication date Author Publisher Affected organizations Affected people Affected agendas Notes
Scalable agent alignment via reward modeling: a research direction 2018-11-19 Jan Leike, David Krueger, Tom Everitt, Miljan Martic, Vishal Maini, Shane Legg arXiv Google DeepMind Recursive reward modeling, Imitation learning, inverse reinforcement learning, Cooperative inverse reinforcement learning, myopic reinforcement learning, iterated amplification, debate This paper introduces the (recursive) reward modeling agenda, discussing its basic outline, challenges, and ways to overcome those challenges. The paper also discusses alternative agendas and their relation to reward modeling.

Similar people

Showing at most 20 people who are most similar in terms of which organizations they have worked at.

Person Number of organizations in common List of organizations in common
Allan Dafoe 2 Centre for the Study of Existential Risk, Future of Humanity Institute
Stuart Russell 2 Centre for the Study of Existential Risk, Center for Human-Compatible AI
Nick Bostrom 2 Centre for the Study of Existential Risk, Future of Humanity Institute
Seán Ó hÉigeartaigh 2 Centre for the Study of Existential Risk, Future of Humanity Institute
Carla Zoe Cremer 2 Centre for the Study of Existential Risk, Future of Humanity Institute
Sam Clarke 2 Centre for the Study of Existential Risk, Future of Humanity Institute
Piers Millett 2 Centre for the Study of Existential Risk, Future of Humanity Institute
Beth Barnes 2 Centre for the Study of Existential Risk, Center for Human-Compatible AI
Ryan Carey 2 Centre for the Study of Existential Risk, Future of Humanity Institute
Alison Gopnik 2 Centre for the Study of Existential Risk, Center for Human-Compatible AI
Michael Cohen 2 Center for Human-Compatible AI, Future of Humanity Institute
Christopher Cundy 2 Center for Human-Compatible AI, Future of Humanity Institute
Daniel Filan 2 Center for Human-Compatible AI, Future of Humanity Institute
Sören Mindermann 2 Center for Human-Compatible AI, Future of Humanity Institute
Simon Beard 1 Centre for the Study of Existential Risk
Seth Baum 1 Centre for the Study of Existential Risk
Adrian Weller 1 Centre for the Study of Existential Risk
Yang Liu 1 Centre for the Study of Existential Risk
Huw Price 1 Centre for the Study of Existential Risk
José Hernández-Orallo 1 Centre for the Study of Existential Risk