Home, About, Compare, , Upcoming people Information for David Krueger
Table of contents
List of positions (1 position)
AI safety relation
Future of Humanity Institute Intern
, , , ,  April 2020: Long-Term Future Fund Grant Recipient
Products (0 products)
Organization documents (0 documents)
Documents (1 document)
Scalable agent alignment via reward modeling: a research direction 2018-11-19
Jan Leike, David Krueger, Tom Everitt, Miljan Martic, Vishal Maini, Shane Legg arXiv
Recursive reward modeling, Imitation learning, inverse reinforcement learning, Cooperative inverse reinforcement learning, myopic reinforcement learning, iterated amplification, debate This paper introduces the (recursive) reward modeling agenda, discussing its basic outline, challenges, and ways to overcome those challenges. The paper also discusses alternative agendas and their relation to reward modeling.
Showing at most 20 people who are most similar in terms of which
organizations they have worked at.