Blog

Author: Rob Bensinger

February 2022 Newsletter

As of yesterday, we've released the final posts in the Late 2021 MIRI Conversations sequence, a collection of (relatively raw and unedited) AI strategy conversations: Ngo's view on alignment difficulty Ngo and Yudkowsky on scientific reasoning and pivotal acts Christiano and Yudkowsky...

January 2022 Newsletter

MIRI updates MIRI's $1.2 million Visible Thoughts Project bounty now has an FAQ, and an example of a successful partial run that you can use to inform your own runs. Scott Alexander reviews the first part of the Yudkowsky/Ngo debate. See also Richard...

December 2021 Newsletter

MIRI is offering $200,000 to build a dataset of AI-dungeon-style writing annotated with the thoughts used in the writing process, and an additional $1,000,000 for scaling that dataset an additional 10x: the Visible Thoughts Project. Additionally, MIRI is in the...

Ngo’s view on alignment difficulty

  This post features a write-up by Richard Ngo on his views, with inline comments.   Color key:   Chat     Google Doc content     Inline comments     13. Follow-ups to the Ngo/Yudkowsky conversation   13.1. Alignment...

Conversation on technology forecasting and gradualism

  This post is a transcript of a multi-day discussion between Paul Christiano, Richard Ngo, Eliezer Yudkowsky, Rob Bensinger, Holden Karnofsky, Rohin Shah, Carl Shulman, Nate Soares, and Jaan Tallinn, following up on the Yudkowsky/Christiano debate in 1, 2, 3,...

More Christiano, Cotra, and Yudkowsky on AI progress

  This post is a transcript of a discussion between Paul Christiano, Ajeya Cotra, and Eliezer Yudkowsky (with some comments from Rob Bensinger, Richard Ngo, and Carl Shulman), continuing from 1, 2, and 3.   Color key:  Chat by Paul...