MIRI Updates
February 2021 Newsletter
MIRI updates Abram Demski distinguishes different versions of the problem of “pointing at” human values in AI alignment. Evan Hubinger discusses “Risks from Learned Optimization” on the AI X-Risk Research Podcast. Eliezer Yudkowsky comments on AI safety via debate and...
January 2021 Newsletter
MIRI updates MIRI’s Evan Hubinger uses a notion of optimization power to define whether AI systems are compatible with the strategy-stealing assumption. MIRI’s Abram Demski discusses debate approaches to AI safety that don’t rely on factored cognition. Evan argues that...
December 2020 Newsletter
MIRI COO Malo Bourgon reviews our past year and discusses our future plans in 2020 Updates and Strategy. Our biggest update is that we've made less concrete progress than we expected on the new research we described in 2018 Update:...
2020 Updates and Strategy
MIRI’s 2020 has been a year of experimentation and adjustment. In response to the COVID-19 pandemic, we largely moved our operations to more rural areas in March, and shifted to a greater emphasis on remote work. We took the opportunity...
November 2020 Newsletter
MIRI researcher Scott Garrabrant has completed his Cartesian Frames sequence. Scott also covers the first two posts' contents in video form. Other MIRI updates Contrary to my previous announcement, MIRI won’t be running a formal fundraiser this year, though we’ll still...
October 2020 Newsletter
Starting today, Scott Garrabrant has begun posting Cartesian Frames, a sequence introducing a new conceptual framework Scott has found valuable for thinking about agency. In Scott's words: Cartesian Frames are “applying reasoning like Pearl's to objects like game theory's, with...