This is a joint post by MIRI Research Associate and DeepMind Research Scientist Ramana Kumar and MIRI Research Fellow Scott Garrabrant, cross-posted from the AI Alignment Forum and LessWrong. Human values and preferences are hard to specify, especially in complex...
This is a guest post by Oliver Habryka, lead developer for LessWrong. Our gratitude to the LessWrong team for the hard work they’ve put into developing this resource, and our congratulations on today’s launch! I am happy to announce that...
This is a guest post by Jed McCaleb, one of MIRI’s top contributors, for our winter fundraiser. A few months ago, several leaders in the scientific community signed an open letter pushing for oversight into the research and development...