Classification of global catastrophic risks connected with artificial intelligence
- A. Turchin, D. C. Denkenberger
Summary
Turchin and Denkenberger present a classification of global catastrophic risks posed by AI, expanding on previously identified risks and introducing new ones. They argue that different types of catastrophes dominate at various stages of AI development, highlight the complexity and diversity of AI risks, and stress the need for tailored AI safety approaches at each developmental stage.
Abstract
A classification of the global catastrophic risks of AI is presented, along with a comprehensive list of previously identified risks. This classification allows the identification of several new risks. We show that at each level of AI’s intelligence power, separate types of possible catastrophes dominate. Our classification demonstrates that the field of AI risks is diverse, and includes many scenarios beyond the commonly discussed cases of a paperclip maximizer or robot-caused unemployment. Global catastrophic failure could happen at various levels of AI development, namely, (1) before it starts self-improvement, (2) during its takeoff, when it uses various instruments to escape its initial confinement, or (3) after it successfully takes over the world and starts to implement its goal system, which could be plainly unaligned, or feature-flawed friendliness. AI could also halt at later stages of its development either due to technical glitches or ontological problems. Overall, we identified around several dozen scenarios of AI-driven global catastrophe. The extent of this list illustrates that there is no one simple solution to the problem of AI safety, and that AI safety theory is complex and must be customized for each AI development level.