This paper aims to contribute to the futurology of a possible artificial intelligence (AI) breakthrough, by reexamining the Omohundro–Bostrom theory for instrumental vs final AI…
Abstract
Purpose
This paper aims to contribute to the futurology of a possible artificial intelligence (AI) breakthrough, by reexamining the Omohundro–Bostrom theory for instrumental vs final AI goals. Does that theory, along with its predictions for what a superintelligent AI would be motivated to do, hold water?
Design/methodology/approach
The standard tools of systematic reasoning and analytic philosophy are used to probe possible weaknesses of Omohundro–Bostrom theory from four different directions: self-referential contradictions, Tegmark’s physics challenge, moral realism and the messy case of human motivations.
Findings
The two cornerstones of Omohundro–Bostrom theory – the orthogonality thesis and the instrumental convergence thesis – are both open to various criticisms that question their validity and scope. These criticisms are however far from conclusive: while they do suggest that a reasonable amount of caution and epistemic humility is attached to predictions derived from the theory, further work will be needed to clarify its scope and to put it on more rigorous foundations.
Originality/value
The practical value of being able to predict AI goals and motivations under various circumstances cannot be overstated: the future of humanity may depend on it. Currently, the only framework available for making such predictions is Omohundro–Bostrom theory, and the value of the present paper is to demonstrate its tentative nature and the need for further scrutiny.
Details
Keywords
Seth D. Baum, Stuart Armstrong, Timoteus Ekenstedt, Olle Häggström, Robin Hanson, Karin Kuhlemann, Matthijs M. Maas, James D. Miller, Markus Salmela, Anders Sandberg, Kaj Sotala, Phil Torres, Alexey Turchin and Roman V. Yampolskiy
This paper aims to formalize long-term trajectories of human civilization as a scientific and ethical field of study. The long-term trajectory of human civilization can be defined…
Abstract
Purpose
This paper aims to formalize long-term trajectories of human civilization as a scientific and ethical field of study. The long-term trajectory of human civilization can be defined as the path that human civilization takes during the entire future time period in which human civilization could continue to exist.
Design/methodology/approach
This paper focuses on four types of trajectories: status quo trajectories, in which human civilization persists in a state broadly similar to its current state into the distant future; catastrophe trajectories, in which one or more events cause significant harm to human civilization; technological transformation trajectories, in which radical technological breakthroughs put human civilization on a fundamentally different course; and astronomical trajectories, in which human civilization expands beyond its home planet and into the accessible portions of the cosmos.
Findings
Status quo trajectories appear unlikely to persist into the distant future, especially in light of long-term astronomical processes. Several catastrophe, technological transformation and astronomical trajectories appear possible.
Originality/value
Some current actions may be able to affect the long-term trajectory. Whether these actions should be pursued depends on a mix of empirical and ethical factors. For some ethical frameworks, these actions may be especially important to pursue.