Chapter

Philosophy and Theory of Artificial Intelligence

Volume 5 of the series Studies in Applied Philosophy, Epistemology and Rational Ethics pp 335-347

Risks and Mitigation Strategies for Oracle AI

  • Stuart ArmstrongAffiliated withFuture of Humanity Institute, University of Oxford

* Final gross prices may vary according to local VAT.

Get Access

Abstract

There is no strong reason to believe human level intelligence represents an upper limit of the capacity of artificial intelligence, should it be realized. This poses serious safety issues, since a superintelligent system would have great power to direct the future according to its possibly flawed goals or motivation systems. Oracle AIs (OAI), confined AIs that can only answer questions, are one particular approach to this problem. However even Oracles are not particularly safe: humans are still vulnerable to traps, social engineering, or simply becoming dependent on the OAI. But OAIs are still strictly safer than general AIs, and there are many extra layers of precautions we can add on top of these. This paper looks at some of them and analyses their strengths and weaknesses.

Keywords

Artificial Intelligence Superintelligence Security Risks Motivational control Capability control