The AI containment problem

How to build an AI prison

Elon Musk plans to build his Tesla Bot, Optimus, so that humans “can run away from it and most likely overpower it” should they ever need to. “Hopefully, that doesn’t ever happen, but you never know,” says Musk. But is this really enough to make an AI safe? The problem of keeping AI contained, and only doing the things we want it to, is a deceptively tricky one, writes Roman V. Yampolskiy.

 

With the likely development of superintelligent programs in the near future, many scientists have raised the issue of safety as it relates to such technology. A common theme in Artificial Intellgence (AI) safety research is the possibility of keeping a super-intelligent agent in a sealed hardware so as to prevent it from doing any harm to humankind.

In this essay we will review specific proposals aimed at creating restricted environments for safely interacting with artificial minds. We will evaluate feasibility of presented proposals and suggest a protocol aimed at enhancing safety and security of such methodologies. While it is unlikely that long-term and secure confinement of AI is possible, we are hopeful that the proposed protocol will give researchers a little more time to find a permanent and satisfactory solution for addressing existential risks associated with appearance of super-intelligent machines.

Continue reading

Enjoy unlimited access to the world's leading thinkers.

Start by exploring our subscription options or joining our mailing list today.

Start Free Trial

Already a subscriber? Log in

Join the conversation

Ronan Gibney 21 June 2022

I wonder if we are missing the point about AI. There is a less complex version of self-directed activity by software that we should be concerned about - namely, that something like an out of control,destructive, self-adaptive virus (like an AI version of COVID) will destroy technology or make it unusable and unfixable. In this situation, the 'supervirus' would be able to use its AI capabilities to adapt to any counter measures that we might apply, always keeping ahead of human attempts to repair or inoculate ourselves against it. Society then becomes paralysed by our lack of technology and we reverse our 'progress' and go back to pre-industrial, low tech days. The AI doesn't have to take over to have a profound effect, it just has to interfere sufficiently and prevent elimination. After all, the history of conquest shows that it was smallpox and the black death that killed more people and influenced than anything else, not marauding high tech overlords. We all.developed some immunity to those viruses, but what happens with a slightly intelligent non-biological virus that can evade all attempts to devise immunity? We may be only one step away from a permanent loss of all high tech advances.