1f65722061820f92741e5f6c46ae063072bfb4bc46768a3ef6a515fb2d12b585481b0d9330244c3263037f6f1f2a06f43048b117bda3db471569360504154a7d

Question: Why not just put it in a box?

Answer: One possible way to ensure the safety of a powerful AI system is to keep it contained in a software environment. There is nothing intrinsically wrong with this procedure - keeping an AI system in a secure software environment would make it safer than letting it roam free. However, even AI systems inside software environments might not be safe enough.

Humans sometimes put dangerous humans inside boxes to limit their ability to influence the external world. Sometimes, these humans escape their boxes. The security of a prison depends on certain assumptions, which can be violated. [https://medium.com/breakingasia/yoshie-shiratori-the-incredible-story-of-a-man-no-prison-could-hold-6d79a67345f5 Yoshie Shiratori] reportedly escaped prison by weakening the door-frame with miso soup and dislocating his shoulders.

Human written software has a [https://spacepolicyonline.com/news/boeing-software-errors-could-have-doomed-starliners-uncrewed-test-flight/ high defect rate]; we should expect a perfectly secure system to be difficult to create. If humans construct a software system they think is secure, it is possible that the security relies on a false assumption. A powerful AI system could potentially learn how its hardware works and manipulate bits to send radio signals. It could fake a malfunction and attempt social engineering when the engineers look at its code. As the saying goes: in order for someone to do something we had imagined was impossible requires only that they have a better imagination.

Experimentally, humans have [https://yudkowsky.net/singularity/aibox/ convinced] other humans to let them out of the box. Spooky.