119ae17987827a9d2ab862b63bb4e505a7c951a5ef448d46d5f37fcd23ae76e54a2b752adb3b2f49f1f67b810508e4114da14f05945dcd251fe5bf5e37b3ad83

Question: Why is AGI dangerous?

Answer: # [https://www.youtube.com/watch?v꞊hEUO6pjwFOo The Orthogonality Thesis]: AI could have almost any goal while at the same time having high intelligence (aka ability to succeed at those goals). This means that we could build a very powerful agent which would not necessarily share human-friendly values. For example, the classic [https://www.lesswrong.com/tag/paperclip-maximizer paperclip maximizer] thought experiment explores this with an AI which has a goal of creating as many paperclips as possible, something that humans are (mostly) indifferent to, and as a side effect ends up destroying humanity to make room for more paperclip factories.

[https://www.lesswrong.com/posts/GNnHHmm8EzePmKzPk/value-is-fragile Complexity of value]: What humans care about is not simple, and the space of all goals is large, so virtually all goals we could program into an AI would lead to worlds not valuable to humans if pursued by a sufficiently powerful agent. If we, for example, did not include our value of diversity of experience, we could end up with a world of endlessly looping simple pleasures, rather than beings living rich lives.

[https://www.youtube.com/watch?v꞊ZeecOKBus3Q Instrumental Convergence]: For almost any goal an AI has there are shared ‘instrumental’ steps, such as acquiring resources, preserving itself, and preserving the contents of its goals. This means that a powerful AI with goals that were not explicitly human-friendly would predictably both take actions that lead to the end of humanity (e.g. using resources humans need to live to further its goals, such as replacing our crop fields with vast numbers of solar panels to power its growth, or using the carbon in our bodies to build things) and prevent us from turning it off or altering its goals.