Question: Why is AI alignment hard?
Answer: There's the "we never figure out how to reliably instill AIs with human friendly goals" filter, which seems pretty challenging, especially with [https://www.youtube.com/watch?v꞊bJLcIBixGj8 inner alignment], solving morality in a way which is possible to code up, interpretability, etc.
There's the "race dynamics mean that even though we know how to build the thing safely the first group to cross the recursive self-improvement line ends up not implementing it safely" which is potentially made worse by the twin issues of "maybe robustly aligned AIs are much harder to build" and "maybe robustly aligned AIs are much less compute efficient".
There's the "we solved the previous problems but writing perfectly reliably code in a whole new domain is hard and there is some fatal bug which we don't find until too late" filter. The paper [https://arxiv.org/abs/1701.04739 The Pursuit of Exploitable Bugs in Machine Learning] explores this.
For a much more in depth analysis, see [https://ai-alignment.com/ai-alignment-landscape-d3773c37ae38 Paul Christiano's AI Alignment Landscape] talk and [https://www.alignmentforum.org/posts/WXvt8bxYnwBYpy9oT/the-main-sources-of-ai-risk The Main Sources of AI Risk?].