Liron gives us an intuition about how hard the AI alignment problem is at its core, drawing a contrast between the AI’s recursive self-improvement to build up its capabilities to be way super-human level and the thing we’re asking it to do at the same time, which is to be biased in the way that humans are biased and maintain the fragile human values.