RogerDearnaley comments on Motivating Alignment of LLM-Powered Agents: Easy for AGI, Hard for ASI?