[Question] What’s the simplest concrete unsolved problem in AI alignment?

In your preferred area of AI alignment, what is the simplest concrete unsolved problem?

By “simplest”, ideally the problem has been solved when any of the conditions are weakened. However, this isn’t always possible, so a simpler solved version of the problem could also work (e.g., Goldbach’s weak conjecture is known to be true.)

By “concrete”, I mean something where given the statement of the problem and a proposed solution, a neutral third party would be able to consistently determine whether it’s solved or not (e.g., not “explain [some theory] in a good way”).

No comments.