I don’t know yet, but researchers have some preliminary thoughts, which I’m hoping to write about in the future. Also I realized that what I actually meant to say is “expected explicit utility maximizers are the wrong framework to use”, not utility functions—I’ve edited the parent comment to reflect this. CIRL comes to mind as published work that’s moving in a direction away from “expected explicit utility maximizers”, even though it does involve a reward function—it involves a human-robot system that together are optimizing some expected utility, but the robot itself is not maximizing some explicitly represented utility function.
I don’t know yet, but researchers have some preliminary thoughts, which I’m hoping to write about in the future. Also I realized that what I actually meant to say is “expected explicit utility maximizers are the wrong framework to use”, not utility functions—I’ve edited the parent comment to reflect this. CIRL comes to mind as published work that’s moving in a direction away from “expected explicit utility maximizers”, even though it does involve a reward function—it involves a human-robot system that together are optimizing some expected utility, but the robot itself is not maximizing some explicitly represented utility function.