I actually wonder whether information asymmetries will even be sustainable in the long run, or whether we’ll eventually approach a world where access to information is relatively equalized (that will all depend on how centralized things become). If things do become fairly equalized, the issue may shift from information asymmetries and how to control them to attention asymmetries. I argued something like that here.
I think you hit on this some by shifting the argument from “what can be known?” to “what should be done with what is known?” But I think the next problem may be more than just good attention-allocation decisions. Rather, it may be more so establishing proper defenses against AI-powered attention manipulation, since that seems a likely next locus of control.
I actually wonder whether information asymmetries will even be sustainable in the long run, or whether we’ll eventually approach a world where access to information is relatively equalized (that will all depend on how centralized things become). If things do become fairly equalized, the issue may shift from information asymmetries and how to control them to attention asymmetries. I argued something like that here.
I think you hit on this some by shifting the argument from “what can be known?” to “what should be done with what is known?” But I think the next problem may be more than just good attention-allocation decisions. Rather, it may be more so establishing proper defenses against AI-powered attention manipulation, since that seems a likely next locus of control.