Some sort of “coordination takeoff” seems not-impossible to me: set up some sort of platform that’s simultaneously massively profitable/addictive/viral and optimizes for e. g. approximating the ground truth.
Prediction markets were supposed to be that, and some sufficiently clever wrapper on them might yet get there.
Twitter’s community notes are another case study, where good, sufficiently cynical incentive design leads to unsupervised selection of truth-ish statements.
This post has been sitting in my head for years. If scaled up, it might produce a sort of white-box “superpersuasion engine” that could then be tuned for raising the sanity waterline.
Intuitively, I think it’s possible there’s some sort of idea from this reference class that would take off explosively if properly implemented, and then fix our civilization. But I haven’t gone beyond idle thinking regarding it.
Some sort of “coordination takeoff” seems not-impossible to me: set up some sort of platform that’s simultaneously massively profitable/addictive/viral and optimizes for e. g. approximating the ground truth.
Prediction markets were supposed to be that, and some sufficiently clever wrapper on them might yet get there.
Twitter’s community notes are another case study, where good, sufficiently cynical incentive design leads to unsupervised selection of truth-ish statements.
This post has been sitting in my head for years. If scaled up, it might produce a sort of white-box “superpersuasion engine” that could then be tuned for raising the sanity waterline.
Intuitively, I think it’s possible there’s some sort of idea from this reference class that would take off explosively if properly implemented, and then fix our civilization. But I haven’t gone beyond idle thinking regarding it.