Appreciate the thought here. I don’t see any of this as intrinsically objectionable, but I’m still not really sure what the root issues you’re trying to resolve are. What are examples of cases you’re either worried about the current system specifically failing, or areas where we just little don’t have anything even trying to handle a particular use-case?
In the previous thread, you mentioned “voting weight rewards people who invest time on the site, which may not be the thing you want”, and I’m not sure if this was meant to help with that case or not.
I’m still not really sure what the root issues you’re trying to resolve are. What are examples of cases you’re either worried about the current system specifically failing, or areas where we just little don’t have anything even trying to handle a particular use-case?
Sure, I can list some examples, but first note that while I agree that examples are useful, focusing on them too much is not a good way in general to think about designing systems.
A good design can preempt issues that you would never have predicted could happen; a bad design will screw you up in similarly unpredictable ways. What you want to look out for is designs which reflect the computational structure of the problem you are trying to solve. If you cannot think about it in these terms, I don’t think you’ll be persuaded by any specific list of benefits of the proposed system, such as:
multiple voting is natural, safe, and easy to implement
reduced room for downvote abuse, spamming and nepotism (see example in the post)
it’s possible to change how static karma translates into voting power without disrupting the ecosystem (because the calculations you change affect the marginal voting power, not the total voting power)
users can choose their voting strategy (e.g. many low-impact votes or few high-impact ones) without being incentivized to waste time (in the current system, more clicks = more impact)
moderation power is quantitative, allowing things like partial trust, trial periods and temporary stand-ins without bad feelings afterward (instead of “become moderator” and then “unbecome moderator” we have “I trust you enough to give you 1000 moderation power, no strings attached—let’s see what you do with it”)
each user is incentivized to balance downvotes against upvotes (in the current system, the incentive is to downvote everything you don’t upvote to double your signal—and this would stopped only by some artifical intervention?)
Etc. etc. Does it feel like you could generate more pros (and cons) if you set a 5 minute timer? Ultimately, there is nothing I can say to replace you figuring these out.
I like this idea. Ultimately the primary constraint on almost any feature on LessWrong is UI complexity, and so there is a very strong prior against any specific passing the very high bar to make it into the final UI, but this is a pretty good contender.
I am particularly interested in more ideas for communicating this to the user in a way that makes intuitive sense, and that they can understand with existing systems and analogies they are already familiar with.
The other big constraint on UI design are hedonic-gradients. While often a system can be economically optimal, because of hyperbolic discounting and naive reinforcement learning, you often end up with really bad equilibria if one of the core interactions on your site is not fun to use. This in particular limits the degree to which you can force the user to spend a limited number of resources, since it both strongly increases mental overhead (instead of just asking themselves “yay or nay?” after reading a comment, they now need to reason about their limited budget and compare it to alternative options), and because people hate spending limited resources (which results in me having played through 4 final fantasy games, and only using two health potions in over 200 hours of gameplay, because I really hate giving up limited resources, and I might really need them in the next fight, even though I never, ever will).
Ultimately the primary constraint on almost any feature on LessWrong is UI complexity, and so there is a very strong prior against any specific passing the very high bar to make it into the final UI
On the low end, you can fit the idea entirely inside of the existing UI, as a new fancy way of calculating voting weights under the hood (and allowing multiple clicks on the voting buttons).
Then, in a rough order of less to more shocking to users:
showing the user some indication of how many points their one click is currently worth
showing how many unused “voting points” they still have (privately)
showing a breakdown of recevied feedback into positive and negative votes
some simple configuration that allows to change the default allocation of budget to one click (e.g. how many percent, or pick a fixed value)
And that’s probably all you ever need?
This in particular limits the degree to which you can force the user to spend a limited number of resources, since it both strongly increases mental overhead (instead of just asking themselves “yay or nay?” after reading a comment, they now need to reason about their limited budget and compare it to alternative options)
This should be much less of an issue if the configuration of this is global and has reasonable defaults. Then it’s pretty much reduced to “new fancy way of calculating voting weights”, and the users should be fine with just being roughly aware that if they vote lots lots or don’t post anything on their own, their individual votes will have less weight.
in the current system, the incentive is to downvote everything you don’t upvote to double your signal
For some reason, I don’t do that. The interesting comments I upvote, the few annoying ones I downvote, but half or more comments fall in the “meh” territory where I neither appreciate them, nor mind them, so I simply do not vote on those.
Not voting is not exactly halfway between an upvote and a downvote. Upvote costs you a click, downvote costs you a click, but no-vote costs nothing.
Sure, and that’s probably what almost all users do. But the situation is still perverse: the broken incentives of the system are fighting against your private incentive to not waste effort.
This kind of conflict is especially bad if people have different levels of the internal incentive, but also bad even they don’t, because on the margin it pushes everyone to act slightly against their preferences. (I don’t think this particular case is really so bad, but the more general phenomenon is and that’s what you get if you design systems with poor incentives)