Not sure why you replied in three different places. I will (try to) reply to all of them here.
I did this so that you could easily reply to them separately, since they were separate responses.
I do not consider linking to those Aella and Duncan posts a literature review, nor do I consider them central examples of work on this topic.
I did not link them for that reason. I linked them to ask whether my understanding of the general problem you’re pointing to is correct: “Especially bad consequences relative to other instances of this mistake because the topic relates to people’s relationship with their experience of suffering and potentially unfair dismissals of suffering, which can very easily cause damage to readers or encourage readers to cause damage to others.”
I am not going to do a literature review on your behalf.
Fair. I was simply wondering whether or not you had something to back up your claim that this topic has been covered “quite extensively”.
Your explanation of how you will be careful gave me no confidence; the cases I’m worried about are related to people modeling others as undergoing ‘fake’ suffering, and ignoring their suffering on that basis. This is one of the major nexuses of abuse stumbled into by people interested in cognition. You have to take extreme care not to be misread and wielded in this way, and it just really looks like you have no interest in exercising that care. You’re just not going to anticipate all of the different ways this kind of frame can be damaging to someone and forbid them one by one.
I would like to be clear that I do not intend to claim that Newcomblike suffering is fake in any way. Suffering is a subjective experience. It is equally real whether it comes from physical pain, emotional pain, or an initially false belief that quickly becomes true. Hopefully posting it in a place like Lesswrong will keep it mostly away from the eyes of those who will fail to see this point.
I again ask though, how would a literature review help at all?
I’d look at Buddhist accounts of suffering as a starting point.
This does vibe as possibly relevant.
If you’re going to invite people to sink hundreds of cumulative person hours into reading your thing, you really should actually try to make it good, and part of that is having any familiarity at all with relevant background material.
I’m not sure how to feel about this general attitude towards posting. I think with most things I would rather err on the side of posting something bad; I think a lot of great stuff goes unwritten because people’s standards on themselves are too high (of course, Scott’s law of advice reversal applies here, but I think, given I’ve only posted a handfull of times, I’m on the “doesn’t post enough” end of the spectrum). I try to start all of my posts with a TLDR, so that people who aren’t interested or who think they might be harmed by my post can steer clear. Beyond this, I think it’s the readers’ responsibility to avoid content that will harm them or others.
Humans do substantial work on AI r&d, but we haven’t been very effective at alignment research. (At least, according to the view that says alignment is very hard, which typically also says that basically all of our current “alignment” techniques will not scale at all.)
Yup, this is very possible.