Also, I often see them claim their AI ethics work (train a model not to offend the average Berkeley humanities grad—possibly not useless, I suppose, but not exactly going to save our lightcone) is important alignment work.
Wait, you don’t think this (I mean the training, not the offending) is a safety problem in and of itself? (See also my previous comment about this.)
Wait, you don’t think this (I mean the training, not the offending) is a safety problem in and of itself? (See also my previous comment about this.)