I like it, it is worth a try because it could be very helpful if it works!
A possible objection is that “you can’t mentor others on something you suck yourself,” and this would require AGI capable of making valuable LessWrong comments themselves, which may be similarly hard to automating AI research (considering the math/programming advantages of LLMs).
This objection doesn’t doom your idea, because even if the AI is bad at writing valuable comments, and bad at judging valuable comments written by itself, it may be good at judging the failure modes where a human writes a bad comments. It could still work and is worth a try!
I like it, it is worth a try because it could be very helpful if it works!
A possible objection is that “you can’t mentor others on something you suck yourself,” and this would require AGI capable of making valuable LessWrong comments themselves, which may be similarly hard to automating AI research (considering the math/programming advantages of LLMs).
This objection doesn’t doom your idea, because even if the AI is bad at writing valuable comments, and bad at judging valuable comments written by itself, it may be good at judging the failure modes where a human writes a bad comments. It could still work and is worth a try!