I mean there is ~no prior art here because humanity just invented LLMs last ~tuesday.
Okay j’/k, there may be some. But, I think you’re imagining “the LLM is judging whether the content as good” as opposed to “the LLM is given formulaic rules to evaluate posts for, and it returns ‘yes/no/maybe’ for each of those evaluations.”
The question here is more “is it possible to construct rules that are useful?”
(in the conversation that generated this idea, one person noted “on my youtube channel, it’d be pretty great if I could just identify any comment that mentions someone’s appearance and have it automoderated as ‘off topic’”. If we were trying this on a LessWrong-like community, the rules I might want to try to implement would probably be subtler and I don’t know if LLMs could actually pull them off).
I mean there is ~no prior art here because humanity just invented LLMs last ~tuesday.
Okay j’/k, there may be some. But, I think you’re imagining “the LLM is judging whether the content as good” as opposed to “the LLM is given formulaic rules to evaluate posts for, and it returns ‘yes/no/maybe’ for each of those evaluations.”
The question here is more “is it possible to construct rules that are useful?”
(in the conversation that generated this idea, one person noted “on my youtube channel, it’d be pretty great if I could just identify any comment that mentions someone’s appearance and have it automoderated as ‘off topic’”. If we were trying this on a LessWrong-like community, the rules I might want to try to implement would probably be subtler and I don’t know if LLMs could actually pull them off).