It actually wouldn’t surprise me if it could be done by a human alignment theorist working with an existing GPT, where the GPT serves mostly as a source of ideas.
It actually wouldn’t surprise me if it could be done by a human alignment theorist working with an existing GPT, where the GPT serves mostly as a source of ideas.