[Question] GPT learning from smarter texts?

(Epistemic status: I do not understand GPT deeply, so this is just a random idea.)

If I understand it correctly, GPT learns from existing texts. Lots of texts.

Would it be possible to make GPT smarter by simply giving it smarter text? Of course, writing tons of smarter text would be a lot of work, but what about annotating the existing text, like “take this more seriously” and “take this less seriously”? (From technical perspective, maybe the GPT should read the text marked as serious five times?) Assuming that the annonation is roughly correct, would this improve the results?

*

If yes, the problem is how to select smarter texts, especially if we want lots of them? But I think some good guesses can be made:

  • High-school textbooks. They should be educational and relatively uncontroversial (settled science). There probably are some reviews of textbooks, so only annotate the ones with good reviews.

  • Some parts of Reddit are probably much better than average, in the sense that smart content gets upvoted. So include the upvoted comments.