GTP4 capable of limited recursive improving?

Apparently with reflection technique (answer-critique-improve) GTP4 capable of giving much better answers. But that implies it should be capable of doing essentially Alpha Go Zero type of learning! It can’t do complete self play from zero as there is no ground truth for it to learn from, but that basically burns all hopes of having further progress bottlenecked by data. Also, while still severely lacking, it constitutes limited self-improvement capability. Not only we have unclear distinction of GTP4 being an AGI or not, but it also have some slight self improvement capabilities! We really are boiling frog, aren’t we?