I think almost all of these are things that I’d only think after I’d already noticed confusion, and most are things I’d never say in my head anyway. A little way into the list I thought “Wait, did he just ask ChatGPT for different ways to say “I’m confused”?”.
I expect there are things that pop up in my inner monologue when I’m confused about something, that I wouldn’t notice, and it would be very useful to have a list of such phrases, but your list contains ~none of them.
Edit: Actually the last three are reasonable. Are they human written?
Covid was a big learning experience for me, but I’d like to think about more than one example. Covid is interesting because, compared to my examples of birth control and animal-free meat, it seems like with covid humanity smashed the technical problem out of the park, but still overall failed by my lights because of the political situation.
How likely does it seem that we could get full marks on solving alignment but still fail due to politics? I tend to think of building a properly aligned AGI as a straightforward win condition, but that’s not a very deeply considered view. I guess we could solve it on a whiteboard somewhere but for political reasons it doesn’t get implemented in time?