A major thread of my thoughts these days is “can we make AI more philosophically competent relative their own overall capability growth?”. I’m not sure if it’s doable because the things you’d need to be good at philosophy are pretty central capabilities-ish-things. (i.e. ability to reason precisely, notice confusion, convert confusion into useful questions, etc)
Re “can AI advisors help?”
A major thread of my thoughts these days is “can we make AI more philosophically competent relative their own overall capability growth?”. I’m not sure if it’s doable because the things you’d need to be good at philosophy are pretty central capabilities-ish-things. (i.e. ability to reason precisely, notice confusion, convert confusion into useful questions, etc)
Curious if you have any thoughts on that.