I think it’s very good that OpenAI and Google DeepMind are pursuing complementary approaches in this sense.
As long as both teams keep publishing their advances, this sounds like a win-win. This way the whole field makes better progress, compared to the situation where everyone is following the same general approach.
to be clear, this post is just my personal opinion, and is not necessarily representative of the beliefs of the openai interpretability team as a whole
I think it’s very good that OpenAI and Google DeepMind are pursuing complementary approaches in this sense.
As long as both teams keep publishing their advances, this sounds like a win-win. This way the whole field makes better progress, compared to the situation where everyone is following the same general approach.
to be clear, this post is just my personal opinion, and is not necessarily representative of the beliefs of the openai interpretability team as a whole
Thanks, sure.
And I am simplifying quite a bit (the whole field is much larger anyway).
Mostly, I mean to say I do hope people diversify and not converge in their approaches to interpretability.