Hi, I joined a few days ago and I’m looking forward to contributing to this great community.
I’m transitioning back to research from startups. Currently based in London.
I’m particularly interested in mechanistic interpretability, chain-of-thought monitoring, and reasoning model interpretability. I’m excited to engage with the thoughtful discussions here on alignment and to collaborate with others.
What’s your view on sceptic claims about RL on transformer LMs like https://arxiv.org/abs/2504.13837v2 or one that CoT instruction yields better results than <thinking> training?
Hi, I joined a few days ago and I’m looking forward to contributing to this great community.
I’m transitioning back to research from startups. Currently based in London.
I’m particularly interested in mechanistic interpretability, chain-of-thought monitoring, and reasoning model interpretability. I’m excited to engage with the thoughtful discussions here on alignment and to collaborate with others.
What’s your view on sceptic claims about RL on transformer LMs like https://arxiv.org/abs/2504.13837v2 or one that CoT instruction yields better results than <thinking> training?