RSS

RohanS

Karma: 411

I aim to promote welfare and reduce suffering as much as possible. This has led me to work on AGI safety research. I am particularly interested in foundation model agents (FMAs): systems like AutoGPT and Operator that equip foundation models with memory, tool use, and other affordances so they can perform multi-step tasks autonomously.

Previously, I completed an undergrad in CS and Math at Columbia, where I helped run Columbia Effective Altruism and Columbia AI Alignment Club (CAIAC).

Hid­den Rea­son­ing in LLMs: A Taxonomy

25 Aug 2025 22:43 UTC
62 points
8 comments12 min readLW link

How we spent our first two weeks as an in­de­pen­dent AI safety re­search group

11 Aug 2025 19:32 UTC
28 points
0 comments10 min readLW link

Ex­tract-and-Eval­u­ate Mon­i­tor­ing Can Sig­nifi­cantly En­hance CoT Mon­i­tor Perfor­mance (Re­search Note)

8 Aug 2025 10:41 UTC
51 points
7 comments10 min readLW link

Effi­ciently De­tect­ing Hid­den Rea­son­ing with a Small Pre­dic­tor Model

13 Jul 2025 16:04 UTC
33 points
3 comments16 min readLW link

Aether July 2025 Update

1 Jul 2025 21:08 UTC
23 points
7 comments3 min readLW link

Ro­hanS’s Shortform

RohanS31 Dec 2024 16:11 UTC
3 points
28 comments1 min readLW link