RSS

RohanS

Karma: 514

Hi, I’m Rohan! I aim to promote welfare and reduce suffering as much as possible for all sentient beings, which has led me to work on AGI safety research. I am particularly interested in foundation model agents (FMAs): systems like AutoGPT and Operator that equip foundation models with memory, tool use, and other affordances so they can perform multi-step tasks autonomously.

I am the founder of Aether, an independent research lab focused on foundation model agent safety. I’m also a PhD student at the University of Toronto, where I am supervised by Professor Zhijing Jin and continue to run Aether. Previously, I completed an undergrad in CS and Math at Columbia, where I helped run Columbia Effective Altruism and Columbia AI Alignment Club (CAIAC). I have done research internships with AI Safety Hub Labs (now LASR Labs), UC Berkeley’s Center for Human-Compatible AI (CHAI), and the ML Alignment & Theory Scholars (MATS) program.

I love playing tennis, listening to rock and indie pop music, playing social deduction games, reading fantasy books, watching a fairly varied set of TV shows and movies, and playing the saxophone, among other things.

A quick, el­e­gant deriva­tion of Bayes’ Theorem

RohanS23 Jan 2026 1:40 UTC
34 points
5 comments1 min readLW link

Ex­plor­ing Re­in­force­ment Learn­ing Effects on Chain-of-Thought Legibility

6 Jan 2026 3:04 UTC
41 points
3 comments21 min readLW link

Aether is hiring tech­ni­cal AI safety researchers

5 Jan 2026 22:27 UTC
20 points
0 comments2 min readLW link

Hid­den Rea­son­ing in LLMs: A Taxonomy

25 Aug 2025 22:43 UTC
73 points
12 comments12 min readLW link

How we spent our first two weeks as an in­de­pen­dent AI safety re­search group

11 Aug 2025 19:32 UTC
32 points
0 comments10 min readLW link

Ex­tract-and-Eval­u­ate Mon­i­tor­ing Can Sig­nifi­cantly En­hance CoT Mon­i­tor Perfor­mance (Re­search Note)

8 Aug 2025 10:41 UTC
51 points
7 comments10 min readLW link

Effi­ciently De­tect­ing Hid­den Rea­son­ing with a Small Pre­dic­tor Model

13 Jul 2025 16:04 UTC
34 points
3 comments16 min readLW link

Aether July 2025 Update

1 Jul 2025 21:08 UTC
26 points
7 comments3 min readLW link

Ro­hanS’s Shortform

RohanS31 Dec 2024 16:11 UTC
3 points
35 comments1 min readLW link

~80 In­ter­est­ing Ques­tions about Foun­da­tion Model Agent Safety

28 Oct 2024 16:37 UTC
48 points
4 comments15 min readLW link

Trans­form­ers Ex­plained (Again)

RohanS22 Oct 2024 4:06 UTC
4 points
0 comments18 min readLW link

Ap­ply to Aether—In­de­pen­dent LLM Agent Safety Re­search Group

RohanS21 Aug 2024 9:47 UTC
13 points
0 comments7 min readLW link
(forum.effectivealtruism.org)

Notes on “How do we be­come con­fi­dent in the safety of a ma­chine learn­ing sys­tem?”

RohanS26 Oct 2023 3:13 UTC
4 points
0 comments13 min readLW link

Quick Thoughts on Lan­guage Models

RohanS18 Jul 2023 20:38 UTC
6 points
0 comments4 min readLW link

~100 In­ter­est­ing Questions

RohanS30 Mar 2023 13:57 UTC
53 points
18 comments9 min readLW link

A Thor­ough In­tro­duc­tion to Abstraction

RohanS13 Jan 2023 0:30 UTC
9 points
1 comment18 min readLW link

Con­tent and Take­aways from SERI MATS Train­ing Pro­gram with John Wentworth

RohanS24 Dec 2022 4:17 UTC
28 points
3 comments12 min readLW link

Fol­low along with Columbia EA’s Ad­vanced AI Safety Fel­low­ship!

RohanS2 Jul 2022 17:45 UTC
3 points
0 comments2 min readLW link
(forum.effectivealtruism.org)