RSS

LawrenceC(Lawrence Chan)

Karma: 4,564

I do AI Alignment research. Currently at ARC Evals, though I still dabble in grantmaking and interpretability in my spare time.

I’m also currently on leave from my PhD at UC Berkeley’s CHAI.

Obligatory research billboard website: https://​​chanlawrence.me/​​

What I would do if I wasn’t at ARC Evals

LawrenceC5 Sep 2023 19:19 UTC
212 points
8 comments13 min readLW link

Nat­u­ral Ab­strac­tions: Key claims, The­o­rems, and Critiques

16 Mar 2023 16:37 UTC
206 points
20 comments45 min readLW link

Causal Scrub­bing: a method for rigor­ously test­ing in­ter­pretabil­ity hy­pothe­ses [Red­wood Re­search]

3 Dec 2022 0:58 UTC
195 points
35 comments20 min readLW link1 review

Sam Alt­man fired from OpenAI

LawrenceC17 Nov 2023 20:42 UTC
192 points
75 comments1 min readLW link
(openai.com)

Shard The­ory in Nine Th­e­ses: a Distil­la­tion and Crit­i­cal Appraisal

LawrenceC19 Dec 2022 22:52 UTC
138 points
30 comments18 min readLW link

Eval­u­a­tions (of new AI Safety re­searchers) can be noisy

LawrenceC5 Feb 2023 4:15 UTC
130 points
10 comments16 min readLW link

An­thropic re­lease Claude 3, claims >GPT-4 Performance

LawrenceC4 Mar 2024 18:23 UTC
114 points
40 comments2 min readLW link
(www.anthropic.com)

Touch re­al­ity as soon as pos­si­ble (when do­ing ma­chine learn­ing re­search)

LawrenceC3 Jan 2023 19:11 UTC
107 points
7 comments8 min readLW link

Sam Alt­man: “Plan­ning for AGI and be­yond”

LawrenceC24 Feb 2023 20:28 UTC
104 points
54 comments6 min readLW link
(openai.com)

Paper: Dis­cov­er­ing novel al­gorithms with AlphaTen­sor [Deep­mind]

LawrenceC5 Oct 2022 16:20 UTC
82 points
18 comments1 min readLW link
(www.deepmind.com)

OpenAI/​Microsoft an­nounce “next gen­er­a­tion lan­guage model” in­te­grated into Bing/​Edge

LawrenceC7 Feb 2023 20:38 UTC
79 points
4 comments1 min readLW link
(blogs.microsoft.com)

Paper: Con­sti­tu­tional AI: Harm­less­ness from AI Feed­back (An­thropic)

LawrenceC16 Dec 2022 22:12 UTC
68 points
11 comments1 min readLW link
(www.anthropic.com)

Paper: The Ca­pac­ity for Mo­ral Self-Cor­rec­tion in Large Lan­guage Models (An­thropic)

LawrenceC16 Feb 2023 19:47 UTC
65 points
9 comments1 min readLW link
(arxiv.org)

Open Phil re­leases RFPs on LLM Bench­marks and Forecasting

LawrenceC11 Nov 2023 3:01 UTC
53 points
0 comments2 min readLW link
(www.openphilanthropy.org)

Paper: Su­per­po­si­tion, Me­moriza­tion, and Dou­ble Des­cent (An­thropic)

LawrenceC5 Jan 2023 17:54 UTC
53 points
11 comments1 min readLW link
(transformer-circuits.pub)