In Defense of At­tempt­ing Hard Things, and my story of the Lev­er­age ecosystem

Cathleen17 Dec 2021 23:08 UTC
115 points
43 comments1 min readLW link2 reviews
(cathleensdiscoveries.com)

[Question] Get­ting di­ag­nosed for ADHD if I don’t plan on tak­ing meds?

vroomerify17 Dec 2021 19:27 UTC
6 points
6 comments1 min readLW link

Ven­ture Gran­ters, The VCs of pub­lic goods, in­cen­tiviz­ing good dreams

mako yass17 Dec 2021 8:57 UTC
12 points
9 comments12 min readLW link

Un­der­stand the ex­po­nen­tial func­tion: R0 of the COVID

Yandong Zhang17 Dec 2021 6:44 UTC
−6 points
17 comments1 min readLW link

Some mo­ti­va­tions to gra­di­ent hack

peterbarnett17 Dec 2021 3:06 UTC
8 points
0 comments6 min readLW link

Blog Respectably

lsusr17 Dec 2021 1:23 UTC
13 points
4 comments1 min readLW link

The Case for Rad­i­cal Op­ti­mism about Interpretability

Quintin Pope16 Dec 2021 23:38 UTC
66 points
16 comments8 min readLW link1 review

-

Alice K16 Dec 2021 23:03 UTC
2 points
2 comments1 min readLW link

Ev­i­dence Sets: Towards In­duc­tive-Bi­ases based Anal­y­sis of Pro­saic AGI

bayesian_kitten16 Dec 2021 22:41 UTC
22 points
10 comments21 min readLW link

Hous­ing Mar­kets, Satis­ficers, and One-Track Goodhart

J Bostock16 Dec 2021 21:38 UTC
2 points
2 comments2 min readLW link

Covid 12/​16: On Your Marks

Zvi16 Dec 2021 21:00 UTC
53 points
36 comments9 min readLW link
(thezvi.wordpress.com)

Re­views of “Is power-seek­ing AI an ex­is­ten­tial risk?”

Joe Carlsmith16 Dec 2021 20:48 UTC
79 points
20 comments1 min readLW link

The “Other” Option

jsteinhardt16 Dec 2021 20:20 UTC
24 points
1 comment7 min readLW link
(bounded-regret.ghost.io)

What Ca­plan’s “Miss­ing Mood” Heuris­tic Is Really For

DirectedEvolution16 Dec 2021 19:47 UTC
32 points
7 comments4 min readLW link

Sub­way Slides

jefftk16 Dec 2021 19:30 UTC
11 points
2 comments1 min readLW link
(www.jefftk.com)

Viru­lence Management

harsimony16 Dec 2021 19:25 UTC
4 points
0 comments3 min readLW link
(harsimony.wordpress.com)

Omicron Post #7

Zvi16 Dec 2021 17:30 UTC
155 points
41 comments12 min readLW link
(thezvi.wordpress.com)

[Question] Where can one learn deep in­tu­itions about in­for­ma­tion the­ory?

Valentine16 Dec 2021 15:47 UTC
67 points
27 comments2 min readLW link

Elic­i­ta­tion for Model­ing Trans­for­ma­tive AI Risks

Davidmanheim16 Dec 2021 15:24 UTC
30 points
2 comments9 min readLW link

An Open Let­ter to the Monas­tic Academy and com­mu­nity members

HS202116 Dec 2021 9:04 UTC
44 points
46 comments1 min readLW link

Five Miss­ing Moods

mike_hawke16 Dec 2021 1:25 UTC
14 points
3 comments3 min readLW link

Mo­ti­va­tions, Nat­u­ral Selec­tion, and Cur­ricu­lum Engineering

Oliver Sourbut16 Dec 2021 1:07 UTC
16 points
0 comments42 min readLW link

Univer­sal­ity and the “Filter”

maggiehayes16 Dec 2021 0:47 UTC
10 points
2 comments11 min readLW link

More power to you

jasoncrawford15 Dec 2021 23:50 UTC
16 points
14 comments1 min readLW link
(rootsofprogress.org)

My Overview of the AI Align­ment Land­scape: A Bird’s Eye View

Neel Nanda15 Dec 2021 23:44 UTC
127 points
9 comments15 min readLW link

SmartPoop 1.0: An AI Safety Science-Fiction

Lê Nguyên Hoang15 Dec 2021 22:28 UTC
7 points
1 comment1 min readLW link

Bay Area Ra­tion­al­ist Field Day

Raj Thimmiah15 Dec 2021 19:57 UTC
7 points
1 comment1 min readLW link

Fram­ing ap­proaches to al­ign­ment and the hard prob­lem of AI cognition

ryan_greenblatt15 Dec 2021 19:06 UTC
16 points
15 comments27 min readLW link

South Bay ACX/​LW Pre-Holi­day Get-Together

IS15 Dec 2021 16:58 UTC
5 points
0 comments1 min readLW link

Leverage

lsusr15 Dec 2021 5:20 UTC
23 points
2 comments1 min readLW link

We’ll Always Have Crazy

[DEACTIVATED] Duncan Sabien15 Dec 2021 2:55 UTC
36 points
22 comments13 min readLW link

2020 Re­view: The Dis­cus­sion Phase

Vaniver15 Dec 2021 1:12 UTC
55 points
14 comments2 min readLW link

The Nat­u­ral Ab­strac­tion Hy­poth­e­sis: Im­pli­ca­tions and Evidence

CallumMcDougall14 Dec 2021 23:14 UTC
37 points
8 comments19 min readLW link

Robin Han­son’s “Hu­mans are Early”

Raemon14 Dec 2021 22:07 UTC
11 points
0 comments2 min readLW link
(www.overcomingbias.com)

Ngo’s view on al­ign­ment difficulty

14 Dec 2021 21:34 UTC
63 points
7 comments17 min readLW link

A pro­posed sys­tem for ideas jumpstart

Just Learning14 Dec 2021 21:01 UTC
4 points
2 comments3 min readLW link

Should we rely on the speed prior for safety?

Marc Carauleanu14 Dec 2021 20:45 UTC
14 points
5 comments5 min readLW link

ARC’s first tech­ni­cal re­port: Elic­it­ing La­tent Knowledge

14 Dec 2021 20:09 UTC
225 points
90 comments1 min readLW link3 reviews
(docs.google.com)

ARC is hiring!

14 Dec 2021 20:09 UTC
63 points
2 comments1 min readLW link

In­ter­lude: Agents as Automobiles

Daniel Kokotajlo14 Dec 2021 18:49 UTC
26 points
6 comments5 min readLW link

Zvi’s Thoughts on the Sur­vival and Flour­ish­ing Fund (SFF)

Zvi14 Dec 2021 14:30 UTC
186 points
65 comments64 min readLW link1 review
(thezvi.wordpress.com)

Con­se­quen­tial­ism & corrigibility

Steven Byrnes14 Dec 2021 13:23 UTC
66 points
27 comments7 min readLW link

De­ci­sion The­ory Break­down—Per­sonal At­tempt at a Review

Jake Arft-Guatelli14 Dec 2021 0:40 UTC
4 points
1 comment8 min readLW link

Mys­tery Hunt 2022

Scott Garrabrant13 Dec 2021 21:57 UTC
30 points
5 comments1 min readLW link

En­abling More Feed­back for AI Safety Researchers

frances_lorenz13 Dec 2021 20:10 UTC
17 points
0 comments3 min readLW link

Lan­guage Model Align­ment Re­search Internships

Ethan Perez13 Dec 2021 19:53 UTC
74 points
1 comment1 min readLW link

Omicron Post #6

Zvi13 Dec 2021 18:00 UTC
89 points
30 comments8 min readLW link
(thezvi.wordpress.com)

Anal­y­sis of Bird Box (2018)

TekhneMakre13 Dec 2021 17:30 UTC
11 points
3 comments5 min readLW link

Solv­ing In­ter­pretabil­ity Week

Logan Riggs13 Dec 2021 17:09 UTC
11 points
5 comments1 min readLW link

Un­der­stand­ing and con­trol­ling auto-in­duced dis­tri­bu­tional shift

L Rudolf L13 Dec 2021 14:59 UTC
32 points
4 comments16 min readLW link