[Question] Con­vince me that hu­man­ity is as doomed by AGI as Yud­kowsky et al., seems to believe

YitzApr 10, 2022, 9:02 PM
92 points
141 comments2 min readLW link

Emo­tion­ally Con­fronting a Prob­a­bly-Doomed World: Against Mo­ti­va­tion Via Dig­nity Points

TurnTroutApr 10, 2022, 6:45 PM
154 points
7 comments9 min readLW link

[Question] Does non-ac­cess to out­puts pre­vent re­cur­sive self-im­prove­ment?

Gunnar_ZarnckeApr 10, 2022, 6:37 PM
15 points
0 comments1 min readLW link

A Brief Ex­cur­sion Into Molec­u­lar Neuroscience

JanApr 10, 2022, 5:55 PM
48 points
8 comments19 min readLW link
(universalprior.substack.com)

Fi­nally En­ter­ing Alignment

Ulisse MiniApr 10, 2022, 5:01 PM
80 points
8 comments2 min readLW link

Schel­ling Meetup Toronto

Sean AubinApr 10, 2022, 1:58 PM
3 points
0 comments1 min readLW link

Is Fish­e­rian Ru­n­away Gra­di­ent Hack­ing?

Ryan KiddApr 10, 2022, 1:47 PM
15 points
6 comments4 min readLW link

Worse than an un­al­igned AGI

ShmiApr 10, 2022, 3:35 AM
−1 points
11 comments1 min readLW link

Time-Time Tradeoffs

Orpheus16Apr 10, 2022, 2:33 AM
18 points
1 comment3 min readLW link
(forum.effectivealtruism.org)

Bos­ton Con­tra: Fully Gen­der-Free

jefftkApr 10, 2022, 12:40 AM
3 points
12 comments1 min readLW link
(www.jefftk.com)

[Question] Hid­den com­ments set­tings not work­ing?

TLWApr 9, 2022, 11:15 PM
4 points
2 comments1 min readLW link

God­shat­ter Ver­sus Leg­i­bil­ity: A Fun­da­men­tally Differ­ent Ap­proach To AI Alignment

LukeOnlineApr 9, 2022, 9:43 PM
15 points
14 comments7 min readLW link

A con­crete bet offer to those with short AGI timelines

Apr 9, 2022, 9:41 PM
199 points
120 comments5 min readLW link

New: use The Non­lin­ear Library to listen to the top LessWrong posts of all time

KatWoodsApr 9, 2022, 8:50 PM
39 points
9 comments8 min readLW link

140 Cog­ni­tive Bi­ases You Should Know

André FerrettiApr 9, 2022, 5:15 PM
8 points
7 comments1 min readLW link

Strate­gies for keep­ing AIs nar­row in the short term

RossinApr 9, 2022, 4:42 PM
9 points
3 comments3 min readLW link

Hyper­bolic takeoff

Ege ErdilApr 9, 2022, 3:57 PM
18 points
7 comments10 min readLW link
(www.metaculus.com)

Elicit: Lan­guage Models as Re­search Assistants

Apr 9, 2022, 2:56 PM
71 points
6 comments13 min readLW link

Emer­gent Ven­tures/​Sch­midt (new grantor for in­di­vi­d­ual re­searchers)

gwernApr 9, 2022, 2:41 PM
21 points
6 comments1 min readLW link
(marginalrevolution.com)

AI safety: the ul­ti­mate trol­ley problem

chaosmageApr 9, 2022, 12:05 PM
−21 points
6 comments1 min readLW link

AMA Con­jec­ture, A New Align­ment Startup

adamShimiApr 9, 2022, 9:43 AM
47 points
42 comments1 min readLW link

[Question] What ad­vice do you have for some­one strug­gling to de­tach their grim-o-me­ter?

Zorger74Apr 9, 2022, 7:35 AM
6 points
3 comments1 min readLW link

[Question] Can AI sys­tems have ex­tremely im­pres­sive out­puts and also not need to be al­igned be­cause they aren’t gen­eral enough or some­thing?

WilliamKielyApr 9, 2022, 6:03 AM
6 points
3 comments1 min readLW link

Buy-in Be­fore Randomization

jefftkApr 9, 2022, 1:30 AM
26 points
9 comments1 min readLW link
(www.jefftk.com)

Why In­stru­men­tal Goals are not a big AI Safety Problem

Jonathan PaulsonApr 9, 2022, 12:10 AM
0 points
7 comments3 min readLW link

A method of writ­ing con­tent eas­ily with lit­tle anxiety

jessicataApr 8, 2022, 10:11 PM
64 points
19 comments3 min readLW link
(unstableontology.com)

Good Heart Dona­tion Lot­tery Winner

Gordon Seidoh WorleyApr 8, 2022, 8:34 PM
21 points
0 comments1 min readLW link

Roam Re­search Mo­bile is Out!

Logan RiggsApr 8, 2022, 7:05 PM
12 points
0 comments1 min readLW link

Progress Re­port 4: logit lens redux

Nathan Helm-BurgerApr 8, 2022, 6:35 PM
4 points
0 comments2 min readLW link

[Question] What would the cre­ation of al­igned AGI look like for us?

PerhapsApr 8, 2022, 6:05 PM
3 points
4 comments1 min readLW link

Con­vinc­ing All Ca­pa­bil­ity Researchers

Logan RiggsApr 8, 2022, 5:40 PM
120 points
70 comments3 min readLW link

Lan­guage Model Tools for Align­ment Research

Logan RiggsApr 8, 2022, 5:32 PM
28 points
0 comments2 min readLW link

Take­aways From 3 Years Work­ing In Ma­chine Learning

George3d6Apr 8, 2022, 5:14 PM
35 points
10 comments11 min readLW link
(www.epistem.ink)

[RETRACTED] It’s time for EA lead­er­ship to pull the short-timelines fire alarm.

Not RelevantApr 8, 2022, 4:07 PM
115 points
166 comments4 min readLW link

Boulder ACX Meetup, Sun Apr 24

Josh SacksApr 8, 2022, 3:43 PM
5 points
4 comments1 min readLW link

AIs should learn hu­man prefer­ences, not biases

Stuart_ArmstrongApr 8, 2022, 1:45 PM
10 points
0 comments1 min readLW link

We Are Con­jec­ture, A New Align­ment Re­search Startup

Connor LeahyApr 8, 2022, 11:40 AM
197 points
25 comments4 min readLW link

Differ­ent per­spec­tives on con­cept extrapolation

Stuart_ArmstrongApr 8, 2022, 10:42 AM
48 points
8 comments5 min readLW link1 review

[Question] Is there a pos­si­bil­ity that the up­com­ing scal­ing of data in lan­guage mod­els causes A.G.I.?

ArtMiApr 8, 2022, 6:56 AM
2 points
0 comments1 min readLW link

Good Heart Week Is Over!

Ben PaceApr 8, 2022, 6:43 AM
55 points
35 comments1 min readLW link

The Ra­tion­al­ist-Etcetera Di­as­pora: A SPREADSHEET!!

Amelia BedeliaApr 8, 2022, 5:43 AM
25 points
2 comments1 min readLW link

AI Align­ment and Recognition

Chris_LeongApr 8, 2022, 5:39 AM
7 points
2 comments1 min readLW link

Na­ture’s an­swer to the ex­plore/​ex­ploit problem

lizard_brainApr 8, 2022, 5:13 AM
5 points
1 comment1 min readLW link

Edge cases don’t in­val­i­date the rule

Adam SelkerApr 8, 2022, 4:17 AM
6 points
5 comments2 min readLW link

Re­v­erse (in­tent) al­ign­ment may al­low for safer Oracles

azsantoskApr 8, 2022, 2:48 AM
4 points
0 comments4 min readLW link

Sum­mary: “In­ter­net Search tips” by Gw­ern Branwen

Pablo RepettoApr 8, 2022, 2:02 AM
12 points
2 comments4 min readLW link
(pabloernesto.github.io)

Maxwell Peter­son’s High­lighted Posts

Maxwell PetersonApr 8, 2022, 1:34 AM
5 points
0 comments1 min readLW link

Foot-Chord­ing Chords

jefftk8 Apr 2022 1:10 UTC
8 points
0 comments1 min readLW link
(www.jefftk.com)

Deep­Mind: The Pod­cast—Ex­cerpts on AGI

WilliamKiely7 Apr 2022 22:09 UTC
99 points
12 comments5 min readLW link

Con­vinc­ing Your Brain That Hu­man­ity is Evil is Easy

Johannes C. Mayer7 Apr 2022 21:39 UTC
14 points
4 comments2 min readLW link