[Question] Con­vince me that hu­man­ity is as doomed by AGI as Yud­kowsky et al., seems to believe

Yitz10 Apr 2022 21:02 UTC
92 points
141 comments2 min readLW link

Emo­tion­ally Con­fronting a Prob­a­bly-Doomed World: Against Mo­ti­va­tion Via Dig­nity Points

TurnTrout10 Apr 2022 18:45 UTC
151 points
7 comments9 min readLW link

[Question] Does non-ac­cess to out­puts pre­vent re­cur­sive self-im­prove­ment?

Gunnar_Zarncke10 Apr 2022 18:37 UTC
15 points
0 comments1 min readLW link

A Brief Ex­cur­sion Into Molec­u­lar Neuroscience

Jan10 Apr 2022 17:55 UTC
48 points
8 comments19 min readLW link
(universalprior.substack.com)

Fi­nally En­ter­ing Alignment

Ulisse Mini10 Apr 2022 17:01 UTC
79 points
8 comments2 min readLW link

Schel­ling Meetup Toronto

Sean Aubin10 Apr 2022 13:58 UTC
3 points
0 comments1 min readLW link

Is Fish­e­rian Ru­n­away Gra­di­ent Hack­ing?

Ryan Kidd10 Apr 2022 13:47 UTC
15 points
6 comments4 min readLW link

Worse than an un­al­igned AGI

shminux10 Apr 2022 3:35 UTC
−1 points
11 comments1 min readLW link

Time-Time Tradeoffs

Akash10 Apr 2022 2:33 UTC
17 points
1 comment3 min readLW link
(forum.effectivealtruism.org)

Bos­ton Con­tra: Fully Gen­der-Free

jefftk10 Apr 2022 0:40 UTC
3 points
12 comments1 min readLW link
(www.jefftk.com)

[Question] Hid­den com­ments set­tings not work­ing?

TLW9 Apr 2022 23:15 UTC
4 points
2 comments1 min readLW link

God­shat­ter Ver­sus Leg­i­bil­ity: A Fun­da­men­tally Differ­ent Ap­proach To AI Alignment

LukeOnline9 Apr 2022 21:43 UTC
15 points
14 comments7 min readLW link

A con­crete bet offer to those with short AGI timelines

9 Apr 2022 21:41 UTC
198 points
116 comments5 min readLW link

New: use The Non­lin­ear Library to listen to the top LessWrong posts of all time

KatWoods9 Apr 2022 20:50 UTC
39 points
9 comments8 min readLW link

140 Cog­ni­tive Bi­ases You Should Know

André Ferretti9 Apr 2022 17:15 UTC
7 points
7 comments1 min readLW link

Strate­gies for keep­ing AIs nar­row in the short term

Rossin9 Apr 2022 16:42 UTC
9 points
3 comments3 min readLW link

Hyper­bolic takeoff

Ege Erdil9 Apr 2022 15:57 UTC
17 points
7 comments10 min readLW link
(www.metaculus.com)

Elicit: Lan­guage Models as Re­search Assistants

9 Apr 2022 14:56 UTC
71 points
6 comments13 min readLW link

Emer­gent Ven­tures/​Sch­midt (new grantor for in­di­vi­d­ual re­searchers)

gwern9 Apr 2022 14:41 UTC
21 points
6 comments1 min readLW link
(marginalrevolution.com)

AI safety: the ul­ti­mate trol­ley problem

chaosmage9 Apr 2022 12:05 UTC
−21 points
6 comments1 min readLW link

AMA Con­jec­ture, A New Align­ment Startup

adamShimi9 Apr 2022 9:43 UTC
47 points
42 comments1 min readLW link

[Question] What ad­vice do you have for some­one strug­gling to de­tach their grim-o-me­ter?

Zorger749 Apr 2022 7:35 UTC
6 points
3 comments1 min readLW link

[Question] Can AI sys­tems have ex­tremely im­pres­sive out­puts and also not need to be al­igned be­cause they aren’t gen­eral enough or some­thing?

WilliamKiely9 Apr 2022 6:03 UTC
6 points
3 comments1 min readLW link

Buy-in Be­fore Randomization

jefftk9 Apr 2022 1:30 UTC
26 points
9 comments1 min readLW link
(www.jefftk.com)

Why In­stru­men­tal Goals are not a big AI Safety Problem

Jonathan Paulson9 Apr 2022 0:10 UTC
0 points
7 comments3 min readLW link

A method of writ­ing con­tent eas­ily with lit­tle anxiety

jessicata8 Apr 2022 22:11 UTC
64 points
19 comments3 min readLW link
(unstableontology.com)

Good Heart Dona­tion Lot­tery Winner

Gordon Seidoh Worley8 Apr 2022 20:34 UTC
21 points
0 comments1 min readLW link

Roam Re­search Mo­bile is Out!

Logan Riggs8 Apr 2022 19:05 UTC
12 points
0 comments1 min readLW link

Progress Re­port 4: logit lens redux

Nathan Helm-Burger8 Apr 2022 18:35 UTC
3 points
0 comments2 min readLW link

[Question] What would the cre­ation of al­igned AGI look like for us?

Perhaps8 Apr 2022 18:05 UTC
3 points
4 comments1 min readLW link

Con­vinc­ing All Ca­pa­bil­ity Researchers

Logan Riggs8 Apr 2022 17:40 UTC
120 points
70 comments3 min readLW link

Lan­guage Model Tools for Align­ment Research

Logan Riggs8 Apr 2022 17:32 UTC
28 points
0 comments2 min readLW link

Take­aways From 3 Years Work­ing In Ma­chine Learning

George3d68 Apr 2022 17:14 UTC
34 points
10 comments11 min readLW link
(www.epistem.ink)

[RETRACTED] It’s time for EA lead­er­ship to pull the short-timelines fire alarm.

Not Relevant8 Apr 2022 16:07 UTC
109 points
163 comments4 min readLW link

Boulder ACX Meetup, Sun Apr 24

Josh Sacks8 Apr 2022 15:43 UTC
5 points
4 comments1 min readLW link

AIs should learn hu­man prefer­ences, not biases

Stuart_Armstrong8 Apr 2022 13:45 UTC
10 points
0 comments1 min readLW link

We Are Con­jec­ture, A New Align­ment Re­search Startup

Connor Leahy8 Apr 2022 11:40 UTC
197 points
25 comments4 min readLW link

Differ­ent per­spec­tives on con­cept extrapolation

Stuart_Armstrong8 Apr 2022 10:42 UTC
48 points
8 comments5 min readLW link1 review

[Question] Is there a pos­si­bil­ity that the up­com­ing scal­ing of data in lan­guage mod­els causes A.G.I.?

ArtMi8 Apr 2022 6:56 UTC
2 points
0 comments1 min readLW link

Good Heart Week Is Over!

Ben Pace8 Apr 2022 6:43 UTC
55 points
35 comments1 min readLW link

The Ra­tion­al­ist-Etcetera Di­as­pora: A SPREADSHEET!!

Amelia Bedelia8 Apr 2022 5:43 UTC
25 points
2 comments1 min readLW link

AI Align­ment and Recognition

Chris_Leong8 Apr 2022 5:39 UTC
7 points
2 comments1 min readLW link

Na­ture’s an­swer to the ex­plore/​ex­ploit problem

lizard_brain8 Apr 2022 5:13 UTC
5 points
1 comment1 min readLW link

Edge cases don’t in­val­i­date the rule

Adam Selker8 Apr 2022 4:17 UTC
6 points
5 comments2 min readLW link

Re­v­erse (in­tent) al­ign­ment may al­low for safer Oracles

azsantosk8 Apr 2022 2:48 UTC
4 points
0 comments4 min readLW link

Sum­mary: “In­ter­net Search tips” by Gw­ern Branwen

Pablo Repetto8 Apr 2022 2:02 UTC
12 points
2 comments4 min readLW link
(pabloernesto.github.io)

Maxwell Peter­son’s High­lighted Posts

Maxwell Peterson8 Apr 2022 1:34 UTC
5 points
0 comments1 min readLW link

Foot-Chord­ing Chords

jefftk8 Apr 2022 1:10 UTC
8 points
0 comments1 min readLW link
(www.jefftk.com)

Deep­Mind: The Pod­cast—Ex­cerpts on AGI

WilliamKiely7 Apr 2022 22:09 UTC
99 points
11 comments5 min readLW link

Con­vinc­ing Your Brain That Hu­man­ity is Evil is Easy

Johannes C. Mayer7 Apr 2022 21:39 UTC
13 points
4 comments2 min readLW link