Naive Hy­pothe­ses on AI Alignment

Shoshannah Tekofsky2 Jul 2022 19:03 UTC
98 points
29 comments5 min readLW link

The Tree of Life: Stan­ford AI Align­ment The­ory of Change

Gabriel Mukobi2 Jul 2022 18:36 UTC
24 points
0 comments14 min readLW link

Fol­low along with Columbia EA’s Ad­vanced AI Safety Fel­low­ship!

RohanS2 Jul 2022 17:45 UTC
3 points
0 comments2 min readLW link
(forum.effectivealtruism.org)

Wel­come to Analo­gia! (Chap­ter 7)

Justin Bullock2 Jul 2022 17:04 UTC
5 points
0 comments11 min readLW link

[Question] What about tran­shu­mans and be­yond?

AlignmentMirror2 Jul 2022 13:58 UTC
7 points
6 comments1 min readLW link

Goal-di­rect­ed­ness: tack­ling complexity

Morgan_Rogers2 Jul 2022 13:51 UTC
8 points
0 comments38 min readLW link

Liter­a­ture recom­men­da­tions July 2022

ChristianKl2 Jul 2022 9:14 UTC
17 points
9 comments1 min readLW link

Deon­tolog­i­cal Evil

lsusr2 Jul 2022 6:57 UTC
38 points
4 comments2 min readLW link

Could an AI Align­ment Sand­box be use­ful?

Michael Soareverix2 Jul 2022 5:06 UTC
2 points
1 comment1 min readLW link

Five views of Bayes’ Theorem

Adam Scherlis2 Jul 2022 2:25 UTC
38 points
4 comments1 min readLW link

[Linkpost] Ex­is­ten­tial Risk Anal­y­sis in Em­piri­cal Re­search Papers

Dan H2 Jul 2022 0:09 UTC
40 points
0 comments1 min readLW link
(arxiv.org)

Agenty AGI – How Tempt­ing?

PeterMcCluskey1 Jul 2022 23:40 UTC
22 points
3 comments5 min readLW link
(www.bayesianinvestor.com)

AXRP Epi­sode 16 - Prepar­ing for De­bate AI with Ge­offrey Irving

DanielFilan1 Jul 2022 22:20 UTC
20 points
0 comments37 min readLW link

[Question] Ex­am­ples of prac­ti­cal im­pli­ca­tions of Judea Pearl’s Causal­ity work

ChristianKl1 Jul 2022 20:58 UTC
23 points
6 comments1 min readLW link

Minerva

Algon1 Jul 2022 20:06 UTC
36 points
6 comments2 min readLW link
(ai.googleblog.com)

Disarm­ing status

sano1 Jul 2022 20:00 UTC
−4 points
1 comment6 min readLW link

Paper: Fore­cast­ing world events with neu­ral nets

1 Jul 2022 19:40 UTC
39 points
3 comments4 min readLW link

Refram­ing the AI Risk

Thane Ruthenis1 Jul 2022 18:44 UTC
26 points
7 comments6 min readLW link

Who is this MSRayne per­son any­way?

MSRayne1 Jul 2022 17:32 UTC
32 points
30 comments11 min readLW link

Limer­ence Messes Up Your Ra­tion­al­ity Real Bad, Yo

Raemon1 Jul 2022 16:53 UTC
121 points
41 comments3 min readLW link2 reviews

[Link] On the para­dox of tol­er­ance in re­la­tion to fas­cism and on­line con­tent mod­er­a­tion – Un­sta­ble Ontology

Kenny1 Jul 2022 16:43 UTC
5 points
0 comments1 min readLW link

Trends in GPU price-performance

1 Jul 2022 15:51 UTC
85 points
12 comments1 min readLW link1 review
(epochai.org)

[Question] How to deal with non-schedu­la­ble one-off stim­u­lus-re­sponse-pair-like situ­a­tions when plan­ning/​or­ganis­ing pro­jects?

mikbp1 Jul 2022 15:22 UTC
2 points
3 comments1 min readLW link

What Is The True Name of Mo­du­lar­ity?

1 Jul 2022 14:55 UTC
38 points
10 comments12 min readLW link

Defin­ing Op­ti­miza­tion in a Deeper Way Part 1

J Bostock1 Jul 2022 14:03 UTC
7 points
0 comments2 min readLW link

Safetywashing

Adam Scholl1 Jul 2022 11:56 UTC
255 points
20 comments1 min readLW link2 reviews

[Question] AGI al­ign­ment with what?

AlignmentMirror1 Jul 2022 10:22 UTC
6 points
10 comments1 min readLW link

Open & Wel­come Thread—July 2022

Kaj_Sotala1 Jul 2022 7:47 UTC
20 points
61 comments1 min readLW link

[Question] What is the con­trast to coun­ter­fac­tual rea­son­ing?

Dominic Roser1 Jul 2022 7:39 UTC
5 points
10 comments1 min readLW link

Meio­sis is all you need

Metacelsus1 Jul 2022 7:39 UTC
37 points
3 comments2 min readLW link
(denovo.substack.com)

[Question] How to Nav­i­gate Eval­u­at­ing Poli­ti­cized Re­search?

Davis_Kingsley1 Jul 2022 5:59 UTC
11 points
1 comment1 min readLW link

One is (al­most) nor­mal in base π

Adam Scherlis1 Jul 2022 4:05 UTC
14 points
0 comments1 min readLW link
(adam.scherlis.com)

AI safety uni­ver­sity groups: a promis­ing op­por­tu­nity to re­duce ex­is­ten­tial risk

mic1 Jul 2022 3:59 UTC
14 points
0 comments11 min readLW link

Look­ing back on my al­ign­ment PhD

TurnTrout1 Jul 2022 3:19 UTC
318 points
63 comments11 min readLW link

Selec­tion pro­cesses for subagents

Ryan Kidd30 Jun 2022 23:57 UTC
36 points
2 comments9 min readLW link

[Question] Cry­on­ics-ad­ja­cent question

Flaglandbase30 Jun 2022 23:03 UTC
12 points
3 comments1 min readLW link

Fore­casts are not enough

Ege Erdil30 Jun 2022 22:00 UTC
43 points
5 comments5 min readLW link

Mur­phyjitsu: an In­ner Si­mu­la­tor algorithm

CFAR!Duncan30 Jun 2022 21:50 UTC
62 points
23 comments11 min readLW link2 reviews

GPT-3 Catch­ing Fish in Morse Code

Megan Kinniment30 Jun 2022 21:22 UTC
117 points
27 comments8 min readLW link

Me­tacog­ni­tion in the Rat

Jacob Falkovich30 Jun 2022 20:53 UTC
19 points
0 comments6 min readLW link

On viewquakes

Dalton Mabery30 Jun 2022 20:08 UTC
6 points
0 comments2 min readLW link

The Track Record of Fu­tur­ists Seems … Fine

HoldenKarnofsky30 Jun 2022 19:40 UTC
91 points
25 comments12 min readLW link
(www.cold-takes.com)

Quick sur­vey on AI al­ign­ment resources

frances_lorenz30 Jun 2022 19:09 UTC
14 points
0 comments1 min readLW link

[Linkpost] Solv­ing Quan­ti­ta­tive Rea­son­ing Prob­lems with Lan­guage Models

Yitz30 Jun 2022 18:58 UTC
76 points
15 comments2 min readLW link
(storage.googleapis.com)

Failing to fix a dan­ger­ous intersection

alyssavance30 Jun 2022 18:09 UTC
110 points
17 comments2 min readLW link

Most Func­tions Have Un­de­sir­able Global Extrema

En Kepeig30 Jun 2022 17:10 UTC
8 points
5 comments3 min readLW link

He­donis­tic Iso­topes:

Trozxzr30 Jun 2022 16:49 UTC
1 point
0 comments1 min readLW link

Abadar­ian Trades

David Udell30 Jun 2022 16:41 UTC
16 points
22 comments2 min readLW link

Covid 6/​30/​22: Vac­cine Up­date Update

Zvi30 Jun 2022 14:00 UTC
32 points
6 comments12 min readLW link
(thezvi.wordpress.com)

[Question] How should I talk about op­ti­mal but not sub­game-op­ti­mal play?

JamesFaville30 Jun 2022 13:58 UTC
5 points
1 comment3 min readLW link