Reflec­tive Consequentialism

Adam Zerner18 Nov 2022 23:56 UTC
21 points
14 comments4 min readLW link

Value Created vs. Value Extracted

Sable18 Nov 2022 21:34 UTC
8 points
6 comments6 min readLW link
(affablyevil.substack.com)

gen­er­al­ized wireheading

Tamsin Leake18 Nov 2022 20:18 UTC
25 points
7 comments2 min readLW link
(carado.moe)

The Disas­trously Con­fi­dent And Inac­cu­rate AI

Sharat Jacob Jacob18 Nov 2022 19:06 UTC
13 points
0 comments13 min readLW link

How AI Fails Us: A non-tech­ni­cal view of the Align­ment Problem

testingthewaters18 Nov 2022 19:02 UTC
7 points
0 comments2 min readLW link
(ethics.harvard.edu)

[Question] Is there any policy for a fair treat­ment of AIs whose friendli­ness is in doubt?

nahoj18 Nov 2022 19:01 UTC
15 points
10 comments1 min readLW link

SBF, Pas­cal’s Mug­ging, and a Pro­posed Solution

Cole Killian18 Nov 2022 18:39 UTC
−1 points
5 comments5 min readLW link
(colekillian.com)

Distil­la­tion of “How Likely Is De­cep­tive Align­ment?”

NickGabs18 Nov 2022 16:31 UTC
24 points
4 comments10 min readLW link

Con­tra Chords

jefftk18 Nov 2022 16:20 UTC
12 points
1 comment7 min readLW link
(www.jefftk.com)

[Question] Up­dates on scal­ing laws for foun­da­tion mod­els from ′ Tran­scend­ing Scal­ing Laws with 0.1% Ex­tra Com­pute’

Nick_Greig18 Nov 2022 12:46 UTC
15 points
2 comments1 min readLW link

Hal­i­fax, NS – Monthly Ra­tion­al­ist, EA, and ACX Meetup

Ideopunk18 Nov 2022 11:45 UTC
10 points
0 comments1 min readLW link

In­tro­duc­ing The Log­i­cal Foun­da­tion, A Plan to End Poverty With Guaran­teed Income

Michael Simm18 Nov 2022 8:13 UTC
9 points
23 comments1 min readLW link

My Deon­tol­ogy Says Nar­row-Mind­ed­ness is Always Wrong

LVSN18 Nov 2022 6:11 UTC
6 points
2 comments1 min readLW link

AI Ethics != Ai Safety

Dentin18 Nov 2022 3:02 UTC
2 points
0 comments1 min readLW link

Don’t de­sign agents which ex­ploit ad­ver­sar­ial inputs

18 Nov 2022 1:48 UTC
70 points
64 comments12 min readLW link

Eng­ineer­ing Monose­man­tic­ity in Toy Models

18 Nov 2022 1:43 UTC
75 points
7 comments3 min readLW link
(arxiv.org)

AGIs may value in­trin­sic re­wards more than ex­trin­sic ones

catubc17 Nov 2022 21:49 UTC
8 points
6 comments4 min readLW link

LLMs may cap­ture key com­po­nents of hu­man agency

catubc17 Nov 2022 20:14 UTC
27 points
0 comments4 min readLW link

Mastodon Replies as Comments

jefftk17 Nov 2022 20:10 UTC
20 points
0 comments1 min readLW link
(www.jefftk.com)

An­nounc­ing the Progress Forum

jasoncrawford17 Nov 2022 19:26 UTC
83 points
9 comments1 min readLW link

[Question] What kind of bias is this?

Daniel Samuel17 Nov 2022 18:44 UTC
3 points
2 comments1 min readLW link

AI Fore­cast­ing Re­search Ideas

Jsevillamol17 Nov 2022 17:37 UTC
21 points
2 comments1 min readLW link

Re­sults from the in­ter­pretabil­ity hackathon

17 Nov 2022 14:51 UTC
81 points
0 comments6 min readLW link
(alignmentjam.com)

Covid 11/​17/​22: Slow Recovery

Zvi17 Nov 2022 14:50 UTC
33 points
3 comments4 min readLW link
(thezvi.wordpress.com)

Sadly, FTX

Zvi17 Nov 2022 14:30 UTC
133 points
18 comments47 min readLW link
(thezvi.wordpress.com)

Deon­tol­ogy and virtue ethics as “effec­tive the­o­ries” of con­se­quen­tial­ist ethics

Jan_Kulveit17 Nov 2022 14:11 UTC
63 points
9 comments1 min readLW link1 review

The Ground Truth Prob­lem (Or, Why Eval­u­at­ing In­ter­pretabil­ity Meth­ods Is Hard)

Jessica Rumbelow17 Nov 2022 11:06 UTC
27 points
2 comments2 min readLW link

[Question] [Per­sonal Ques­tion] Can any­one help me nav­i­gate this po­ten­tially painful in­ter­per­sonal dy­namic ra­tio­nally?

SlainLadyMondegreen17 Nov 2022 8:53 UTC
9 points
3 comments4 min readLW link

Mas­sive Scal­ing Should be Frowned Upon

harsimony17 Nov 2022 8:43 UTC
4 points
6 comments5 min readLW link

[Question] Why are prof­itable com­pa­nies lay­ing off staff?

Yair Halberstadt17 Nov 2022 6:19 UTC
15 points
10 comments1 min readLW link

Dis­cus­sion: Was SBF a naive util­i­tar­ian, or a so­ciopath?

NicholasKross17 Nov 2022 2:52 UTC
0 points
4 comments1 min readLW link

Kel­sey Piper’s re­cent in­ter­view of SBF

agucova16 Nov 2022 20:30 UTC
51 points
29 comments1 min readLW link

The Echo Principle

Jonathan Moregård16 Nov 2022 20:09 UTC
4 points
0 comments3 min readLW link
(honestliving.substack.com)

[Question] Is there some rea­son LLMs haven’t seen broader use?

tailcalled16 Nov 2022 20:04 UTC
25 points
27 comments1 min readLW link

When should we be sur­prised that an in­ven­tion took “so long”?

jasoncrawford16 Nov 2022 20:04 UTC
32 points
11 comments4 min readLW link
(rootsofprogress.org)

Ques­tions about Value Lock-in, Pa­ter­nal­ism, and Empowerment

Sam F. Brown16 Nov 2022 15:33 UTC
13 points
2 comments12 min readLW link
(sambrown.eu)

If Pro­fes­sional In­vestors Missed This...

jefftk16 Nov 2022 15:00 UTC
37 points
18 comments3 min readLW link
(www.jefftk.com)

Disagree­ment with bio an­chors that lead to shorter timelines

Marius Hobbhahn16 Nov 2022 14:40 UTC
75 points
17 comments7 min readLW link1 review

Cur­rent themes in mechanis­tic in­ter­pretabil­ity research

16 Nov 2022 14:14 UTC
89 points
2 comments12 min readLW link

Un­pack­ing “Shard The­ory” as Hunch, Ques­tion, The­ory, and Insight

Jacy Reese Anthis16 Nov 2022 13:54 UTC
31 points
9 comments2 min readLW link

Mir­a­cles and why not to be­lieve them

mruwnik16 Nov 2022 12:07 UTC
4 points
0 comments2 min readLW link

[Question] How do peo­ple do re­mote re­search col­lab­o­ra­tions effec­tively?

Krieger16 Nov 2022 11:51 UTC
8 points
0 comments1 min readLW link

Method of state­ments: an al­ter­na­tive to taboo

Q Home16 Nov 2022 10:57 UTC
7 points
0 comments41 min readLW link

The two con­cep­tions of Ac­tive In­fer­ence: an in­tel­li­gence ar­chi­tec­ture and a the­ory of agency

Roman Leventov16 Nov 2022 9:30 UTC
15 points
0 comments4 min readLW link

Devel­oper ex­pe­rience for the motivation

Adam Zerner16 Nov 2022 7:12 UTC
49 points
7 comments4 min readLW link

Progress links and tweets, 2022-11-15

jasoncrawford16 Nov 2022 3:21 UTC
9 points
0 comments2 min readLW link
(rootsofprogress.org)

EA & LW Fo­rums Weekly Sum­mary (7th Nov − 13th Nov 22′)

Zoe Williams16 Nov 2022 3:04 UTC
19 points
0 comments1 min readLW link

The FTX Saga—Simplified

Annapurna16 Nov 2022 2:42 UTC
44 points
10 comments7 min readLW link
(jorgevelez.substack.com)

Utili­tar­i­anism and the idea of a “ra­tio­nal agent” are fun­da­men­tally in­con­sis­tent with reality

banev16 Nov 2022 0:19 UTC
−4 points
1 comment1 min readLW link

[Question] Is the speed of train­ing large mod­els go­ing to in­crease sig­nifi­cantly in the near fu­ture due to Cere­bras An­dromeda?

Amal 15 Nov 2022 22:50 UTC
12 points
11 comments1 min readLW link