[Question] Is Paul Chris­ti­ano still as op­ti­mistic about Ap­proval-Directed Agents as he was in 2018?

Chris_Leong14 Dec 2022 23:28 UTC
8 points
0 comments1 min readLW link

«Boundaries», Part 3b: Align­ment prob­lems in terms of bound­aries

Andrew_Critch14 Dec 2022 22:34 UTC
72 points
7 comments13 min readLW link

Align­ing al­ign­ment with performance

Marv K14 Dec 2022 22:19 UTC
2 points
0 comments2 min readLW link

Con­trary to List of Lethal­ity’s point 22, al­ign­ment’s door num­ber 2

False Name14 Dec 2022 22:01 UTC
−2 points
5 comments22 min readLW link

Kol­mogorov Com­plex­ity and Si­mu­la­tion Hypothesis

False Name14 Dec 2022 22:01 UTC
−3 points
0 comments7 min readLW link

[Question] Stan­ley Meyer’s wa­ter fuel cell

mikbp14 Dec 2022 21:19 UTC
2 points
6 comments1 min readLW link

all claw, no world — and other thoughts on the uni­ver­sal distribution

Tamsin Leake14 Dec 2022 18:55 UTC
15 points
0 comments7 min readLW link
(carado.moe)

[Question] Is the AI timeline too short to have chil­dren?

Yoreth14 Dec 2022 18:32 UTC
38 points
20 comments1 min readLW link

Pre­dict­ing GPU performance

14 Dec 2022 16:27 UTC
60 points
26 comments1 min readLW link
(epochai.org)

[In­com­plete] What is Com­pu­ta­tion Any­way?

DragonGod14 Dec 2022 16:17 UTC
16 points
1 comment13 min readLW link
(arxiv.org)

Chair Hang­ing Peg

jefftk14 Dec 2022 15:30 UTC
11 points
0 comments1 min readLW link
(www.jefftk.com)

My AGI safety re­search—2022 re­view, ’23 plans

Steven Byrnes14 Dec 2022 15:15 UTC
51 points
10 comments7 min readLW link

Ex­tract­ing and Eval­u­at­ing Causal Direc­tion in LLMs’ Activations

14 Dec 2022 14:33 UTC
29 points
5 comments11 min readLW link

Key Mostly Out­ward-Fac­ing Facts From the Story of VaccinateCA

Zvi14 Dec 2022 13:30 UTC
61 points
2 comments23 min readLW link
(thezvi.wordpress.com)

Dis­cov­er­ing La­tent Knowl­edge in Lan­guage Models Without Supervision

Xodarap14 Dec 2022 12:32 UTC
45 points
1 comment1 min readLW link
(arxiv.org)

[Question] COVID China Per­sonal Ad­vice (No mRNA vax, pos­si­ble hos­pi­tal over­load, bug-chas­ing edi­tion)

Lao Mein14 Dec 2022 10:31 UTC
20 points
11 comments1 min readLW link

Beyond a bet­ter world

Davidmanheim14 Dec 2022 10:18 UTC
14 points
7 comments4 min readLW link
(progressforum.org)

Proof as mere strong evidence

adamShimi14 Dec 2022 8:56 UTC
28 points
16 comments2 min readLW link
(epistemologicalvigilance.substack.com)

Try­ing to dis­am­biguate differ­ent ques­tions about whether RLHF is “good”

Buck14 Dec 2022 4:03 UTC
106 points
47 comments7 min readLW link1 review

[Question] How can one liter­ally buy time (from x-risk) with money?

Alex_Altair13 Dec 2022 19:24 UTC
24 points
3 comments1 min readLW link

[Question] Best in­tro­duc­tory overviews of AGI safety?

JakubK13 Dec 2022 19:01 UTC
21 points
9 comments2 min readLW link
(forum.effectivealtruism.org)

Ap­pli­ca­tions open for AGI Safety Fun­da­men­tals: Align­ment Course

13 Dec 2022 18:31 UTC
48 points
0 comments2 min readLW link

What Does It Mean to Align AI With Hu­man Values?

Algon13 Dec 2022 16:56 UTC
8 points
3 comments1 min readLW link
(www.quantamagazine.org)

It Takes Two Parac­eta­mol?

Eli_13 Dec 2022 16:29 UTC
33 points
10 comments2 min readLW link

[In­terim re­search re­port] Tak­ing fea­tures out of su­per­po­si­tion with sparse autoencoders

13 Dec 2022 15:41 UTC
137 points
22 comments22 min readLW link2 reviews

[Question] Is the ChatGPT-simu­lated Linux vir­tual ma­chine real?

Kenoubi13 Dec 2022 15:41 UTC
18 points
7 comments1 min readLW link

Ex­is­ten­tial AI Safety is NOT sep­a­rate from near-term applications

scasper13 Dec 2022 14:47 UTC
37 points
17 comments3 min readLW link

What is the cor­re­la­tion be­tween up­vot­ing and benefit to read­ers of LW?

banev13 Dec 2022 14:26 UTC
8 points
15 comments1 min readLW link

Limits of Superintelligence

Aleksei Petrenko13 Dec 2022 12:19 UTC
1 point
5 comments1 min readLW link

Bay 2022 Solstice

Raemon13 Dec 2022 8:58 UTC
17 points
0 comments1 min readLW link

Last day to nom­i­nate things for the Re­view. Also, 2019 books still ex­ist.

Raemon13 Dec 2022 8:53 UTC
15 points
0 comments1 min readLW link

AI al­ign­ment is dis­tinct from its near-term applications

paulfchristiano13 Dec 2022 7:10 UTC
254 points
21 comments2 min readLW link
(ai-alignment.com)

Take 10: Fine-tun­ing with RLHF is aes­thet­i­cally un­satis­fy­ing.

Charlie Steiner13 Dec 2022 7:04 UTC
37 points
3 comments2 min readLW link

[Question] Are law­suits against AGI com­pa­nies ex­tend­ing AGI timelines?

SlowingAGI13 Dec 2022 6:00 UTC
1 point
1 comment1 min readLW link

EA & LW Fo­rums Weekly Sum­mary (5th Dec − 11th Dec 22′)

Zoe Williams13 Dec 2022 2:53 UTC
7 points
0 comments1 min readLW link

Align­ment with ar­gu­ment-net­works and as­sess­ment-predictions

Tor Økland Barstad13 Dec 2022 2:17 UTC
10 points
5 comments45 min readLW link

Re­vis­it­ing al­gorith­mic progress

13 Dec 2022 1:39 UTC
94 points
15 comments2 min readLW link1 review
(arxiv.org)

An ex­plo­ra­tion of GPT-2′s em­bed­ding weights

Adam Scherlis13 Dec 2022 0:46 UTC
42 points
4 comments10 min readLW link

12 ca­reer-re­lated ques­tions that may (or may not) be helpful for peo­ple in­ter­ested in al­ign­ment research

Akash12 Dec 2022 22:36 UTC
20 points
0 comments2 min readLW link

Con­cept ex­trap­o­la­tion for hy­poth­e­sis generation

12 Dec 2022 22:09 UTC
20 points
2 comments3 min readLW link

Let’s go meta: Gram­mat­i­cal knowl­edge and self-refer­en­tial sen­tences [ChatGPT]

Bill Benzon12 Dec 2022 21:50 UTC
5 points
0 comments9 min readLW link

D&D.Sci De­cem­ber 2022 Eval­u­a­tion and Ruleset

abstractapplic12 Dec 2022 21:21 UTC
14 points
7 comments2 min readLW link

Log-odds are bet­ter than Probabilities

Robert_AIZI12 Dec 2022 20:10 UTC
22 points
4 comments4 min readLW link
(aizi.substack.com)

Ben­galuru LW/​ACX So­cial Meetup—De­cem­ber 2022

faiz12 Dec 2022 19:30 UTC
4 points
0 comments1 min readLW link

Psy­cholog­i­cal Di­sor­ders and Problems

12 Dec 2022 18:15 UTC
39 points
6 comments1 min readLW link

Con­fus­ing the goal and the path

adamShimi12 Dec 2022 16:42 UTC
44 points
7 comments1 min readLW link
(epistemologicalvigilance.substack.com)

Mean­ingful things are those the uni­verse pos­sesses a se­man­tics for

Abhimanyu Pallavi Sudhir12 Dec 2022 16:03 UTC
16 points
14 comments14 min readLW link

Trade­offs in com­plex­ity, ab­strac­tion, and generality

12 Dec 2022 15:55 UTC
32 points
0 comments2 min readLW link

Green Line Ex­ten­sion Open­ing Dates

jefftk12 Dec 2022 14:40 UTC
12 points
0 comments1 min readLW link
(www.jefftk.com)

Join the AI Test­ing Hackathon this Friday

Esben Kran12 Dec 2022 14:24 UTC
10 points
0 comments1 min readLW link