Paper: Con­sti­tu­tional AI: Harm­less­ness from AI Feed­back (An­thropic)

LawrenceC16 Dec 2022 22:12 UTC
68 points
11 comments1 min readLW link
(www.anthropic.com)

Vaguely in­ter­ested in Effec­tive Altru­ism? Please Take the Offi­cial 2022 EA Survey

Peter Wildeford16 Dec 2022 21:07 UTC
22 points
4 comments1 min readLW link
(rethinkpriorities.qualtrics.com)

Ab­stract con­cepts and met­al­in­gual defi­ni­tion: Does ChatGPT un­der­stand jus­tice and char­ity?

Bill Benzon16 Dec 2022 21:01 UTC
2 points
0 comments13 min readLW link

Beyond the mo­ment of invention

jasoncrawford16 Dec 2022 20:18 UTC
35 points
0 comments2 min readLW link
(rootsofprogress.org)

[Question] What’s the best time-effi­cient al­ter­na­tive to the Se­quences?

trevor16 Dec 2022 20:17 UTC
6 points
7 comments1 min readLW link

Can we effi­ciently ex­plain model be­hav­iors?

paulfchristiano16 Dec 2022 19:40 UTC
64 points
3 comments9 min readLW link
(ai-alignment.com)

Proper scor­ing rules don’t guaran­tee pre­dict­ing fixed points

16 Dec 2022 18:22 UTC
68 points
8 comments21 min readLW link

A learned agent is not the same as a learn­ing agent

Ben Amitay16 Dec 2022 17:27 UTC
4 points
5 comments4 min readLW link

[Question] Col­lege Selec­tion Ad­vice for Tech­ni­cal Alignment

TempCollegeAsk16 Dec 2022 17:11 UTC
11 points
8 comments1 min readLW link

How im­por­tant are ac­cu­rate AI timelines for the op­ti­mal spend­ing sched­ule on AI risk in­ter­ven­tions?

Tristan Cook16 Dec 2022 16:05 UTC
27 points
2 comments1 min readLW link

In­tro­duc­ing Shrubgrazer

jefftk16 Dec 2022 14:50 UTC
22 points
0 comments2 min readLW link
(www.jefftk.com)

Paper: Trans­form­ers learn in-con­text by gra­di­ent descent

LawrenceC16 Dec 2022 11:10 UTC
28 points
11 comments2 min readLW link
(arxiv.org)

Will Machines Ever Rule the World? MLAISU W50

Esben Kran16 Dec 2022 11:03 UTC
12 points
7 comments4 min readLW link
(newsletter.apartresearch.com)

AI over­hangs de­pend on whether al­gorithms, com­pute and data are sub­sti­tutes or complements

NathanBarnard16 Dec 2022 2:23 UTC
2 points
0 comments3 min readLW link

AI Safety Move­ment Builders should help the com­mu­nity to op­ti­mise three fac­tors: con­trib­u­tors, con­tri­bu­tions and coordination

peterslattery15 Dec 2022 22:50 UTC
4 points
0 comments6 min readLW link

Mask­ing to Avoid Miss­ing Things

jefftk15 Dec 2022 21:00 UTC
17 points
2 comments1 min readLW link
(www.jefftk.com)

Con­sider work­ing more hours and tak­ing more stimulants

Arjun Panickssery15 Dec 2022 20:38 UTC
36 points
11 comments1 min readLW link

We’ve stepped over the thresh­old into the Fourth Arena, but don’t rec­og­nize it

Bill Benzon15 Dec 2022 20:22 UTC
2 points
0 comments7 min readLW link

[Question] How is ARC plan­ning to use ELK?

jacquesthibs15 Dec 2022 20:11 UTC
24 points
5 comments1 min readLW link

How “Dis­cov­er­ing La­tent Knowl­edge in Lan­guage Models Without Su­per­vi­sion” Fits Into a Broader Align­ment Scheme

Collin15 Dec 2022 18:22 UTC
243 points
39 comments16 min readLW link1 review

High-level hopes for AI alignment

HoldenKarnofsky15 Dec 2022 18:00 UTC
58 points
3 comments19 min readLW link
(www.cold-takes.com)

Two Dog­mas of LessWrong

omnizoid15 Dec 2022 17:56 UTC
−6 points
155 comments69 min readLW link

Covid 12/​15/​22: China’s Wave Begins

Zvi15 Dec 2022 16:20 UTC
32 points
7 comments10 min readLW link
(thezvi.wordpress.com)

The next decades might be wild

Marius Hobbhahn15 Dec 2022 16:10 UTC
175 points
42 comments41 min readLW link1 review

Ba­sic build­ing blocks of de­pen­dent type theory

Thomas Kehrenberg15 Dec 2022 14:54 UTC
47 points
8 comments13 min readLW link

AI Ne­o­re­al­ism: a threat model & suc­cess crite­rion for ex­is­ten­tial safety

davidad15 Dec 2022 13:42 UTC
64 points
1 comment3 min readLW link

Who should write the defini­tive post on Ziz?

NicholasKross15 Dec 2022 6:37 UTC
3 points
45 comments3 min readLW link

[Question] Is Paul Chris­ti­ano still as op­ti­mistic about Ap­proval-Directed Agents as he was in 2018?

Chris_Leong14 Dec 2022 23:28 UTC
8 points
0 comments1 min readLW link

«Boundaries», Part 3b: Align­ment prob­lems in terms of bound­aries

Andrew_Critch14 Dec 2022 22:34 UTC
72 points
7 comments13 min readLW link

Align­ing al­ign­ment with performance

Marv K14 Dec 2022 22:19 UTC
2 points
0 comments2 min readLW link

Con­trary to List of Lethal­ity’s point 22, al­ign­ment’s door num­ber 2

False Name14 Dec 2022 22:01 UTC
−2 points
5 comments22 min readLW link

Kol­mogorov Com­plex­ity and Si­mu­la­tion Hypothesis

False Name14 Dec 2022 22:01 UTC
−3 points
0 comments7 min readLW link

[Question] Stan­ley Meyer’s wa­ter fuel cell

mikbp14 Dec 2022 21:19 UTC
2 points
6 comments1 min readLW link

all claw, no world — and other thoughts on the uni­ver­sal distribution

Tamsin Leake14 Dec 2022 18:55 UTC
15 points
0 comments7 min readLW link
(carado.moe)

[Question] Is the AI timeline too short to have chil­dren?

Yoreth14 Dec 2022 18:32 UTC
38 points
20 comments1 min readLW link

Pre­dict­ing GPU performance

14 Dec 2022 16:27 UTC
60 points
26 comments1 min readLW link
(epochai.org)

[In­com­plete] What is Com­pu­ta­tion Any­way?

DragonGod14 Dec 2022 16:17 UTC
16 points
1 comment13 min readLW link
(arxiv.org)

Chair Hang­ing Peg

jefftk14 Dec 2022 15:30 UTC
11 points
0 comments1 min readLW link
(www.jefftk.com)

My AGI safety re­search—2022 re­view, ’23 plans

Steven Byrnes14 Dec 2022 15:15 UTC
51 points
10 comments7 min readLW link

Ex­tract­ing and Eval­u­at­ing Causal Direc­tion in LLMs’ Activations

14 Dec 2022 14:33 UTC
29 points
5 comments11 min readLW link

Key Mostly Out­ward-Fac­ing Facts From the Story of VaccinateCA

Zvi14 Dec 2022 13:30 UTC
61 points
2 comments23 min readLW link
(thezvi.wordpress.com)

Dis­cov­er­ing La­tent Knowl­edge in Lan­guage Models Without Supervision

Xodarap14 Dec 2022 12:32 UTC
45 points
1 comment1 min readLW link
(arxiv.org)

[Question] COVID China Per­sonal Ad­vice (No mRNA vax, pos­si­ble hos­pi­tal over­load, bug-chas­ing edi­tion)

Lao Mein14 Dec 2022 10:31 UTC
20 points
11 comments1 min readLW link

Beyond a bet­ter world

Davidmanheim14 Dec 2022 10:18 UTC
14 points
7 comments4 min readLW link
(progressforum.org)

Proof as mere strong evidence

adamShimi14 Dec 2022 8:56 UTC
28 points
16 comments2 min readLW link
(epistemologicalvigilance.substack.com)

Try­ing to dis­am­biguate differ­ent ques­tions about whether RLHF is “good”

Buck14 Dec 2022 4:03 UTC
106 points
47 comments7 min readLW link1 review

[Question] How can one liter­ally buy time (from x-risk) with money?

Alex_Altair13 Dec 2022 19:24 UTC
24 points
3 comments1 min readLW link

[Question] Best in­tro­duc­tory overviews of AGI safety?

JakubK13 Dec 2022 19:01 UTC
21 points
9 comments2 min readLW link
(forum.effectivealtruism.org)

Ap­pli­ca­tions open for AGI Safety Fun­da­men­tals: Align­ment Course

13 Dec 2022 18:31 UTC
48 points
0 comments2 min readLW link

What Does It Mean to Align AI With Hu­man Values?

Algon13 Dec 2022 16:56 UTC
8 points
3 comments1 min readLW link
(www.quantamagazine.org)