What’s next for in­stru­men­tal ra­tio­nal­ity?

Andrew_Critch23 Jul 2022 22:55 UTC
63 points
7 comments1 min readLW link

Easy guide for run­ning a lo­cal Ra­tion­al­ity meetup

Nikita Sokolsky23 Jul 2022 22:52 UTC
13 points
1 comment6 min readLW link

Cu­rat­ing “The Epistemic Se­quences” (list v.0.1)

Andrew_Critch23 Jul 2022 22:17 UTC
65 points
12 comments7 min readLW link

Room Opening

jefftk23 Jul 2022 21:00 UTC
8 points
3 comments4 min readLW link
(www.jefftk.com)

A Bias Against Altruism

Lone Pine23 Jul 2022 20:44 UTC
58 points
30 comments2 min readLW link

What En­vi­ron­ment Prop­er­ties Select Agents For World-Model­ing?

Thane Ruthenis23 Jul 2022 19:27 UTC
24 points
1 comment12 min readLW link

Which sin­gu­lar­ity schools plus the no sin­gu­lar­ity school was right?

Noosphere8923 Jul 2022 15:16 UTC
9 points
26 comments9 min readLW link

Ba­sic Post Scarcity Q&A

lorepieri23 Jul 2022 13:43 UTC
3 points
0 comments1 min readLW link
(lorenzopieri.com)

Ro­bust­ness to Scal­ing Down: More Im­por­tant Than I Thought

adamShimi23 Jul 2022 11:40 UTC
37 points
5 comments3 min readLW link

Eat­ing Boogers

George3d623 Jul 2022 11:20 UTC
17 points
5 comments6 min readLW link
(www.epistem.ink)

On Akra­sia, Habits and Re­ward Maximization

Aiyen23 Jul 2022 8:34 UTC
13 points
1 comment6 min readLW link

Which val­ues are sta­ble un­der on­tol­ogy shifts?

Richard_Ngo23 Jul 2022 2:40 UTC
73 points
48 comments3 min readLW link
(thinkingcomplete.blogspot.com)

Try­ing out Prompt Eng­ineer­ing on TruthfulQA

Megan Kinniment23 Jul 2022 2:04 UTC
10 points
0 comments8 min readLW link

Con­nor Leahy on Dy­ing with Dig­nity, EleutherAI and Conjecture

Michaël Trazzi22 Jul 2022 18:44 UTC
194 points
29 comments14 min readLW link
(theinsideview.ai)

Wy­clif’s Dust: the miss­ing chapter

David Hugh-Jones22 Jul 2022 18:27 UTC
9 points
0 comments4 min readLW link
(wyclif.substack.com)

Mak­ing DALL-E Count

DirectedEvolution22 Jul 2022 9:11 UTC
23 points
12 comments4 min readLW link

One-day ap­plied ra­tio­nal­ity work­shop in Ber­lin Aug 29 (af­ter LWCW)

[DEACTIVATED] Duncan Sabien22 Jul 2022 7:58 UTC
30 points
5 comments2 min readLW link

In­ter­nal Dou­ble Crux

CFAR!Duncan22 Jul 2022 4:34 UTC
88 points
15 comments12 min readLW link

Con­di­tion­ing Gen­er­a­tive Models with Restrictions

Adam Jermyn21 Jul 2022 20:33 UTC
18 points
4 comments8 min readLW link

Our Ex­ist­ing Solu­tions to AGI Align­ment (semi-safe)

Michael Soareverix21 Jul 2022 19:00 UTC
12 points
1 comment3 min readLW link

Chang­ing the world through slack & hobbies

Steven Byrnes21 Jul 2022 18:11 UTC
258 points
13 comments10 min readLW link

Which per­son­al­ities do we find in­tol­er­able?

weathersystems21 Jul 2022 15:56 UTC
10 points
3 comments6 min readLW link

YouTubeTV and Spoilers

Zvi21 Jul 2022 13:50 UTC
16 points
6 comments8 min readLW link
(thezvi.wordpress.com)

Covid 7/​21/​22: Fea­tur­ing ASPR

Zvi21 Jul 2022 13:50 UTC
27 points
0 comments14 min readLW link
(thezvi.wordpress.com)

[Question] How much to op­ti­mize for the short-timelines sce­nario?

SoerenMind21 Jul 2022 10:47 UTC
20 points
3 comments1 min readLW link

Is Gas Green?

ChristianKl21 Jul 2022 10:30 UTC
19 points
19 comments1 min readLW link

Why are poli­ti­ci­ans po­larized?

ErnestScribbler21 Jul 2022 8:17 UTC
13 points
24 comments7 min readLW link

[AN #173] Re­cent lan­guage model re­sults from DeepMind

Rohin Shah21 Jul 2022 2:30 UTC
37 points
9 comments8 min readLW link
(mailchi.mp)

Don’t take the or­ga­ni­za­tional chart literally

lc21 Jul 2022 0:56 UTC
52 points
21 comments4 min readLW link

Per­sonal fore­cast­ing ret­ro­spec­tive: 2020-2022

elifland21 Jul 2022 0:07 UTC
35 points
3 comments8 min readLW link
(www.foxy-scout.com)

Defin­ing Op­ti­miza­tion in a Deeper Way Part 3

J Bostock20 Jul 2022 22:06 UTC
8 points
0 comments2 min readLW link

Cog­ni­tive Risks of Ado­les­cent Binge Drinking

20 Jul 2022 21:10 UTC
70 points
12 comments10 min readLW link
(acesounderglass.com)

Why AGI Timeline Re­search/​Dis­course Might Be Overrated

Noosphere8920 Jul 2022 20:26 UTC
5 points
0 comments1 min readLW link
(forum.effectivealtruism.org)

En­light­en­ment Values in a Vuln­er­a­ble World

Maxwell Tabarrok20 Jul 2022 19:52 UTC
15 points
6 comments31 min readLW link
(maximumprogress.substack.com)

Coun­ter­ing ar­gu­ments against work­ing on AI safety

Rauno Arike20 Jul 2022 18:23 UTC
7 points
2 comments7 min readLW link

A Short In­tro to Humans

Ben Amitay20 Jul 2022 15:28 UTC
1 point
1 comment7 min readLW link

How to Diver­sify Con­cep­tual Align­ment: the Model Be­hind Refine

adamShimi20 Jul 2022 10:44 UTC
87 points
11 comments8 min readLW link

[Question] What are the sim­plest ques­tions in ap­plied ra­tio­nal­ity where you don’t know the an­swer to?

ChristianKl20 Jul 2022 9:53 UTC
26 points
11 comments1 min readLW link

AI Safety Cheat­sheet /​ Quick Reference

Zohar Jackson20 Jul 2022 9:39 UTC
3 points
0 comments1 min readLW link
(github.com)

Get­ting Un­stuck on Counterfactuals

Chris_Leong20 Jul 2022 5:31 UTC
7 points
1 comment2 min readLW link

Pit­falls with Proofs

scasper19 Jul 2022 22:21 UTC
19 points
21 comments8 min readLW link

A daily rou­tine I do for my AI safety re­search work

scasper19 Jul 2022 21:58 UTC
21 points
7 comments1 min readLW link

Progress links and tweets, 2022-07-19

jasoncrawford19 Jul 2022 20:50 UTC
11 points
1 comment1 min readLW link
(rootsofprogress.org)

Ap­pli­ca­tions are open for CFAR work­shops in Prague this fall!

John Steidley19 Jul 2022 18:29 UTC
64 points
3 comments2 min readLW link

Sex­ual Abuse at­ti­tudes might be infohazardous

Pseudonymous Otter19 Jul 2022 18:06 UTC
254 points
71 comments1 min readLW link

Spend­ing Up­date 2022

jefftk19 Jul 2022 14:10 UTC
28 points
0 comments3 min readLW link
(www.jefftk.com)

Abram Dem­ski’s ELK thoughts and pro­posal—distillation

Rubi J. Hudson19 Jul 2022 6:57 UTC
16 points
8 comments16 min readLW link

Bounded com­plex­ity of solv­ing ELK and its implications

Rubi J. Hudson19 Jul 2022 6:56 UTC
11 points
4 comments18 min readLW link

Help ARC eval­u­ate ca­pa­bil­ities of cur­rent lan­guage mod­els (still need peo­ple)

Beth Barnes19 Jul 2022 4:55 UTC
95 points
6 comments2 min readLW link

A Cri­tique of AI Align­ment Pessimism

ExCeph19 Jul 2022 2:28 UTC
9 points
1 comment9 min readLW link