Is check­ing that a state of the world is not dystopian eas­ier than con­struct­ing a non-dystopian state?

No77e27 Dec 2022 20:57 UTC
5 points
3 comments1 min readLW link

Crypto-cur­rency as pro-al­ign­ment mechanism

False Name27 Dec 2022 17:45 UTC
−10 points
2 comments2 min readLW link

My Reser­va­tions about Dis­cov­er­ing La­tent Knowl­edge (Burns, Ye, et al)

Robert_AIZI27 Dec 2022 17:27 UTC
50 points
0 comments4 min readLW link
(aizi.substack.com)

Things that can kill you quickly: What ev­ery­one should know about first aid

jasoncrawford27 Dec 2022 16:23 UTC
166 points
21 comments2 min readLW link1 review
(jasoncrawford.org)

[Question] Why The Fo­cus on Ex­pected Utility Max­imisers?

DragonGod27 Dec 2022 15:49 UTC
116 points
84 comments3 min readLW link

Pre­sump­tive Listen­ing: stick­ing to fa­mil­iar con­cepts and miss­ing the outer rea­son­ing paths

Remmelt27 Dec 2022 15:40 UTC
−14 points
8 comments2 min readLW link
(mflb.com)

Mere ex­po­sure effect: Bias in Eval­u­at­ing AGI X-Risks

27 Dec 2022 14:05 UTC
0 points
2 comments1 min readLW link

Hous­ing and Trans­porta­tion Roundup #2

Zvi27 Dec 2022 13:10 UTC
25 points
0 comments12 min readLW link
(thezvi.wordpress.com)

[Question] Are tul­pas moral pa­tients?

ChristianKl27 Dec 2022 11:30 UTC
16 points
28 comments1 min readLW link

Reflec­tions on my 5-month al­ign­ment up­skil­ling grant

Jay Bailey27 Dec 2022 10:51 UTC
82 points
4 comments8 min readLW link

In­sti­tu­tions Can­not Res­train Dark-Triad AI Exploitation

27 Dec 2022 10:34 UTC
5 points
0 comments5 min readLW link
(mflb.com)

In­tro­duc­tion: Bias in Eval­u­at­ing AGI X-Risks

27 Dec 2022 10:27 UTC
1 point
0 comments3 min readLW link

MDPs and the Bel­l­man Equa­tion, In­tu­itively Explained

Jack O'Brien27 Dec 2022 5:50 UTC
11 points
3 comments14 min readLW link

How ‘Hu­man-Hu­man’ dy­nam­ics give way to ‘Hu­man-AI’ and then ‘AI-AI’ dynamics

27 Dec 2022 3:16 UTC
−2 points
5 comments2 min readLW link
(mflb.com)

Nine Points of Col­lec­tive Insanity

27 Dec 2022 3:14 UTC
−2 points
3 comments1 min readLW link
(mflb.com)

Frac­tional Resignation

jefftk27 Dec 2022 2:30 UTC
18 points
6 comments1 min readLW link
(www.jefftk.com)

[Question] What poli­cies have most thor­oughly crip­pled (oth­er­wise-promis­ing) in­dus­tries or tech­nolo­gies?

benwr27 Dec 2022 2:25 UTC
40 points
4 comments1 min readLW link

Re­cent ad­vances in Nat­u­ral Lan­guage Pro­cess­ing—Some Woolly spec­u­la­tions (2019 es­say on se­man­tics and lan­guage mod­els)

philosophybear27 Dec 2022 2:11 UTC
1 point
0 comments7 min readLW link

Against Agents as an Ap­proach to Aligned Trans­for­ma­tive AI

DragonGod27 Dec 2022 0:47 UTC
12 points
9 comments2 min readLW link

Can we effi­ciently dis­t­in­guish differ­ent mechanisms?

paulfchristiano27 Dec 2022 0:20 UTC
88 points
30 comments16 min readLW link
(ai-alignment.com)

Air-gap­ping eval­u­a­tion and support

Ryan Kidd26 Dec 2022 22:52 UTC
53 points
1 comment2 min readLW link

Slightly against al­ign­ing with neo-luddites

Matthew Barnett26 Dec 2022 22:46 UTC
104 points
31 comments4 min readLW link

Avoid­ing per­pet­ual risk from TAI

scasper26 Dec 2022 22:34 UTC
15 points
6 comments5 min readLW link

An­nounc­ing: The In­de­pen­dent AI Safety Registry

Shoshannah Tekofsky26 Dec 2022 21:22 UTC
53 points
9 comments1 min readLW link

Are men harder to help?

braces26 Dec 2022 21:11 UTC
35 points
1 comment2 min readLW link

[Question] How much should I up­date on the fact that my den­tist is named Den­nis?

MichaelDickens26 Dec 2022 19:11 UTC
2 points
3 comments1 min readLW link

Theod­icy and the simu­la­tion hy­poth­e­sis, or: The prob­lem of simu­la­tor evil

philosophybear26 Dec 2022 18:55 UTC
6 points
12 comments19 min readLW link
(philosophybear.substack.com)

Safety of Self-Assem­bled Neu­ro­mor­phic Hardware

Can26 Dec 2022 18:51 UTC
15 points
2 comments10 min readLW link
(forum.effectivealtruism.org)

Co­her­ent ex­trap­o­lated dreaming

Alex Flint26 Dec 2022 17:29 UTC
38 points
10 comments17 min readLW link

An overview of some promis­ing work by ju­nior al­ign­ment researchers

Akash26 Dec 2022 17:23 UTC
34 points
0 comments4 min readLW link

Sols­tice song: Here Lies the Dragon

jchan26 Dec 2022 16:08 UTC
8 points
1 comment2 min readLW link

The Use­ful­ness Paradigm

Aprillion (Peter Hozák)26 Dec 2022 13:23 UTC
3 points
4 comments1 min readLW link

Look­ing Back on Posts From 2022

Zvi26 Dec 2022 13:20 UTC
49 points
8 comments17 min readLW link
(thezvi.wordpress.com)

Analo­gies be­tween Soft­ware Re­v­erse Eng­ineer­ing and Mechanis­tic Interpretability

26 Dec 2022 12:26 UTC
34 points
6 comments11 min readLW link
(www.neelnanda.io)

Mlyyrczo

lsusr26 Dec 2022 7:58 UTC
41 points
14 comments3 min readLW link

Causal ab­strac­tions vs infradistributions

Pablo Villalobos26 Dec 2022 0:21 UTC
20 points
0 comments6 min readLW link

Con­crete Steps to Get Started in Trans­former Mechanis­tic Interpretability

Neel Nanda25 Dec 2022 22:21 UTC
54 points
7 comments12 min readLW link
(www.neelnanda.io)

It’s time to worry about on­line pri­vacy again

Malmesbury25 Dec 2022 21:05 UTC
66 points
23 comments6 min readLW link

[Heb­bian Nat­u­ral Ab­strac­tions] Math­e­mat­i­cal Foundations

25 Dec 2022 20:58 UTC
15 points
2 comments6 min readLW link
(www.snellessen.com)

[Question] Or­a­cle AGI—How can it es­cape, other than se­cu­rity is­sues? (Steganog­ra­phy?)

RationalSieve25 Dec 2022 20:14 UTC
3 points
6 comments1 min readLW link

YCom­bi­na­tor fraud rates

Xodarap25 Dec 2022 19:21 UTC
56 points
3 comments1 min readLW link

How evolu­tion­ary lineages of LLMs can plan their own fu­ture and act on these plans

Roman Leventov25 Dec 2022 18:11 UTC
39 points
16 comments8 min readLW link

Ac­cu­rate Models of AI Risk Are Hyper­ex­is­ten­tial Exfohazards

Thane Ruthenis25 Dec 2022 16:50 UTC
30 points
38 comments9 min readLW link

ChatGPT is our Wright Brothers moment

Ron J25 Dec 2022 16:26 UTC
10 points
9 comments1 min readLW link

The Med­i­ta­tion on Winter

Raemon25 Dec 2022 16:12 UTC
58 points
3 comments3 min readLW link

I’ve up­dated to­wards AI box­ing be­ing sur­pris­ingly easy

Noosphere8925 Dec 2022 15:40 UTC
8 points
20 comments2 min readLW link

Take 14: Cor­rigi­bil­ity isn’t that great.

Charlie Steiner25 Dec 2022 13:04 UTC
15 points
3 comments3 min readLW link

Sim­plified Level Up

jefftk25 Dec 2022 13:00 UTC
12 points
16 comments2 min readLW link
(www.jefftk.com)

Hyper­finite graphs ~ manifolds

Alok Singh25 Dec 2022 12:24 UTC
11 points
5 comments2 min readLW link

In­con­sis­tent math is great

Alok Singh25 Dec 2022 3:20 UTC
1 point
2 comments1 min readLW link