D&D.Sci 5E: Re­turn of the League of Defenders

aphyerMay 26, 2023, 8:39 PM
42 points
11 comments3 min readLW link

Seek­ing (Paid) Case Stud­ies on Standards

HoldenKarnofskyMay 26, 2023, 5:58 PM
69 points
9 comments11 min readLW link

Con­di­tional Pre­dic­tion with Zero-Sum Train­ing Solves Self-Fulfilling Prophecies

May 26, 2023, 5:44 PM
88 points
13 comments24 min readLW link

Re­quest: stop ad­vanc­ing AI capabilities

So8resMay 26, 2023, 5:42 PM
154 points
24 comments1 min readLW link

Bandgaps, Brains, and Bioweapons: The limi­ta­tions of com­pu­ta­tional sci­ence and what it means for AGI

titotalMay 26, 2023, 3:57 PM
36 points
20 commentsLW link

The Amer­i­can In­for­ma­tion Revolu­tion in Global Perspective

jasoncrawfordMay 26, 2023, 12:39 PM
16 points
1 comment5 min readLW link
(rootsofprogress.org)

He­lio-Se­lenic Laser Te­lescope (in SPACE!?)

Alexander Gietelink OldenzielMay 26, 2023, 11:24 AM
8 points
2 comments4 min readLW link

[Question] Why is vi­o­lence against AI labs a taboo?

ArisCMay 26, 2023, 8:00 AM
−21 points
63 comments1 min readLW link

Where do you lie on two axes of world ma­nipu­la­bil­ity?

Max HMay 26, 2023, 3:04 AM
31 points
15 comments3 min readLW link

Some thoughts on au­tomat­ing al­ign­ment research

Lukas FinnvedenMay 26, 2023, 1:50 AM
30 points
4 comments6 min readLW link

[Question] What’s your view­point on the like­li­hood of GPT-5 be­ing able to au­tonomously cre­ate, train, and im­ple­ment an AI su­pe­rior to GPT-5?

Super AGIMay 26, 2023, 1:43 AM
7 points
15 comments1 min readLW link

Be­fore smart AI, there will be many mediocre or spe­cial­ized AIs

Lukas FinnvedenMay 26, 2023, 1:38 AM
58 points
14 comments9 min readLW link1 review

how hu­mans are aligned

bhauthMay 26, 2023, 12:09 AM
14 points
3 comments1 min readLW link

[Question] What ve­gan food re­sources have you found use­ful?

ElizabethMay 25, 2023, 10:46 PM
29 points
6 commentsLW link

Mob and Bailey

ScrewtapeMay 25, 2023, 10:14 PM
82 points
17 comments7 min readLW link1 review

Look At What’s In Front Of You (Con­clu­sion to The Nuts and Bolts of Nat­u­ral­ism)

LoganStrohlMay 25, 2023, 7:00 PM
50 points
1 comment2 min readLW link

[Mar­ket] Will AI xrisk seem to be han­dled se­ri­ously by the end of 2026?

tailcalledMay 25, 2023, 6:51 PM
15 points
2 comments1 min readLW link
(manifold.markets)

[Question] What should my col­lege ma­jor be if I want to do AI al­ign­ment re­search?

metachiralityMay 25, 2023, 6:23 PM
8 points
7 comments1 min readLW link

Is be­hav­ioral safety “solved” in non-ad­ver­sar­ial con­di­tions?

Robert_AIZIMay 25, 2023, 5:56 PM
26 points
8 comments2 min readLW link
(aizi.substack.com)

Book Re­view: How Minds Change

bc4026bd4aaa5b7feMay 25, 2023, 5:55 PM
313 points
52 comments15 min readLW link

Self-ad­ministered EMDR with­out a ther­a­pist is very use­ful for a lot of things!

EternallyBlissfulMay 25, 2023, 5:54 PM
49 points
12 comments11 min readLW link

Re­cur­ren­tGPT: a loom-type tool with a twist

mishkaMay 25, 2023, 5:09 PM
10 points
0 comments3 min readLW link
(arxiv.org)

The Ge­nie in the Bot­tle: An In­tro­duc­tion to AI Align­ment and Risk

SnorkelfarsanMay 25, 2023, 4:30 PM
5 points
1 comment25 min readLW link

AI #13: Po­ten­tial Al­gorith­mic Improvements

ZviMay 25, 2023, 3:40 PM
45 points
4 comments67 min readLW link
(thezvi.wordpress.com)

Solv­ing the Mechanis­tic In­ter­pretabil­ity challenges: EIS VII Challenge 2

May 25, 2023, 3:37 PM
71 points
1 comment13 min readLW link

Malthu­sian Com­pe­ti­tion (not as bad as it seems)

Logan ZoellnerMay 25, 2023, 3:30 PM
6 points
11 comments2 min readLW link

You Don’t Always Need Indexes

jefftkMay 25, 2023, 2:20 PM
22 points
6 comments1 min readLW link
(www.jefftk.com)

The­o­ries of Biolog­i­cal Inspiration

Eric ZhangMay 25, 2023, 1:07 PM
7 points
3 comments1 min readLW link

Eval­u­at­ing strate­gic rea­son­ing in GPT models

phelps-sgMay 25, 2023, 11:51 AM
4 points
1 comment8 min readLW link

Re­quire­ments for a STEM-ca­pa­ble AGI Value Learner (my Case for Less Doom)

RogerDearnaleyMay 25, 2023, 9:26 AM
33 points
3 comments15 min readLW link

Align­ment solu­tions for weak AI don’t (nec­es­sar­ily) scale to strong AI

Michael TontchevMay 25, 2023, 8:26 AM
6 points
0 comments5 min readLW link

[Question] What fea­tures would you like to see in a per­sonal for­cast­ing /​ pre­dic­tion track­ing app?

regnargMay 25, 2023, 8:18 AM
9 points
0 comments1 min readLW link

An­nounc­ing the Con­fido app: bring­ing fore­cast­ing to everyone

regnargMay 25, 2023, 8:18 AM
6 points
2 comments10 min readLW link
(forum.effectivealtruism.org)

But What If We Ac­tu­ally Want To Max­i­mize Paper­clips?

snerxMay 25, 2023, 7:13 AM
−17 points
6 comments7 min readLW link

Deep­Mind: Model eval­u­a­tion for ex­treme risks

Zach Stein-PerlmanMay 25, 2023, 3:00 AM
94 points
12 comments1 min readLW link1 review
(arxiv.org)

Why I’m Not (Yet) A Full-Time Tech­ni­cal Align­ment Researcher

Nicholas / Heather KrossMay 25, 2023, 1:26 AM
41 points
21 comments4 min readLW link
(www.thinkingmuchbetter.com)

Two ideas for al­ign­ment, per­pet­ual mu­tual dis­trust and induction

APaleBlueDotMay 25, 2023, 12:56 AM
1 point
2 comments4 min readLW link

Eval­u­at­ing Ev­i­dence Re­con­struc­tions of Mock Crimes -Sub­mis­sion 2

Alan E DunneMay 24, 2023, 10:17 PM
−1 points
1 comment3 min readLW link

[Linkpost] In­ter­pretabil­ity Dreams

DanielFilanMay 24, 2023, 9:08 PM
39 points
2 comments2 min readLW link
(transformer-circuits.pub)

Rishi Su­nak men­tions “ex­is­ten­tial threats” in talk with OpenAI, Deep­Mind, An­thropic CEOs

May 24, 2023, 9:06 PM
34 points
1 comment1 min readLW link
(www.gov.uk)

If you’re not a morn­ing per­son, con­sider quit­ting allergy pills

Brendan LongMay 24, 2023, 8:11 PM
8 points
3 comments1 min readLW link

Adum­bra­tions on AGI from an outsider

nicholashaldenMay 24, 2023, 5:41 PM
57 points
44 comments8 min readLW link
(nicholashalden.home.blog)

Open Thread With Ex­per­i­men­tal Fea­ture: Reactions

jimrandomhMay 24, 2023, 4:46 PM
101 points
189 comments3 min readLW link

A re­jec­tion of the Orthog­o­nal­ity Thesis

ArisCMay 24, 2023, 4:37 PM
−2 points
11 comments2 min readLW link
(medium.com)

Aligned AI via mon­i­tor­ing ob­jec­tives in Au­toGPT-like systems

Paul CologneseMay 24, 2023, 3:59 PM
27 points
4 comments4 min readLW link

The Office of Science and Tech­nol­ogy Policy put out a re­quest for in­for­ma­tion on A.I.

HiroSakurabaMay 24, 2023, 1:33 PM
59 points
4 comments1 min readLW link
(www.whitehouse.gov)

ChatGPT (May 2023) on De­sign­ing Friendly Superintelligence

Mitchell_PorterMay 24, 2023, 10:47 AM
5 points
0 comments1 min readLW link
(singularitypolitics.wordpress.com)

No—AI is just as en­ergy-effi­cient as your brain.

Maxwell ClarkeMay 24, 2023, 2:30 AM
11 points
7 comments1 min readLW link

[Question] What pro­jects and efforts are there to pro­mote AI safety re­search?

Christopher KingMay 24, 2023, 12:33 AM
4 points
0 comments1 min readLW link

My May 2023 pri­ori­ties for AI x-safety: more em­pa­thy, more unifi­ca­tion of con­cerns, and less vil­ifi­ca­tion of OpenAI

Andrew_CritchMay 24, 2023, 12:02 AM
268 points
39 comments8 min readLW link