Ex­ams-Only Universities

Mati_Roy6 Nov 2022 22:05 UTC
80 points
40 comments2 min readLW link

Democ­racy Is in Danger, but Not for the Rea­sons You Think

ExCeph6 Nov 2022 21:15 UTC
−7 points
4 comments12 min readLW link
(ginnungagapfoundation.wordpress.com)

Play­ground Game: Monster

jefftk6 Nov 2022 16:00 UTC
14 points
4 comments1 min readLW link
(www.jefftk.com)

[Question] Has Pas­cal’s Mug­ging prob­lem been com­pletely solved yet?

EniScien6 Nov 2022 12:52 UTC
3 points
11 comments1 min readLW link

[Question] Should I Pur­sue a PhD?

DragonGod6 Nov 2022 10:58 UTC
8 points
8 comments2 min readLW link

You won’t solve al­ign­ment with­out agent foundations

Mikhail Samin6 Nov 2022 8:07 UTC
29 points
3 comments8 min readLW link

Word-Dis­tance vs Idea-Dis­tance: The Case for Lanoitaring

Sable6 Nov 2022 5:25 UTC
7 points
7 comments7 min readLW link
(affablyevil.substack.com)

Ap­ple Cider Syrup

jefftk6 Nov 2022 2:10 UTC
11 points
6 comments1 min readLW link
(www.jefftk.com)

What is epi­ge­net­ics?

Metacelsus6 Nov 2022 1:24 UTC
78 points
4 comments6 min readLW link
(denovo.substack.com)

Response

Jarred Filmer6 Nov 2022 1:03 UTC
29 points
2 comments12 min readLW link

[Question] Has any­one in­creased their AGI timelines?

Darren McKee6 Nov 2022 0:03 UTC
39 points
12 comments1 min readLW link

Take­aways from a sur­vey on AI al­ign­ment resources

DanielFilan5 Nov 2022 23:40 UTC
73 points
10 comments6 min readLW link1 review
(danielfilan.com)

Un­pri­ca­ble In­for­ma­tion and Cer­tifi­cate Hell

eva_5 Nov 2022 22:56 UTC
13 points
2 comments6 min readLW link

Recom­mend HAIST re­sources for as­sess­ing the value of RLHF-re­lated al­ign­ment research

5 Nov 2022 20:58 UTC
26 points
9 comments3 min readLW link

In­stead of tech­ni­cal re­search, more peo­ple should fo­cus on buy­ing time

5 Nov 2022 20:43 UTC
101 points
45 comments14 min readLW link

Prov­ably Hon­est—A First Step

Srijanak De5 Nov 2022 19:18 UTC
10 points
2 comments8 min readLW link

Should AI fo­cus on prob­lem-solv­ing or strate­gic plan­ning? Why not both?

Oliver Siegel5 Nov 2022 19:17 UTC
−12 points
3 comments1 min readLW link

How to store hu­man val­ues on a computer

Oliver Siegel5 Nov 2022 19:17 UTC
−12 points
17 comments1 min readLW link

The Slip­pery Slope from DALLE-2 to Deep­fake Anarchy

scasper5 Nov 2022 14:53 UTC
17 points
9 comments11 min readLW link

When can a mimic sur­prise you? Why gen­er­a­tive mod­els han­dle seem­ingly ill-posed problems

David Johnston5 Nov 2022 13:19 UTC
8 points
4 comments16 min readLW link

My sum­mary of “Prag­matic AI Safety”

Eleni Angelou5 Nov 2022 12:54 UTC
3 points
0 comments5 min readLW link

Re­view of the Challenge

SD Marlow5 Nov 2022 6:38 UTC
−14 points
5 comments2 min readLW link

Spec­trum of Independence

jefftk5 Nov 2022 2:40 UTC
43 points
7 comments1 min readLW link
(www.jefftk.com)

[pa­per link] In­ter­pret­ing sys­tems as solv­ing POMDPs: a step to­wards a for­mal un­der­stand­ing of agency

the gears to ascension5 Nov 2022 1:06 UTC
13 points
2 comments1 min readLW link
(www.semanticscholar.org)

Me­tac­u­lus is seek­ing Soft­ware Engineers

dschwarz5 Nov 2022 0:42 UTC
18 points
0 comments1 min readLW link
(apply.workable.com)

Should we “go against na­ture”?

jasoncrawford4 Nov 2022 22:14 UTC
10 points
3 comments2 min readLW link
(rootsofprogress.org)

How much should we care about non-hu­man an­i­mals?

bokov4 Nov 2022 21:36 UTC
17 points
8 comments2 min readLW link
(www.lesswrong.com)

For ELK truth is mostly a distraction

c.trout4 Nov 2022 21:14 UTC
44 points
0 comments21 min readLW link

Toy Models and Tegum Products

Adam Jermyn4 Nov 2022 18:51 UTC
28 points
7 comments5 min readLW link

Fol­low up to med­i­cal miracle

Elizabeth4 Nov 2022 18:00 UTC
77 points
5 comments6 min readLW link
(acesounderglass.com)

Cross-Void Optimization

pneumynym4 Nov 2022 17:47 UTC
1 point
1 comment8 min readLW link

Monthly Shorts 10/​22

Celer4 Nov 2022 16:30 UTC
12 points
0 comments6 min readLW link
(keller.substack.com)

Weekly Roundup #4

Zvi4 Nov 2022 15:00 UTC
42 points
1 comment6 min readLW link
(thezvi.wordpress.com)

A new place to dis­cuss cog­ni­tive sci­ence, ethics and hu­man alignment

Daniel_Friedrich4 Nov 2022 14:34 UTC
3 points
4 comments2 min readLW link
(www.facebook.com)

A new­comer’s guide to the tech­ni­cal AI safety field

zeshen4 Nov 2022 14:29 UTC
42 points
3 comments10 min readLW link

[Question] Are al­ign­ment re­searchers de­vot­ing enough time to im­prov­ing their re­search ca­pac­ity?

Carson Jones4 Nov 2022 0:58 UTC
13 points
3 comments3 min readLW link

[Question] Don’t you think RLHF solves outer al­ign­ment?

Charbel-Raphaël4 Nov 2022 0:36 UTC
9 points
23 comments1 min readLW link

Mechanis­tic In­ter­pretabil­ity as Re­v­erse Eng­ineer­ing (fol­low-up to “cars and elephants”)

David Scott Krueger (formerly: capybaralet)3 Nov 2022 23:19 UTC
28 points
3 comments1 min readLW link

[Question] Could a Supreme Court suit work to solve NEPA prob­lems?

ChristianKl3 Nov 2022 21:10 UTC
15 points
0 comments1 min readLW link

[Video] How hav­ing Fast Fourier Trans­forms sooner could have helped with Nu­clear Disar­ma­ment—Veritasium

mako yass3 Nov 2022 21:04 UTC
17 points
1 comment1 min readLW link
(www.youtube.com)

Fur­ther con­sid­er­a­tions on the Ev­i­den­tial­ist’s Wager

Martín Soto3 Nov 2022 20:06 UTC
3 points
9 comments8 min readLW link

AI as a Civ­i­liza­tional Risk Part 6/​6: What can be done

PashaKamyshev3 Nov 2022 19:48 UTC
2 points
4 comments4 min readLW link

A Mys­tery About High Di­men­sional Con­cept Encoding

Fabien Roger3 Nov 2022 17:05 UTC
46 points
13 comments7 min readLW link

Why do we post our AI safety plans on the In­ter­net?

Peter S. Park3 Nov 2022 16:02 UTC
4 points
4 comments11 min readLW link

Mul­ti­ple De­ploy-Key Repos

jefftk3 Nov 2022 15:10 UTC
15 points
0 comments1 min readLW link
(www.jefftk.com)

Covid 11/​3/​22: Ask­ing Forgiveness

Zvi3 Nov 2022 13:50 UTC
23 points
3 comments6 min readLW link
(thezvi.wordpress.com)

Ad­ver­sar­ial Poli­cies Beat Pro­fes­sional-Level Go AIs

sanxiyn3 Nov 2022 13:27 UTC
31 points
35 comments1 min readLW link
(goattack.alignmentfund.org)

K-types vs T-types — what pri­ors do you have?

Cleo Nardo3 Nov 2022 11:29 UTC
74 points
25 comments7 min readLW link

In­for­ma­tion Mar­kets 2: Op­ti­mally Shaped Re­ward Bets

eva_3 Nov 2022 11:08 UTC
9 points
0 comments3 min readLW link

The Ra­tional Utili­tar­ian Love Move­ment (A His­tor­i­cal Ret­ro­spec­tive)

Caleb Biddulph3 Nov 2022 7:11 UTC
3 points
0 comments1 min readLW link
(ratutilove.substack.com)