[Question] What’s up with the font size in the Mark­down text ed­i­tor?

Ege Erdil14 May 2022 21:12 UTC
7 points
1 comment1 min readLW link

[Link post] Promis­ing Paths to Align­ment—Con­nor Leahy | Talk

frances_lorenz14 May 2022 16:01 UTC
34 points
0 comments1 min readLW link

Inequal­ity is in­sep­a­rable from mar­kets

NathanBarnard14 May 2022 13:39 UTC
22 points
7 comments3 min readLW link

Pre­dict­ing the Elec­tions with Deep Learn­ing—Part 1 - Results

Quentin Chenevier14 May 2022 12:54 UTC
0 points
0 comments1 min readLW link

Clar­ify­ing the con­fu­sion around in­ner alignment

Rauno Arike13 May 2022 23:05 UTC
29 points
0 comments11 min readLW link

Costs and benefits of am­nio­cen­te­sis for nor­mal pregnancies

braces13 May 2022 22:47 UTC
13 points
4 comments3 min readLW link

Frame for Take-Off Speeds to in­form com­pute gov­er­nance & scal­ing alignment

Logan Riggs13 May 2022 22:23 UTC
15 points
2 comments2 min readLW link

Align­ment as Constraints

Logan Riggs13 May 2022 22:07 UTC
10 points
0 comments2 min readLW link

How close to nu­clear war did we get over Cuba?

NathanBarnard13 May 2022 19:58 UTC
13 points
0 comments10 min readLW link

Against Time in Agent Models

johnswentworth13 May 2022 19:55 UTC
62 points
13 comments3 min readLW link

Agency As a Nat­u­ral Abstraction

Thane Ruthenis13 May 2022 18:02 UTC
55 points
9 comments13 min readLW link

Fermi es­ti­ma­tion of the im­pact you might have work­ing on AI safety

Fabien Roger13 May 2022 17:49 UTC
6 points
0 comments1 min readLW link

“Tech com­pany sin­gu­lar­i­ties”, and steer­ing them to re­duce x-risk

Andrew_Critch13 May 2022 17:24 UTC
75 points
11 comments4 min readLW link

An ob­ser­va­tion about Hub­inger et al.’s frame­work for learned optimization

Spencer Becker-Kahn13 May 2022 16:20 UTC
34 points
9 comments8 min readLW link

[Question] The Eco­nomics of a New En­ergy Source

hatta_afiq13 May 2022 14:08 UTC
2 points
13 comments1 min readLW link

[Question] Still pos­si­ble to change user­name?

gabrielrecc13 May 2022 13:41 UTC
7 points
4 comments1 min readLW link

[Rough notes, BAIS] Hu­man val­ues and cycli­cal preferences

13 May 2022 13:28 UTC
5 points
0 comments4 min readLW link

[Question] Can mod­er­a­tors fix old se­quences posts?

EniScien13 May 2022 12:30 UTC
10 points
1 comment1 min readLW link

Deep­Mind is hiring for the Scal­able Align­ment and Align­ment Teams

13 May 2022 12:17 UTC
150 points
34 comments9 min readLW link

Thoughts on AI Safety Camp

Charlie Steiner13 May 2022 7:16 UTC
32 points
8 comments7 min readLW link

Deferring

owencb12 May 2022 23:56 UTC
18 points
2 comments11 min readLW link

RLHF

Ansh Radhakrishnan12 May 2022 21:18 UTC
18 points
5 comments5 min readLW link

[Question] What to do when start­ing a busi­ness in an im­mi­nent-AGI world?

ryan_b12 May 2022 21:07 UTC
25 points
7 comments1 min readLW link

In­ter­pretabil­ity’s Align­ment-Solv­ing Po­ten­tial: Anal­y­sis of 7 Scenarios

Evan R. Murphy12 May 2022 20:01 UTC
53 points
0 comments59 min readLW link

In­tro­duc­tion to the se­quence: In­ter­pretabil­ity Re­search for the Most Im­por­tant Century

Evan R. Murphy12 May 2022 19:59 UTC
16 points
0 comments8 min readLW link

A ten­ta­tive di­alogue with a Friendly-boxed-su­per-AGI on brain uploads

Ramiro P.12 May 2022 19:40 UTC
1 point
12 comments4 min readLW link

The Last Paperclip

Logan Zoellner12 May 2022 19:25 UTC
61 points
15 comments17 min readLW link

Deep­mind’s Gato: Gen­er­al­ist Agent

Daniel Kokotajlo12 May 2022 16:01 UTC
165 points
62 comments1 min readLW link

“A Gen­er­al­ist Agent”: New Deep­Mind Publication

1a3orn12 May 2022 15:30 UTC
79 points
43 comments1 min readLW link

Covid 5/​12/​22: Other Priorities

Zvi12 May 2022 13:30 UTC
31 points
4 comments15 min readLW link
(thezvi.wordpress.com)

[Question] How would pub­lic me­dia out­lets need to be gov­erned to cover all poli­ti­cal views?

ChristianKl12 May 2022 12:55 UTC
13 points
14 comments1 min readLW link

[Question] What’s keep­ing con­cerned ca­pa­bil­ities gain re­searchers from leav­ing the field?

sovran12 May 2022 12:16 UTC
19 points
4 comments1 min readLW link

Pos­i­tive out­comes un­der an un­al­igned AGI takeover

Yitz12 May 2022 7:45 UTC
19 points
10 comments3 min readLW link

[Question] What are your recom­men­da­tions for tech­ni­cal AI al­ign­ment pod­casts?

Evan_Gaensbauer11 May 2022 21:52 UTC
5 points
4 comments1 min readLW link

Grace­fully cor­rect­ing un­cal­ibrated shame

AF202211 May 2022 19:51 UTC
−31 points
34 comments5 min readLW link

[In­tro to brain-like-AGI safety] 14. Con­trol­led AGI

Steven Byrnes11 May 2022 13:17 UTC
41 points
25 comments19 min readLW link

Pro­jec­tLawful.com: Eliezer’s lat­est story, past 1M words

Eliezer Yudkowsky11 May 2022 6:18 UTC
213 points
112 comments1 min readLW link4 reviews

An In­side View of AI Alignment

Ansh Radhakrishnan11 May 2022 2:16 UTC
32 points
2 comments2 min readLW link

Fight­ing in var­i­ous places for a re­ally long time

KatjaGrace11 May 2022 1:50 UTC
36 points
12 comments4 min readLW link
(worldspiritsockpuppet.com)

Stuff I might do if I had covid

KatjaGrace11 May 2022 0:00 UTC
39 points
9 comments1 min readLW link
(worldspiritsockpuppet.com)

Crises Don’t Need Your Software

GabrielExists10 May 2022 21:06 UTC
59 points
18 comments6 min readLW link

Ceiling Fan Air Filter

jefftk10 May 2022 14:20 UTC
18 points
9 comments1 min readLW link
(www.jefftk.com)

The limits of AI safety via debate

Marius Hobbhahn10 May 2022 13:33 UTC
29 points
7 comments10 min readLW link

Ex­am­in­ing Arm­strong’s cat­e­gory of gen­er­al­ized models

Morgan_Rogers10 May 2022 9:07 UTC
14 points
0 comments7 min readLW link

Dath Ilani Rule of Law

David Udell10 May 2022 6:17 UTC
18 points
25 comments4 min readLW link

AI safety should be made more ac­cessible us­ing non text-based media

Massimog10 May 2022 3:14 UTC
2 points
4 comments4 min readLW link

LessWrong Now Has Dark Mode

jimrandomh10 May 2022 1:21 UTC
134 points
31 comments1 min readLW link

Con­di­tions for math­e­mat­i­cal equiv­alence of Stochas­tic Gra­di­ent Des­cent and Nat­u­ral Selection

Oliver Sourbut9 May 2022 21:38 UTC
61 points
19 comments8 min readLW link1 review
(www.oliversourbut.net)

AI Align­ment YouTube Playlists

9 May 2022 21:33 UTC
30 points
4 comments1 min readLW link

When is AI safety re­search harm­ful?

NathanBarnard9 May 2022 18:19 UTC
2 points
0 comments8 min readLW link