[Question] How to bet against civ­i­liza­tional ad­e­quacy?

Wei Dai12 Aug 2022 23:33 UTC
54 points
17 comments1 min readLW link

In­fant AI Scenario

Nathan112312 Aug 2022 21:20 UTC
1 point
0 comments3 min readLW link

Deep­Mind al­ign­ment team opinions on AGI ruin arguments

Vika12 Aug 2022 21:06 UTC
376 points
37 comments14 min readLW link1 review

Dis­solve: The Petty Crimes of Blaise Pascal

JohnBuridan12 Aug 2022 20:04 UTC
17 points
4 comments6 min readLW link

The Host Minds of HBO’s West­world.

Nerret12 Aug 2022 18:53 UTC
1 point
0 comments3 min readLW link

What is es­ti­ma­tional pro­gram­ming? Squig­gle in context

Quinn12 Aug 2022 18:39 UTC
14 points
7 comments7 min readLW link

Over­sight Misses 100% of Thoughts The AI Does Not Think

johnswentworth12 Aug 2022 16:30 UTC
97 points
50 comments1 min readLW link

Timelines ex­pla­na­tion post part 1 of ?

Nathan Helm-Burger12 Aug 2022 16:13 UTC
10 points
1 comment2 min readLW link

A lit­tle play­ing around with Blen­der­bot3

Nathan Helm-Burger12 Aug 2022 16:06 UTC
9 points
0 comments1 min readLW link

Refin­ing the Sharp Left Turn threat model, part 1: claims and mechanisms

12 Aug 2022 15:17 UTC
85 points
4 comments3 min readLW link1 review
(vkrakovna.wordpress.com)

Ar­gu­ment by In­tel­lec­tual Ordeal

lc12 Aug 2022 13:03 UTC
26 points
5 comments5 min readLW link

Anti-squat­ted AI x-risk do­mains index

plex12 Aug 2022 12:01 UTC
56 points
6 comments1 min readLW link

[Question] Perfect Predictors

aditya malik12 Aug 2022 11:51 UTC
2 points
5 comments1 min readLW link

[Question] What are some good ar­gu­ments against build­ing new nu­clear power plants?

RomanS12 Aug 2022 7:32 UTC
16 points
15 comments2 min readLW link

Seek­ing PCK (Ped­a­gog­i­cal Con­tent Knowl­edge)

CFAR!Duncan12 Aug 2022 4:15 UTC
52 points
11 comments5 min readLW link

Ar­tifi­cial in­tel­li­gence wireheading

Big Tony12 Aug 2022 3:06 UTC
5 points
2 comments1 min readLW link

Dis­sected boxed AI

Nathan112312 Aug 2022 2:37 UTC
−8 points
2 comments1 min readLW link

Troll Timers

Screwtape12 Aug 2022 0:55 UTC
29 points
13 comments4 min readLW link

[Question] Se­ri­ously, what goes wrong with “re­ward the agent when it makes you smile”?

TurnTrout11 Aug 2022 22:22 UTC
86 points
42 comments2 min readLW link

En­cul­tured AI Pre-plan­ning, Part 2: Pro­vid­ing a Service

11 Aug 2022 20:11 UTC
33 points
4 comments3 min readLW link

My sum­mary of the al­ign­ment problem

Peter Hroššo11 Aug 2022 19:42 UTC
16 points
3 comments2 min readLW link
(threadreaderapp.com)

Lan­guage mod­els seem to be much bet­ter than hu­mans at next-to­ken prediction

11 Aug 2022 17:45 UTC
182 points
59 comments13 min readLW link1 review

In­tro­duc­ing Past­cast­ing: A tool for fore­cast­ing practice

Sage Future11 Aug 2022 17:38 UTC
95 points
10 comments2 min readLW link2 reviews

Pen­du­lums, Policy-Level De­ci­sion­mak­ing, Sav­ing State

CFAR!Duncan11 Aug 2022 16:47 UTC
26 points
3 comments8 min readLW link

Covid 8/​11/​22: The End Is Never The End

Zvi11 Aug 2022 16:20 UTC
28 points
11 comments16 min readLW link
(thezvi.wordpress.com)

Sin­ga­pore—Small ca­sual din­ner in Chi­na­town #4

Joe Rocca11 Aug 2022 12:30 UTC
3 points
3 comments1 min readLW link

Thoughts on the good reg­u­la­tor theorem

JonasMoss11 Aug 2022 12:08 UTC
12 points
0 comments4 min readLW link

How and why to turn ev­ery­thing into audio

11 Aug 2022 8:55 UTC
54 points
20 comments5 min readLW link

Shard The­ory: An Overview

David Udell11 Aug 2022 5:44 UTC
161 points
34 comments10 min readLW link

[Question] Do ad­vance­ments in De­ci­sion The­ory point to­wards moral ab­solutism?

Nathan112311 Aug 2022 0:59 UTC
0 points
4 comments4 min readLW link

The al­ign­ment prob­lem from a deep learn­ing perspective

Richard_Ngo10 Aug 2022 22:46 UTC
107 points
15 comments27 min readLW link1 review

How much al­ign­ment data will we need in the long run?

Jacob_Hilton10 Aug 2022 21:39 UTC
37 points
15 comments4 min readLW link

On Ego, Rein­car­na­tion, Con­scious­ness and The Uni­verse

qmaury10 Aug 2022 20:21 UTC
−3 points
6 comments5 min readLW link

For­mal­iz­ing Alignment

Marv K10 Aug 2022 18:50 UTC
4 points
0 comments2 min readLW link

How Do We Align an AGI Without Get­ting So­cially Eng­ineered? (Hint: Box It)

10 Aug 2022 18:14 UTC
28 points
30 comments11 min readLW link

Emer­gent Abil­ities of Large Lan­guage Models [Linkpost]

aogara10 Aug 2022 18:02 UTC
25 points
2 comments1 min readLW link
(arxiv.org)

How To Go From In­ter­pretabil­ity To Align­ment: Just Re­tar­get The Search

johnswentworth10 Aug 2022 16:08 UTC
179 points
33 comments3 min readLW link1 review

Us­ing GPT-3 to aug­ment hu­man intelligence

Henrik Karlsson10 Aug 2022 15:54 UTC
52 points
8 comments18 min readLW link
(escapingflatland.substack.com)

ACX meetup [Au­gust]

sallatik10 Aug 2022 9:54 UTC
1 point
1 comment1 min readLW link

Dissent Collusion

Screwtape10 Aug 2022 2:43 UTC
19 points
0 comments3 min readLW link

The Medium Is The Bandage

party girl10 Aug 2022 1:45 UTC
11 points
0 comments10 min readLW link

[Question] Why is in­creas­ing pub­lic aware­ness of AI safety not a pri­or­ity?

FinalFormal210 Aug 2022 1:28 UTC
−5 points
14 comments1 min readLW link

Man­i­fold x CSPI $25k Fore­cast­ing Tournament

David Chee9 Aug 2022 21:13 UTC
5 points
0 comments1 min readLW link
(www.cspicenter.com)

Pro­posal: Con­sider not us­ing dis­tance-di­rec­tion-di­men­sion words in ab­stract discussions

moridinamael9 Aug 2022 20:44 UTC
45 points
18 comments5 min readLW link

[Question] How would two su­per­in­tel­li­gent AIs in­ter­act, if they are un­al­igned with each other?

Nathan11239 Aug 2022 18:58 UTC
4 points
6 comments1 min readLW link

Disagree­ments about Align­ment: Why, and how, we should try to solve them

ojorgensen9 Aug 2022 18:49 UTC
11 points
2 comments16 min readLW link

Progress links and tweets, 2022-08-09

jasoncrawford9 Aug 2022 17:35 UTC
11 points
3 comments1 min readLW link
(rootsofprogress.org)

[Question] Is it pos­si­ble to find ven­ture cap­i­tal for AI re­search org with strong safety fo­cus?

AnonResearch9 Aug 2022 16:12 UTC
6 points
1 comment1 min readLW link

[Question] Many Gods re­fu­ta­tion and In­stru­men­tal Goals. (Proper one)

aditya malik9 Aug 2022 11:59 UTC
0 points
15 comments1 min readLW link

Con­tent gen­er­a­tion. Where do we draw the line?

Q Home9 Aug 2022 10:51 UTC
6 points
7 comments2 min readLW link