RSS

[Question] How do we iden­tify bot­tle­necks to sci­en­tific and tech­nolog­i­cal progress?

NaiveTortoise31 Dec 2018 20:21 UTC
31 points
12 comments2 min readLW link

Why do Con­tem­pla­tive Prac­ti­tion­ers Make so Many Me­ta­phys­i­cal Claims?

romeostevensit31 Dec 2018 19:44 UTC
59 points
15 comments6 min readLW link

[Question] What do you do when you find out you have in­con­sis­tent prob­a­bil­ities?

NunoSempere31 Dec 2018 18:13 UTC
15 points
7 comments1 min readLW link

[Question] Is there a stan­dard dis­cus­sion of veg­e­tar­i­anism/​ve­g­anism?

Sherrinford30 Dec 2018 20:22 UTC
4 points
17 comments1 min readLW link

Con­cep­tual Anal­y­sis for AI Align­ment

David Scott Krueger (formerly: capybaralet)30 Dec 2018 0:46 UTC
26 points
3 comments2 min readLW link

[Question] How did academia en­sure pa­pers were cor­rect in the early 20th Cen­tury?

Ben Pace29 Dec 2018 23:37 UTC
99 points
17 comments2 min readLW link1 review

[Question] What makes peo­ple in­tel­lec­tu­ally ac­tive?

abramdemski29 Dec 2018 22:29 UTC
116 points
71 comments1 min readLW link

Why I ex­pect suc­cess­ful (nar­row) alignment

Tobias_Baumann29 Dec 2018 15:44 UTC
8 points
12 comments1 min readLW link
(s-risks.org)

Pe­nal­iz­ing Im­pact via At­tain­able Utility Preservation

TurnTrout28 Dec 2018 21:46 UTC
24 points
0 comments3 min readLW link
(arxiv.org)

Akra­sia is con­fu­sion about what you want

Gordon Seidoh Worley28 Dec 2018 21:09 UTC
26 points
7 comments9 min readLW link

Isaac Asi­mov’s pre­dic­tions for 2019 from 1984

Jan_Kulveit28 Dec 2018 9:51 UTC
39 points
1 comment2 min readLW link
(www.thestar.com)

1987 Sci-Fi Authors Time­cap­sule Pre­dic­tions For 2012

namespace28 Dec 2018 6:50 UTC
20 points
3 comments1 min readLW link
(web.archive.org)

[Question] In what ways are holi­days good?

DanielFilan28 Dec 2018 0:42 UTC
20 points
19 comments1 min readLW link

[Question] Sun­screen. When? Why? Why not?

Viktor Riabtsev27 Dec 2018 22:04 UTC
5 points
22 comments1 min readLW link

[Question] Can dy­ing peo­ple “hold on” for some­thing they are wait­ing for?

Raemon27 Dec 2018 19:53 UTC
28 points
7 comments1 min readLW link

[Question] What does it mean to “be­lieve” a thing to be true?

Bae's Theorem27 Dec 2018 13:28 UTC
1 point
13 comments1 min readLW link

On Disingenuity

Chris_Leong26 Dec 2018 17:08 UTC
28 points
2 comments1 min readLW link

Rein­ter­pret­ing “AI and Com­pute”

habryka25 Dec 2018 21:12 UTC
30 points
9 comments1 min readLW link
(aiimpacts.org)

Align­ment Newslet­ter #38

Rohin Shah25 Dec 2018 16:10 UTC
9 points
0 comments8 min readLW link
(mailchi.mp)

[Question] What are the ax­ioms of ra­tio­nal­ity?

Yoav Ravid25 Dec 2018 6:47 UTC
2 points
8 comments1 min readLW link

Con­trite Strate­gies and The Need For Standards

sarahconstantin24 Dec 2018 18:30 UTC
125 points
5 comments4 min readLW link
(srconstantin.wordpress.com)

[Video] Why Not Just: Think of AGI Like a Cor­po­ra­tion? (Robert Miles)

habryka23 Dec 2018 21:49 UTC
17 points
1 comment9 min readLW link
(www.youtube.com)

[Question] Best ar­gu­ments against wor­ry­ing about AI risk?

Chris_Leong23 Dec 2018 14:57 UTC
15 points
16 comments1 min readLW link

[Question] Why Don’t Creators Switch to their Own Plat­forms?

Jacob Falkovich23 Dec 2018 4:46 UTC
42 points
17 comments1 min readLW link

Cog­ni­tive Bias of AI Re­searchers?

Mindey22 Dec 2018 9:20 UTC
9 points
7 comments1 min readLW link

Spaghetti Towers

eukaryote22 Dec 2018 5:29 UTC
187 points
28 comments3 min readLW link1 review
(eukaryotewritesblog.com)

An­thropic prob­a­bil­ities and cost functions

Stuart_Armstrong21 Dec 2018 17:54 UTC
16 points
1 comment1 min readLW link

Stand­ing on a pile of corpses

Jsevillamol21 Dec 2018 10:36 UTC
35 points
0 comments2 min readLW link

Sys­tems Eng­ineer­ing and the META Program

ryan_b20 Dec 2018 20:19 UTC
31 points
3 comments1 min readLW link

The Pavlov Strategy

sarahconstantin20 Dec 2018 16:20 UTC
247 points
13 comments4 min readLW link
(srconstantin.wordpress.com)

[Question] What self-help has helped you?

Gordon Seidoh Worley20 Dec 2018 3:31 UTC
35 points
22 comments1 min readLW link

Defin­ing Freedom

pku20 Dec 2018 2:41 UTC
8 points
7 comments3 min readLW link

Rea­sons com­pute may not drive AI ca­pa­bil­ities growth

Tristan H19 Dec 2018 22:13 UTC
42 points
10 comments8 min readLW link

An­thropic para­doxes trans­posed into An­thropic De­ci­sion Theory

Stuart_Armstrong19 Dec 2018 18:07 UTC
17 points
23 comments4 min readLW link

18-month fol­low-up on my self-con­cept work

Kaj_Sotala18 Dec 2018 17:40 UTC
60 points
4 comments8 min readLW link
(kajsotala.fi)

[Question] Ex­pe­riences of Self-deception

Bucky18 Dec 2018 11:10 UTC
15 points
3 comments1 min readLW link

2018 AI Align­ment Liter­a­ture Re­view and Char­ity Comparison

Larks18 Dec 2018 4:46 UTC
190 points
26 comments62 min readLW link1 review

Equiv­alence of State Machines and Coroutines

Martin Sustrik18 Dec 2018 4:40 UTC
12 points
1 comment1 min readLW link
(250bpm.com)

You can be wrong about what you like, and you of­ten are

Adam Zerner17 Dec 2018 23:49 UTC
30 points
21 comments4 min readLW link

Align­ment Newslet­ter #37

Rohin Shah17 Dec 2018 19:10 UTC
25 points
4 comments10 min readLW link
(mailchi.mp)

A sim­ple ap­proach to 5-and-10

Gurkenglas17 Dec 2018 18:33 UTC
5 points
10 comments1 min readLW link

Two Ne­glected Prob­lems in Hu­man-AI Safety

Wei Dai16 Dec 2018 22:13 UTC
98 points
24 comments2 min readLW link

Bab­ble, Learn­ing, and the Typ­i­cal Mind Fallacy

NaiveTortoise16 Dec 2018 16:51 UTC
6 points
0 comments1 min readLW link
(an1lam.github.io)

[Question] What are some con­crete prob­lems about log­i­cal coun­ter­fac­tu­als?

Chris_Leong16 Dec 2018 10:20 UTC
25 points
4 comments1 min readLW link

The E-Coli Test for AI Alignment

johnswentworth16 Dec 2018 8:10 UTC
70 points
24 comments1 min readLW link

on wellunderstoodness

Quinn16 Dec 2018 7:22 UTC
9 points
2 comments4 min readLW link

Sabine “Bee” Hossen­felder (and Robin Han­son) on How to fix Academia with Pre­dic­tion Markets

shminux16 Dec 2018 6:37 UTC
12 points
0 comments1 min readLW link
(backreaction.blogspot.com)

New edi­tion of “Ra­tion­al­ity: From AI to Zom­bies”

Rob Bensinger15 Dec 2018 21:33 UTC
84 points
27 comments2 min readLW link

Gw­ern about cen­taurs: there is no chance that any use­ful man+ma­chine com­bi­na­tion will work to­gether for more than 10 years, as hu­mans soon will be only a liability

avturchin15 Dec 2018 21:32 UTC
33 points
4 comments1 min readLW link
(www.reddit.com)

Ar­gue Poli­tics* With Your Best Friends

sarahconstantin15 Dec 2018 19:00 UTC
75 points
6 comments6 min readLW link
(srconstantin.wordpress.com)

In­ter­pret­ing ge­netic testing

jefftk15 Dec 2018 15:56 UTC
24 points
1 comment2 min readLW link

[Question] What is ab­strac­tion?

Adam Zerner15 Dec 2018 8:36 UTC
25 points
11 comments4 min readLW link

Player vs. Char­ac­ter: A Two-Level Model of Ethics

sarahconstantin14 Dec 2018 19:40 UTC
88 points
27 comments7 min readLW link3 reviews
(srconstantin.wordpress.com)

Med­i­ta­tions on Momentum

Richard Meadows14 Dec 2018 10:53 UTC
103 points
32 comments10 min readLW link

Three AI Safety Re­lated Ideas

Wei Dai13 Dec 2018 21:32 UTC
68 points
38 comments2 min readLW link

An Ex­ten­sive Cat­e­gori­sa­tion of In­finite Paradoxes

Chris_Leong13 Dec 2018 18:36 UTC
−14 points
48 comments13 min readLW link

The Bat and Ball Prob­lem Revisited

drossbucket13 Dec 2018 7:16 UTC
69 points
30 comments15 min readLW link2 reviews

Multi-agent pre­dic­tive minds and AI alignment

Jan_Kulveit12 Dec 2018 23:48 UTC
63 points
18 comments10 min readLW link

In­ter­net Search Tips: how I use Google/​Google Scholar/​Lib­gen

gwern12 Dec 2018 14:50 UTC
51 points
0 comments1 min readLW link
(www.gwern.net)

Should ethi­cists be in­side or out­side a pro­fes­sion?

Eliezer Yudkowsky12 Dec 2018 1:40 UTC
91 points
7 comments9 min readLW link

Align­ment Newslet­ter #36

Rohin Shah12 Dec 2018 1:10 UTC
21 points
0 comments11 min readLW link
(mailchi.mp)

A hun­dred Shakespeares

Stuart_Armstrong11 Dec 2018 23:11 UTC
29 points
5 comments2 min readLW link

Norms of Mem­ber­ship for Vol­un­tary Groups

sarahconstantin11 Dec 2018 22:10 UTC
192 points
10 comments7 min readLW link
(srconstantin.wordpress.com)

Quan­tum im­mor­tal­ity: Is de­cline of mea­sure com­pen­sated by merg­ing timelines?

avturchin11 Dec 2018 19:39 UTC
9 points
8 comments2 min readLW link

Bounded ra­tio­nal­ity abounds in mod­els, not ex­plic­itly defined

Stuart_Armstrong11 Dec 2018 19:34 UTC
14 points
9 comments1 min readLW link

Figur­ing out what Alice wants: non-hu­man Alice

Stuart_Armstrong11 Dec 2018 19:31 UTC
14 points
17 comments2 min readLW link

As­sum­ing we’ve solved X, could we do Y...

Stuart_Armstrong11 Dec 2018 18:13 UTC
31 points
16 comments2 min readLW link

[Question] How Old is Smal­lpox?

Raemon10 Dec 2018 10:50 UTC
44 points
5 comments2 min readLW link

Measly Med­i­ta­tion Measurements

justinpombrio9 Dec 2018 20:54 UTC
62 points
19 comments1 min readLW link

Kin­der­garten in NYC: Much More than You Wanted to Know

Laura B9 Dec 2018 15:36 UTC
36 points
1 comment11 min readLW link

New Rat­fic: Nyssa in the Realm of Possibility

Alicorn9 Dec 2018 5:00 UTC
38 points
0 comments1 min readLW link

[Question] What pre­cisely do we mean by AI al­ign­ment?

Gordon Seidoh Worley9 Dec 2018 2:23 UTC
29 points
8 comments1 min readLW link

[Question] What is “So­cial Real­ity?”

Raemon8 Dec 2018 17:41 UTC
38 points
17 comments1 min readLW link

Pre­dic­tion Mar­kets Are About Be­ing Right

Zvi8 Dec 2018 14:00 UTC
83 points
7 comments7 min readLW link
(thezvi.wordpress.com)

[Question] Why should I care about ra­tio­nal­ity?

TurnTrout8 Dec 2018 3:49 UTC
24 points
5 comments1 min readLW link

Book re­view: Ar­tifi­cial In­tel­li­gence Safety and Security

PeterMcCluskey8 Dec 2018 3:47 UTC
27 points
3 comments8 min readLW link
(www.bayesianinvestor.com)

[Question] Is the hu­man brain a valid choice for the Univer­sal Tur­ing Ma­chine in Solomonoff In­duc­tion?

habryka8 Dec 2018 1:49 UTC
22 points
13 comments1 min readLW link

Tran­shu­man­ists Don’t Need Spe­cial Dispositions

Eliezer Yudkowsky7 Dec 2018 22:24 UTC
96 points
18 comments5 min readLW link

Is cog­ni­tive load a fac­tor in com­mu­nity de­cline?

ryan_b7 Dec 2018 15:45 UTC
18 points
6 comments1 min readLW link

COEDT Equil­ibria in Games

Diffractor6 Dec 2018 18:00 UTC
15 points
0 comments3 min readLW link

Tran­shu­man­ism as Sim­plified Humanism

Eliezer Yudkowsky5 Dec 2018 20:12 UTC
170 points
34 comments5 min readLW link

Why we need a *the­ory* of hu­man values

Stuart_Armstrong5 Dec 2018 16:00 UTC
66 points
15 comments4 min readLW link

Fac­tored Cognition

stuhlmueller5 Dec 2018 1:01 UTC
45 points
6 comments17 min readLW link

Play­ing Politics

sarahconstantin5 Dec 2018 0:30 UTC
97 points
45 comments12 min readLW link
(srconstantin.wordpress.com)

Align­ment Newslet­ter #35

Rohin Shah4 Dec 2018 1:10 UTC
15 points
0 comments6 min readLW link
(mailchi.mp)

Book Re­view—Other Minds: The Oc­to­pus, the Sea, and the Deep Ori­gins of Consciousness

Vaughn Papenhausen3 Dec 2018 8:00 UTC
34 points
18 comments16 min readLW link

Co­her­ence ar­gu­ments do not en­tail goal-di­rected behavior

Rohin Shah3 Dec 2018 3:26 UTC
123 points
69 comments7 min readLW link3 reviews

Benign model-free RL

paulfchristiano2 Dec 2018 4:10 UTC
15 points
1 comment7 min readLW link

Novem­ber 2018 gw­ern.net newsletter

gwern1 Dec 2018 13:57 UTC
35 points
0 comments1 min readLW link
(www.gwern.net)

In­tu­itions about goal-di­rected behavior

Rohin Shah1 Dec 2018 4:25 UTC
54 points
15 comments6 min readLW link

Tak­ing vi­tamin D3 with K2 in the morning

ChristianKl30 Nov 2018 19:00 UTC
53 points
47 comments4 min readLW link

Over­con­fi­dent talk­ing down, hum­ble or hos­tile talk­ing up

ozziegooen30 Nov 2018 12:41 UTC
49 points
19 comments3 min readLW link

Iter­ated Distil­la­tion and Amplification

Ajeya Cotra30 Nov 2018 4:47 UTC
47 points
14 comments6 min readLW link

Mov­ing Factward

Tyrrell_McAllister29 Nov 2018 5:54 UTC
14 points
11 comments1 min readLW link

For­mal Open Prob­lem in De­ci­sion Theory

Scott Garrabrant29 Nov 2018 3:25 UTC
36 points
28 comments4 min readLW link

Reflec­tive or­a­cles as a solu­tion to the con­verse Law­vere problem

SamEisenstat29 Nov 2018 3:23 UTC
31 points
0 comments7 min readLW link

The Ubiquitous Con­verse Law­vere Problem

Scott Garrabrant29 Nov 2018 3:16 UTC
21 points
0 comments2 min readLW link

Hyper­real Brouwer

Scott Garrabrant29 Nov 2018 3:15 UTC
30 points
2 comments6 min readLW link

Coun­ter­in­tu­itive Com­par­a­tive Advantage

Wei Dai28 Nov 2018 20:33 UTC
84 points
8 comments2 min readLW link

Sta­bi­lize-Reflect-Execute

ozziegooen28 Nov 2018 17:26 UTC
30 points
1 comment2 min readLW link