Wire­head­ing as a Pos­si­ble Con­trib­u­tor to Civ­i­liza­tional Decline

avturchin12 Nov 2018 20:33 UTC
3 points
6 comments1 min readLW link
(forum.effectivealtruism.org)

Align­ment Newslet­ter #32

Rohin Shah12 Nov 2018 17:20 UTC
18 points
0 comments12 min readLW link
(mailchi.mp)

AI de­vel­op­ment in­cen­tive gra­di­ents are not uniformly terrible

rk12 Nov 2018 16:27 UTC
21 points
12 comments6 min readLW link

What is be­ing?

Andrew Bindon12 Nov 2018 15:33 UTC
−14 points
20 comments7 min readLW link

Aligned AI, The Scientist

shminux12 Nov 2018 6:36 UTC
12 points
2 comments1 min readLW link

Com­bat vs Nur­ture: Cul­tural Genesis

Ruby12 Nov 2018 2:11 UTC
35 points
12 comments6 min readLW link

Ra­tion­al­ity Is Not Sys­tem­atized Winning

namespace11 Nov 2018 22:05 UTC
36 points
20 comments1 min readLW link
(www.thelastrationalist.com)

“She Wanted It”

sarahconstantin11 Nov 2018 22:00 UTC
117 points
19 comments7 min readLW link
(srconstantin.wordpress.com)

Fu­ture di­rec­tions for am­bi­tious value learning

Rohin Shah11 Nov 2018 15:53 UTC
48 points
9 comments4 min readLW link

Rec­on­cil­ing Left and Right, from the Bot­tom-Up

jesseduffield11 Nov 2018 8:36 UTC
14 points
2 comments15 min readLW link

Com­pet­i­tive Mar­kets as Distributed Backprop

johnswentworth10 Nov 2018 16:47 UTC
52 points
10 comments4 min readLW link1 review

Pro­duc­tivity: In­stru­men­tal Rationality

frdk66610 Nov 2018 14:58 UTC
6 points
7 comments1 min readLW link

Pre­face to the se­quence on iter­ated amplification

paulfchristiano10 Nov 2018 13:24 UTC
43 points
8 comments3 min readLW link

Speci­fi­ca­tion gam­ing ex­am­ples in AI

Samuel Rødal10 Nov 2018 12:00 UTC
24 points
6 comments1 min readLW link
(docs.google.com)

Real-time hiring with pre­dic­tion markets

ryan_b9 Nov 2018 22:10 UTC
17 points
9 comments1 min readLW link

Cur­rent AI Safety Roles for Soft­ware Engineers

ozziegooen9 Nov 2018 20:57 UTC
70 points
9 comments4 min readLW link

Model Mis-speci­fi­ca­tion and In­verse Re­in­force­ment Learning

9 Nov 2018 15:33 UTC
33 points
3 comments16 min readLW link

Open AI: Can we rule out near-term AGI?

ShardPhoenix9 Nov 2018 12:16 UTC
13 points
1 comment1 min readLW link
(www.youtube.com)

Pre­dic­tion-Aug­mented Eval­u­a­tion Systems

ozziegooen9 Nov 2018 10:55 UTC
44 points
12 comments8 min readLW link

Up­date the best text­books on ev­ery sub­ject list

ryan_b8 Nov 2018 20:54 UTC
92 points
14 comments1 min readLW link

Multi-Agent Overop­ti­miza­tion, and Embed­ded Agent World Models

Davidmanheim8 Nov 2018 20:33 UTC
8 points
3 comments3 min readLW link

Help Me Re­fac­tor My­self I am Lost

Pausecafe8 Nov 2018 20:18 UTC
12 points
20 comments2 min readLW link

Embed­ded Curiosities

8 Nov 2018 14:19 UTC
91 points
1 comment2 min readLW link

On first look­ing into Rus­sell’s History

Richard_Ngo8 Nov 2018 11:20 UTC
23 points
6 comments5 min readLW link
(thinkingcomplete.blogspot.com)

Is Copen­hagen or Many Wor­lds true? An ex­per­i­ment. What? Yes.

Jonathanm8 Nov 2018 10:10 UTC
11 points
3 comments1 min readLW link
(arxiv.org)

The new Effec­tive Altru­ism fo­rum just launched

habryka8 Nov 2018 1:59 UTC
27 points
6 comments1 min readLW link

What are Univer­sal In­duc­tors, Again?

Diffractor7 Nov 2018 22:32 UTC
14 points
0 comments7 min readLW link

Burnout: What it is and how to Treat it.

Elizabeth7 Nov 2018 22:02 UTC
49 points
0 comments1 min readLW link
(forum.effectivealtruism.org)

Bayes Questions

Bucky7 Nov 2018 16:54 UTC
21 points
13 comments2 min readLW link

La­tent Vari­ables and Model Mis-Specification

jsteinhardt7 Nov 2018 14:48 UTC
24 points
8 comments9 min readLW link

Paris SSC Meetup

fbreton7 Nov 2018 10:04 UTC
1 point
0 comments1 min readLW link

Refram­ing a Crush: Distill­ing the “like” out of “like like”

squidious7 Nov 2018 2:50 UTC
6 points
0 comments3 min readLW link
(opalsandbonobos.blogspot.com)

Ra­tion­al­ity of demon­strat­ing & voting

bfinn7 Nov 2018 0:09 UTC
24 points
21 comments8 min readLW link

Tri­an­gle SSC Meetup

willbobaggins6 Nov 2018 22:43 UTC
1 point
0 comments1 min readLW link

The Vuln­er­a­ble World Hy­poth­e­sis (by Bostrom)

Ben Pace6 Nov 2018 20:05 UTC
50 points
17 comments4 min readLW link
(nickbostrom.com)

Sub­sys­tem Alignment

6 Nov 2018 16:16 UTC
99 points
12 comments1 min readLW link

Align­ment Newslet­ter #31

Rohin Shah5 Nov 2018 23:50 UTC
17 points
0 comments12 min readLW link
(mailchi.mp)

Oc­topath Trav­eler: Spoiler-Free Review

Zvi5 Nov 2018 17:50 UTC
11 points
1 comment14 min readLW link
(thezvi.wordpress.com)

Spec­u­la­tions on im­prov­ing debating

Richard_Ngo5 Nov 2018 16:10 UTC
22 points
4 comments4 min readLW link
(thinkingcomplete.blogspot.com)

Boulder Slate Star Codex Meetup

corticalcircuitry5 Nov 2018 15:01 UTC
5 points
0 comments1 min readLW link

Hu­mans can be as­signed any val­ues what­so­ever…

Stuart_Armstrong5 Nov 2018 14:26 UTC
54 points
27 comments4 min readLW link

When does ra­tio­nal­ity-as-search have non­triv­ial im­pli­ca­tions?

nostalgebraist4 Nov 2018 22:42 UTC
66 points
11 comments3 min readLW link

Beliefs at differ­ent timescales

Nisan4 Nov 2018 20:10 UTC
25 points
12 comments2 min readLW link

No Really, Why Aren’t Ra­tion­al­ists Win­ning?

Sailor Vulcan4 Nov 2018 18:11 UTC
35 points
90 comments5 min readLW link

Ro­bust Delegation

4 Nov 2018 16:38 UTC
116 points
10 comments1 min readLW link

New GreaterWrong fea­ture: image zoom + image slideshows

Said Achmiz4 Nov 2018 7:34 UTC
37 points
3 comments1 min readLW link

The easy goal in­fer­ence prob­lem is still hard

paulfchristiano3 Nov 2018 14:41 UTC
59 points
20 comments4 min readLW link

My In­tu­itions on Coun­ter­log­i­cal Reasoning

DragonGod2 Nov 2018 21:36 UTC
4 points
0 comments5 min readLW link

Embed­ded World-Models

2 Nov 2018 16:07 UTC
92 points
16 comments1 min readLW link

Kansas City SSC Read­ers Meetup

Todd3262 Nov 2018 10:45 UTC
2 points
0 comments1 min readLW link