Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
Archive
Sequences
About
Search
Log In
All
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
2025
All
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
All
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
Page
1
[Question]
Convince me that humanity is as doomed by AGI as Yudkowsky et al., seems to believe
Yitz
Apr 10, 2022, 9:02 PM
92
points
141
comments
2
min read
LW
link
Emotionally Confronting a Probably-Doomed World: Against Motivation Via Dignity Points
TurnTrout
Apr 10, 2022, 6:45 PM
154
points
7
comments
9
min read
LW
link
[Question]
Does non-access to outputs prevent recursive self-improvement?
Gunnar_Zarncke
Apr 10, 2022, 6:37 PM
15
points
0
comments
1
min read
LW
link
A Brief Excursion Into Molecular Neuroscience
Jan
Apr 10, 2022, 5:55 PM
48
points
8
comments
19
min read
LW
link
(universalprior.substack.com)
Finally Entering Alignment
Ulisse Mini
Apr 10, 2022, 5:01 PM
80
points
8
comments
2
min read
LW
link
Schelling Meetup Toronto
Sean Aubin
Apr 10, 2022, 1:58 PM
3
points
0
comments
1
min read
LW
link
Is Fisherian Runaway Gradient Hacking?
Ryan Kidd
Apr 10, 2022, 1:47 PM
15
points
6
comments
4
min read
LW
link
Worse than an unaligned AGI
Shmi
Apr 10, 2022, 3:35 AM
−1
points
11
comments
1
min read
LW
link
Time-Time Tradeoffs
Orpheus16
Apr 10, 2022, 2:33 AM
18
points
1
comment
3
min read
LW
link
(forum.effectivealtruism.org)
Boston Contra: Fully Gender-Free
jefftk
Apr 10, 2022, 12:40 AM
3
points
12
comments
1
min read
LW
link
(www.jefftk.com)
[Question]
Hidden comments settings not working?
TLW
Apr 9, 2022, 11:15 PM
4
points
2
comments
1
min read
LW
link
Godshatter Versus Legibility: A Fundamentally Different Approach To AI Alignment
LukeOnline
Apr 9, 2022, 9:43 PM
15
points
14
comments
7
min read
LW
link
A concrete bet offer to those with short AGI timelines
Matthew Barnett
and
Tamay
Apr 9, 2022, 9:41 PM
199
points
120
comments
5
min read
LW
link
New: use The Nonlinear Library to listen to the top LessWrong posts of all time
KatWoods
Apr 9, 2022, 8:50 PM
39
points
9
comments
8
min read
LW
link
140 Cognitive Biases You Should Know
André Ferretti
Apr 9, 2022, 5:15 PM
8
points
7
comments
1
min read
LW
link
Strategies for keeping AIs narrow in the short term
Rossin
Apr 9, 2022, 4:42 PM
9
points
3
comments
3
min read
LW
link
Hyperbolic takeoff
Ege Erdil
Apr 9, 2022, 3:57 PM
18
points
7
comments
10
min read
LW
link
(www.metaculus.com)
Elicit: Language Models as Research Assistants
stuhlmueller
and
jungofthewon
Apr 9, 2022, 2:56 PM
71
points
6
comments
13
min read
LW
link
Emergent Ventures/Schmidt (new grantor for individual researchers)
gwern
Apr 9, 2022, 2:41 PM
21
points
6
comments
1
min read
LW
link
(marginalrevolution.com)
AI safety: the ultimate trolley problem
chaosmage
Apr 9, 2022, 12:05 PM
−21
points
6
comments
1
min read
LW
link
AMA Conjecture, A New Alignment Startup
adamShimi
Apr 9, 2022, 9:43 AM
47
points
42
comments
1
min read
LW
link
[Question]
What advice do you have for someone struggling to detach their grim-o-meter?
Zorger74
Apr 9, 2022, 7:35 AM
6
points
3
comments
1
min read
LW
link
[Question]
Can AI systems have extremely impressive outputs and also not need to be aligned because they aren’t general enough or something?
WilliamKiely
Apr 9, 2022, 6:03 AM
6
points
3
comments
1
min read
LW
link
Buy-in Before Randomization
jefftk
Apr 9, 2022, 1:30 AM
26
points
9
comments
1
min read
LW
link
(www.jefftk.com)
Why Instrumental Goals are not a big AI Safety Problem
Jonathan Paulson
Apr 9, 2022, 12:10 AM
0
points
7
comments
3
min read
LW
link
A method of writing content easily with little anxiety
jessicata
Apr 8, 2022, 10:11 PM
64
points
19
comments
3
min read
LW
link
(unstableontology.com)
Good Heart Donation Lottery Winner
Gordon Seidoh Worley
Apr 8, 2022, 8:34 PM
21
points
0
comments
1
min read
LW
link
Roam Research Mobile is Out!
Logan Riggs
Apr 8, 2022, 7:05 PM
12
points
0
comments
1
min read
LW
link
Progress Report 4: logit lens redux
Nathan Helm-Burger
Apr 8, 2022, 6:35 PM
4
points
0
comments
2
min read
LW
link
[Question]
What would the creation of aligned AGI look like for us?
Perhaps
Apr 8, 2022, 6:05 PM
3
points
4
comments
1
min read
LW
link
Convincing All Capability Researchers
Logan Riggs
Apr 8, 2022, 5:40 PM
120
points
70
comments
3
min read
LW
link
Language Model Tools for Alignment Research
Logan Riggs
Apr 8, 2022, 5:32 PM
28
points
0
comments
2
min read
LW
link
Takeaways From 3 Years Working In Machine Learning
George3d6
Apr 8, 2022, 5:14 PM
35
points
10
comments
11
min read
LW
link
(www.epistem.ink)
[RETRACTED] It’s time for EA leadership to pull the short-timelines fire alarm.
Not Relevant
Apr 8, 2022, 4:07 PM
115
points
166
comments
4
min read
LW
link
Boulder ACX Meetup, Sun Apr 24
Josh Sacks
Apr 8, 2022, 3:43 PM
5
points
4
comments
1
min read
LW
link
AIs should learn human preferences, not biases
Stuart_Armstrong
Apr 8, 2022, 1:45 PM
10
points
0
comments
1
min read
LW
link
We Are Conjecture, A New Alignment Research Startup
Connor Leahy
Apr 8, 2022, 11:40 AM
197
points
25
comments
4
min read
LW
link
Different perspectives on concept extrapolation
Stuart_Armstrong
Apr 8, 2022, 10:42 AM
48
points
8
comments
5
min read
LW
link
1
review
[Question]
Is there a possibility that the upcoming scaling of data in language models causes A.G.I.?
ArtMi
Apr 8, 2022, 6:56 AM
2
points
0
comments
1
min read
LW
link
Good Heart Week Is Over!
Ben Pace
Apr 8, 2022, 6:43 AM
55
points
35
comments
1
min read
LW
link
The Rationalist-Etcetera Diaspora: A SPREADSHEET!!
Amelia Bedelia
Apr 8, 2022, 5:43 AM
25
points
2
comments
1
min read
LW
link
AI Alignment and Recognition
Chris_Leong
Apr 8, 2022, 5:39 AM
7
points
2
comments
1
min read
LW
link
Nature’s answer to the explore/exploit problem
lizard_brain
Apr 8, 2022, 5:13 AM
5
points
1
comment
1
min read
LW
link
Edge cases don’t invalidate the rule
Adam Selker
Apr 8, 2022, 4:17 AM
6
points
5
comments
2
min read
LW
link
Reverse (intent) alignment may allow for safer Oracles
azsantosk
Apr 8, 2022, 2:48 AM
4
points
0
comments
4
min read
LW
link
Summary: “Internet Search tips” by Gwern Branwen
Pablo Repetto
Apr 8, 2022, 2:02 AM
12
points
2
comments
4
min read
LW
link
(pabloernesto.github.io)
Maxwell Peterson’s Highlighted Posts
Maxwell Peterson
Apr 8, 2022, 1:34 AM
5
points
0
comments
1
min read
LW
link
Foot-Chording Chords
jefftk
8 Apr 2022 1:10 UTC
8
points
0
comments
1
min read
LW
link
(www.jefftk.com)
DeepMind: The Podcast—Excerpts on AGI
WilliamKiely
7 Apr 2022 22:09 UTC
99
points
12
comments
5
min read
LW
link
Convincing Your Brain That Humanity is Evil is Easy
Johannes C. Mayer
7 Apr 2022 21:39 UTC
14
points
4
comments
2
min read
LW
link
Back to top
Next
N
W
F
A
C
D
E
F
G
H
I
Customize appearance
Current theme:
default
A
C
D
E
F
G
H
I
Less Wrong (text)
Less Wrong (link)
Invert colors
Reset to defaults
OK
Cancel
Hi, I’m Bobby the Basilisk! Click on the minimize button (
) to minimize the theme tweaker window, so that you can see what the page looks like with the current tweaked values. (But remember,
the changes won’t be saved until you click “OK”!
)
Theme tweaker help
Show Bobby the Basilisk
OK
Cancel