Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
Archive
Sequences
About
Search
Log In
All
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
2025
All
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
All
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
Page
1
More power to you
jasoncrawford
Dec 15, 2021, 11:50 PM
16
points
14
comments
1
min read
LW
link
(rootsofprogress.org)
My Overview of the AI Alignment Landscape: A Bird’s Eye View
Neel Nanda
Dec 15, 2021, 11:44 PM
127
points
9
comments
15
min read
LW
link
SmartPoop 1.0: An AI Safety Science-Fiction
Lê Nguyên Hoang
Dec 15, 2021, 10:28 PM
7
points
1
comment
1
min read
LW
link
Bay Area Rationalist Field Day
Raj Thimmiah
Dec 15, 2021, 7:57 PM
7
points
1
comment
1
min read
LW
link
Framing approaches to alignment and the hard problem of AI cognition
ryan_greenblatt
Dec 15, 2021, 7:06 PM
16
points
15
comments
27
min read
LW
link
South Bay ACX/LW Pre-Holiday Get-Together
IS
Dec 15, 2021, 4:58 PM
5
points
0
comments
1
min read
LW
link
Leverage
lsusr
Dec 15, 2021, 5:20 AM
23
points
2
comments
1
min read
LW
link
We’ll Always Have Crazy
Duncan Sabien (Inactive)
Dec 15, 2021, 2:55 AM
36
points
22
comments
13
min read
LW
link
2020 Review: The Discussion Phase
Vaniver
Dec 15, 2021, 1:12 AM
55
points
14
comments
2
min read
LW
link
The Natural Abstraction Hypothesis: Implications and Evidence
CallumMcDougall
Dec 14, 2021, 11:14 PM
39
points
9
comments
19
min read
LW
link
Robin Hanson’s “Humans are Early”
Raemon
Dec 14, 2021, 10:07 PM
11
points
0
comments
2
min read
LW
link
(www.overcomingbias.com)
Ngo’s view on alignment difficulty
Richard_Ngo
and
Eliezer Yudkowsky
Dec 14, 2021, 9:34 PM
63
points
7
comments
17
min read
LW
link
A proposed system for ideas jumpstart
Valentin2026
Dec 14, 2021, 9:01 PM
4
points
2
comments
3
min read
LW
link
Should we rely on the speed prior for safety?
Marc Carauleanu
Dec 14, 2021, 8:45 PM
14
points
5
comments
5
min read
LW
link
ARC’s first technical report: Eliciting Latent Knowledge
paulfchristiano
,
Mark Xu
and
Ajeya Cotra
Dec 14, 2021, 8:09 PM
228
points
90
comments
1
min read
LW
link
3
reviews
(docs.google.com)
ARC is hiring!
paulfchristiano
and
Mark Xu
Dec 14, 2021, 8:09 PM
64
points
2
comments
1
min read
LW
link
Interlude: Agents as Automobiles
Daniel Kokotajlo
Dec 14, 2021, 6:49 PM
26
points
6
comments
5
min read
LW
link
Zvi’s Thoughts on the Survival and Flourishing Fund (SFF)
Zvi
Dec 14, 2021, 2:30 PM
193
points
65
comments
64
min read
LW
link
1
review
(thezvi.wordpress.com)
Consequentialism & corrigibility
Steven Byrnes
Dec 14, 2021, 1:23 PM
70
points
35
comments
7
min read
LW
link
Mystery Hunt 2022
Scott Garrabrant
Dec 13, 2021, 9:57 PM
30
points
5
comments
1
min read
LW
link
Enabling More Feedback for AI Safety Researchers
frances_lorenz
Dec 13, 2021, 8:10 PM
17
points
0
comments
3
min read
LW
link
Language Model Alignment Research Internships
Ethan Perez
Dec 13, 2021, 7:53 PM
74
points
1
comment
1
min read
LW
link
Omicron Post #6
Zvi
Dec 13, 2021, 6:00 PM
89
points
30
comments
8
min read
LW
link
(thezvi.wordpress.com)
Analysis of Bird Box (2018)
TekhneMakre
Dec 13, 2021, 5:30 PM
11
points
3
comments
5
min read
LW
link
Solving Interpretability Week
Logan Riggs
Dec 13, 2021, 5:09 PM
11
points
5
comments
1
min read
LW
link
Understanding and controlling auto-induced distributional shift
L Rudolf L
Dec 13, 2021, 2:59 PM
33
points
4
comments
16
min read
LW
link
A fate worse than death?
RomanS
Dec 13, 2021, 11:05 AM
−25
points
26
comments
2
min read
LW
link
What’s the backward-forward FLOP ratio for Neural Networks?
Marius Hobbhahn
and
Jsevillamol
Dec 13, 2021, 8:54 AM
20
points
12
comments
10
min read
LW
link
Summary of the Acausal Attack Issue for AIXI
Diffractor
Dec 13, 2021, 8:16 AM
12
points
6
comments
4
min read
LW
link
Hard-Coding Neural Computation
MadHatter
Dec 13, 2021, 4:35 AM
34
points
8
comments
27
min read
LW
link
[Question]
Is “gears-level” just a synonym for “mechanistic”?
David Scott Krueger (formerly: capybaralet)
Dec 13, 2021, 4:11 AM
48
points
29
comments
1
min read
LW
link
Baby Nicknames
jefftk
Dec 13, 2021, 2:20 AM
11
points
0
comments
1
min read
LW
link
(www.jefftk.com)
[Question]
Why do governments refer to existential risks primarily in terms of national security?
Evan_Gaensbauer
Dec 13, 2021, 1:05 AM
3
points
3
comments
1
min read
LW
link
[Question]
[Resolved] Who else prefers “AI alignment” to “AI safety?”
Evan_Gaensbauer
Dec 13, 2021, 12:35 AM
5
points
8
comments
1
min read
LW
link
Working through D&D.Sci, problem 1
Pablo Repetto
Dec 12, 2021, 11:10 PM
8
points
2
comments
1
min read
LW
link
(pabloernesto.github.io)
Teaser: Hard-coding Transformer Models
MadHatter
Dec 12, 2021, 10:04 PM
74
points
19
comments
1
min read
LW
link
The Three Mutations of Dark Rationality
DarkRationalist
Dec 12, 2021, 10:01 PM
−15
points
0
comments
2
min read
LW
link
Redwood’s Technique-Focused Epistemic Strategy
adamShimi
Dec 12, 2021, 4:36 PM
48
points
1
comment
7
min read
LW
link
For and Against Lotteries in Elite University Admissions
Sam Enright
Dec 12, 2021, 1:41 PM
10
points
2
comments
3
min read
LW
link
[Question]
Nuclear war anthropics
smountjoy
Dec 12, 2021, 4:54 AM
11
points
7
comments
1
min read
LW
link
Some abstract, non-technical reasons to be non-maximally-pessimistic about AI alignment
Rob Bensinger
Dec 12, 2021, 2:08 AM
70
points
35
comments
7
min read
LW
link
Magna Alta Doctrina
jacob_cannell
Dec 11, 2021, 9:54 PM
60
points
7
comments
28
min read
LW
link
EA Dinner Covid Logistics
jefftk
Dec 11, 2021, 9:50 PM
17
points
7
comments
2
min read
LW
link
(www.jefftk.com)
Transforming myopic optimization to ordinary optimization—Do we want to seek convergence for myopic optimization problems?
tailcalled
Dec 11, 2021, 8:38 PM
12
points
1
comment
5
min read
LW
link
What on Earth is a Series I savings bond?
rossry
Dec 11, 2021, 12:18 PM
11
points
7
comments
7
min read
LW
link
D&D.Sci GURPS Dec 2021: Hunters of Monsters
J Bostock
Dec 11, 2021, 12:13 PM
20
points
21
comments
2
min read
LW
link
Anxiety and computer architecture
Adam Zerner
Dec 11, 2021, 10:37 AM
13
points
8
comments
3
min read
LW
link
[Question]
Reasons to act according to the free will paradigm?
Maciej Jałocha
Dec 11, 2021, 8:44 AM
−3
points
5
comments
1
min read
LW
link
Extrinsic and Intrinsic Moral Frameworks
lsusr
Dec 11, 2021, 5:28 AM
14
points
5
comments
2
min read
LW
link
Moore’s Law, AI, and the pace of progress
Veedrac
Dec 11, 2021, 3:02 AM
128
points
38
comments
24
min read
LW
link
Back to top
Next
N
W
F
A
C
D
E
F
G
H
I
Customize appearance
Current theme:
default
A
C
D
E
F
G
H
I
Less Wrong (text)
Less Wrong (link)
Invert colors
Reset to defaults
OK
Cancel
Hi, I’m Bobby the Basilisk! Click on the minimize button (
) to minimize the theme tweaker window, so that you can see what the page looks like with the current tweaked values. (But remember,
the changes won’t be saved until you click “OK”!
)
Theme tweaker help
Show Bobby the Basilisk
OK
Cancel