Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
Archive
Sequences
About
Search
Log In
All
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
2025
All
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
All
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
Page
2
An ML interpretation of Shard Theory
beren
Jan 3, 2023, 8:30 PM
39
points
5
comments
4
min read
LW
link
Talking to God
abramdemski
Jan 3, 2023, 8:14 PM
30
points
7
comments
2
min read
LW
link
My Advice for Incoming SERI MATS Scholars
Johannes C. Mayer
Jan 3, 2023, 7:25 PM
58
points
6
comments
4
min read
LW
link
Touch reality as soon as possible (when doing machine learning research)
LawrenceC
Jan 3, 2023, 7:11 PM
117
points
9
comments
8
min read
LW
link
1
review
Kolb’s: an approach to consciously get better at anything
jacquesthibs
Jan 3, 2023, 6:16 PM
12
points
1
comment
6
min read
LW
link
[Question]
{M|Im|Am}oral Mazes—any large-scale counterexamples?
Dagon
Jan 3, 2023, 4:43 PM
24
points
4
comments
1
min read
LW
link
Effectively self-studying over the Internet
libai
Jan 3, 2023, 4:23 PM
11
points
1
comment
4
min read
LW
link
Set-like mathematics in type theory
Thomas Kehrenberg
Jan 3, 2023, 2:33 PM
5
points
1
comment
13
min read
LW
link
Monthly Roundup #2
Zvi
Jan 3, 2023, 12:50 PM
23
points
3
comments
23
min read
LW
link
(thezvi.wordpress.com)
Whisper’s Wild Implications
Ollie J
Jan 3, 2023, 12:17 PM
19
points
6
comments
5
min read
LW
link
How to eat potato chips while typing
KatjaGrace
Jan 3, 2023, 11:50 AM
45
points
12
comments
1
min read
LW
link
(worldspiritsockpuppet.com)
[Question]
I have thousands of copies of HPMOR in Russian. How to use them with the most impact?
Mikhail Samin
Jan 3, 2023, 10:21 AM
26
points
3
comments
1
min read
LW
link
Is recursive self-alignment possible?
No77e
Jan 3, 2023, 9:15 AM
5
points
5
comments
1
min read
LW
link
On the naturalistic study of the linguistic behavior of artificial intelligence
Bill Benzon
Jan 3, 2023, 9:06 AM
1
point
0
comments
4
min read
LW
link
SF Severe Weather Warning
stavros
Jan 3, 2023, 6:04 AM
3
points
3
comments
1
min read
LW
link
(news.ycombinator.com)
Status quo bias; System justification: Bias in Evaluating AGI X-Risks
Remmelt
and
flandry19
Jan 3, 2023, 2:50 AM
−11
points
0
comments
1
min read
LW
link
200 COP in MI: Exploring Polysemanticity and Superposition
Neel Nanda
Jan 3, 2023, 1:52 AM
34
points
6
comments
16
min read
LW
link
The need for speed in web frameworks?
Adam Zerner
Jan 3, 2023, 12:06 AM
19
points
2
comments
8
min read
LW
link
[Simulators seminar sequence] #1 Background & shared assumptions
Jan
,
Charlie Steiner
,
Logan Riggs
,
janus
,
jacquesthibs
,
metasemi
,
Michael Oesterle
,
Lucas Teixeira
,
peligrietzer
and
remember
Jan 2, 2023, 11:48 PM
50
points
4
comments
3
min read
LW
link
Linear Algebra Done Right, Axler
David Udell
Jan 2, 2023, 10:54 PM
57
points
6
comments
9
min read
LW
link
MacArthur BART (Filk)
Gordon Seidoh Worley
Jan 2, 2023, 10:50 PM
10
points
1
comment
1
min read
LW
link
Knottiness
abramdemski
Jan 2, 2023, 10:13 PM
43
points
4
comments
2
min read
LW
link
[Question]
Default Sort for Shortforms is Very Bad; How Do I Change It?
DragonGod
Jan 2, 2023, 9:50 PM
15
points
0
comments
1
min read
LW
link
MAKE IT BETTER (a poetic demonstration of the banality of GPT-3)
rogersbacon
Jan 2, 2023, 8:47 PM
7
points
2
comments
5
min read
LW
link
Review of “Make People Better”
Metacelsus
Jan 2, 2023, 8:30 PM
10
points
0
comments
3
min read
LW
link
(denovo.substack.com)
Preparing for Less Privacy
jefftk
Jan 2, 2023, 8:30 PM
23
points
1
comment
2
min read
LW
link
(www.jefftk.com)
Large language models can provide “normative assumptions” for learning human preferences
Stuart_Armstrong
Jan 2, 2023, 7:39 PM
29
points
12
comments
3
min read
LW
link
On the Importance of Open Sourcing Reward Models
elandgre
Jan 2, 2023, 7:01 PM
18
points
5
comments
6
min read
LW
link
Prediction Markets for Science
Vaniver
Jan 2, 2023, 5:55 PM
27
points
7
comments
5
min read
LW
link
Why don’t Rationalists use bidets?
Lakin
Jan 2, 2023, 5:42 PM
31
points
33
comments
2
min read
LW
link
Soft optimization makes the value target bigger
Jeremy Gillen
Jan 2, 2023, 4:06 PM
119
points
20
comments
12
min read
LW
link
Results from the AI testing hackathon
Esben Kran
Jan 2, 2023, 3:46 PM
13
points
0
comments
LW
link
Induction heads—illustrated
CallumMcDougall
Jan 2, 2023, 3:35 PM
130
points
12
comments
3
min read
LW
link
Opportunity Cost Blackmail
adamShimi
Jan 2, 2023, 1:48 PM
70
points
11
comments
2
min read
LW
link
(epistemologicalvigilance.substack.com)
The ultimate limits of alignment will determine the shape of the long term future
beren
Jan 2, 2023, 12:47 PM
34
points
2
comments
6
min read
LW
link
A kernel of Lie theory
Alok Singh
Jan 2, 2023, 9:20 AM
−1
points
8
comments
1
min read
LW
link
(alok.github.io)
Belief Bias: Bias in Evaluating AGI X-Risks
Remmelt
and
flandry19
Jan 2, 2023, 8:59 AM
−10
points
1
comment
1
min read
LW
link
Pacing: inexplicably good
KatjaGrace
Jan 2, 2023, 8:30 AM
39
points
7
comments
1
min read
LW
link
(worldspiritsockpuppet.com)
Alignment, Anger, and Love: Preparing for the Emergence of Superintelligent AI
tavurth
Jan 2, 2023, 6:16 AM
2
points
3
comments
1
min read
LW
link
[Question]
How can total world index fund growth outpace money supply growth over the long term?
pando
Jan 2, 2023, 5:33 AM
4
points
7
comments
1
min read
LW
link
My first year in AI alignment
Alex_Altair
Jan 2, 2023, 1:28 AM
61
points
10
comments
7
min read
LW
link
Sail Over Mountains of ICE...
AnthonyRepetto
Jan 2, 2023, 12:27 AM
26
points
51
comments
7
min read
LW
link
Fun math facts about 2023
Adam Scherlis
Jan 1, 2023, 11:38 PM
9
points
6
comments
1
min read
LW
link
The Thingness of Things
TsviBT
1 Jan 2023 22:19 UTC
51
points
35
comments
10
min read
LW
link
Thoughts On Expanding the AI Safety Community: Benefits and Challenges of Outreach to Non-Technical Professionals
Yashvardhan Sharma
1 Jan 2023 19:21 UTC
4
points
4
comments
7
min read
LW
link
[Question]
Would it be good or bad for the US military to get involved in AI risk?
Grant Demaree
1 Jan 2023 19:02 UTC
50
points
12
comments
1
min read
LW
link
Better New Year’s Goals through Aligning the Elephant and the Rider
moridinamael
1 Jan 2023 17:54 UTC
20
points
0
comments
2
min read
LW
link
(guildoftherose.org)
A Löbian argument pattern for implicit reasoning in natural language: Löbian party invitations
Andrew_Critch
1 Jan 2023 17:39 UTC
23
points
8
comments
7
min read
LW
link
woke offline, anti-woke online
Yair Halberstadt
1 Jan 2023 8:24 UTC
13
points
12
comments
1
min read
LW
link
Summary of 80k’s AI problem profile
JakubK
1 Jan 2023 7:30 UTC
7
points
0
comments
5
min read
LW
link
(forum.effectivealtruism.org)
Previous
Back to top
Next
N
W
F
A
C
D
E
F
G
H
I
Customize appearance
Current theme:
default
A
C
D
E
F
G
H
I
Less Wrong (text)
Less Wrong (link)
Invert colors
Reset to defaults
OK
Cancel
Hi, I’m Bobby the Basilisk! Click on the minimize button (
) to minimize the theme tweaker window, so that you can see what the page looks like with the current tweaked values. (But remember,
the changes won’t be saved until you click “OK”!
)
Theme tweaker help
Show Bobby the Basilisk
OK
Cancel