Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
Archive
Sequences
About
Search
Log In
All
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
2025
All
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
All
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
Page
1
[Simulators seminar sequence] #1 Background & shared assumptions
Jan
,
Charlie Steiner
,
Logan Riggs
,
janus
,
jacquesthibs
,
metasemi
,
Michael Oesterle
,
Lucas Teixeira
,
peligrietzer
and
remember
Jan 2, 2023, 11:48 PM
50
points
4
comments
3
min read
LW
link
Linear Algebra Done Right, Axler
David Udell
Jan 2, 2023, 10:54 PM
57
points
6
comments
9
min read
LW
link
MacArthur BART (Filk)
Gordon Seidoh Worley
Jan 2, 2023, 10:50 PM
10
points
1
comment
1
min read
LW
link
Knottiness
abramdemski
Jan 2, 2023, 10:13 PM
43
points
4
comments
2
min read
LW
link
[Question]
Default Sort for Shortforms is Very Bad; How Do I Change It?
DragonGod
Jan 2, 2023, 9:50 PM
15
points
0
comments
1
min read
LW
link
MAKE IT BETTER (a poetic demonstration of the banality of GPT-3)
rogersbacon
Jan 2, 2023, 8:47 PM
7
points
2
comments
5
min read
LW
link
Review of “Make People Better”
Metacelsus
Jan 2, 2023, 8:30 PM
10
points
0
comments
3
min read
LW
link
(denovo.substack.com)
Preparing for Less Privacy
jefftk
Jan 2, 2023, 8:30 PM
23
points
1
comment
2
min read
LW
link
(www.jefftk.com)
Large language models can provide “normative assumptions” for learning human preferences
Stuart_Armstrong
Jan 2, 2023, 7:39 PM
29
points
12
comments
3
min read
LW
link
On the Importance of Open Sourcing Reward Models
elandgre
Jan 2, 2023, 7:01 PM
18
points
5
comments
6
min read
LW
link
Prediction Markets for Science
Vaniver
Jan 2, 2023, 5:55 PM
27
points
7
comments
5
min read
LW
link
Why don’t Rationalists use bidets?
Lakin
Jan 2, 2023, 5:42 PM
31
points
33
comments
2
min read
LW
link
Soft optimization makes the value target bigger
Jeremy Gillen
Jan 2, 2023, 4:06 PM
119
points
20
comments
12
min read
LW
link
Results from the AI testing hackathon
Esben Kran
Jan 2, 2023, 3:46 PM
13
points
0
comments
LW
link
Induction heads—illustrated
CallumMcDougall
Jan 2, 2023, 3:35 PM
128
points
12
comments
3
min read
LW
link
Opportunity Cost Blackmail
adamShimi
Jan 2, 2023, 1:48 PM
70
points
11
comments
2
min read
LW
link
(epistemologicalvigilance.substack.com)
The ultimate limits of alignment will determine the shape of the long term future
beren
Jan 2, 2023, 12:47 PM
34
points
2
comments
6
min read
LW
link
A kernel of Lie theory
Alok Singh
Jan 2, 2023, 9:20 AM
−1
points
8
comments
1
min read
LW
link
(alok.github.io)
Belief Bias: Bias in Evaluating AGI X-Risks
Remmelt
and
flandry19
Jan 2, 2023, 8:59 AM
−10
points
1
comment
1
min read
LW
link
Pacing: inexplicably good
KatjaGrace
Jan 2, 2023, 8:30 AM
39
points
7
comments
1
min read
LW
link
(worldspiritsockpuppet.com)
Alignment, Anger, and Love: Preparing for the Emergence of Superintelligent AI
tavurth
Jan 2, 2023, 6:16 AM
2
points
3
comments
1
min read
LW
link
[Question]
How can total world index fund growth outpace money supply growth over the long term?
pando
Jan 2, 2023, 5:33 AM
4
points
7
comments
1
min read
LW
link
My first year in AI alignment
Alex_Altair
Jan 2, 2023, 1:28 AM
61
points
10
comments
7
min read
LW
link
Sail Over Mountains of ICE...
AnthonyRepetto
Jan 2, 2023, 12:27 AM
26
points
51
comments
7
min read
LW
link
Fun math facts about 2023
Adam Scherlis
Jan 1, 2023, 11:38 PM
9
points
6
comments
1
min read
LW
link
The Thingness of Things
TsviBT
Jan 1, 2023, 10:19 PM
51
points
35
comments
10
min read
LW
link
Thoughts On Expanding the AI Safety Community: Benefits and Challenges of Outreach to Non-Technical Professionals
Yashvardhan Sharma
Jan 1, 2023, 7:21 PM
4
points
4
comments
7
min read
LW
link
[Question]
Would it be good or bad for the US military to get involved in AI risk?
Grant Demaree
Jan 1, 2023, 7:02 PM
50
points
12
comments
1
min read
LW
link
Better New Year’s Goals through Aligning the Elephant and the Rider
moridinamael
Jan 1, 2023, 5:54 PM
20
points
0
comments
2
min read
LW
link
(guildoftherose.org)
A Löbian argument pattern for implicit reasoning in natural language: Löbian party invitations
Andrew_Critch
Jan 1, 2023, 5:39 PM
23
points
8
comments
7
min read
LW
link
woke offline, anti-woke online
Yair Halberstadt
Jan 1, 2023, 8:24 AM
13
points
12
comments
1
min read
LW
link
Summary of 80k’s AI problem profile
JakubK
Jan 1, 2023, 7:30 AM
7
points
0
comments
5
min read
LW
link
(forum.effectivealtruism.org)
What percent of people work in moral mazes?
Raemon
Jan 1, 2023, 4:33 AM
21
points
9
comments
4
min read
LW
link
Recursive Middle Manager Hell
Raemon
Jan 1, 2023, 4:33 AM
224
points
46
comments
11
min read
LW
link
1
review
Challenge to the notion that anything is (maybe) possible with AGI
Remmelt
and
flandry19
Jan 1, 2023, 3:57 AM
−27
points
4
comments
1
min read
LW
link
(mflb.com)
The Roots of Progress’s 2022 in review
jasoncrawford
Jan 1, 2023, 2:54 AM
14
points
2
comments
15
min read
LW
link
(rootsofprogress.org)
Investing for a World Transformed by AI
PeterMcCluskey
Jan 1, 2023, 2:47 AM
70
points
24
comments
6
min read
LW
link
1
review
(bayesianinvestor.com)
Why Free Will is NOT an illusion
Akira Pyinya
Jan 1, 2023, 2:29 AM
0
points
16
comments
1
min read
LW
link
Localhost Security Messaging
jefftk
Jan 1, 2023, 2:20 AM
7
points
3
comments
1
min read
LW
link
(www.jefftk.com)
0 and 1 aren’t probabilities
Alok Singh
Jan 1, 2023, 12:09 AM
2
points
4
comments
2
min read
LW
link
(en.wikipedia.org)
‘simulator’ framing and confusions about LLMs
Beth Barnes
Dec 31, 2022, 11:38 PM
104
points
11
comments
4
min read
LW
link
Monitoring devices I have loved
Elizabeth
Dec 31, 2022, 10:51 PM
62
points
13
comments
3
min read
LW
link
1
review
Slack matters more than any outcome
Valentine
Dec 31, 2022, 8:11 PM
164
points
56
comments
19
min read
LW
link
1
review
To Be Particular About Morality
AGO
Dec 31, 2022, 7:58 PM
6
points
2
comments
7
min read
LW
link
200 COP in MI: Interpreting Algorithmic Problems
Neel Nanda
Dec 31, 2022, 7:55 PM
33
points
2
comments
10
min read
LW
link
The Feeling of Idea Scarcity
johnswentworth
Dec 31, 2022, 5:34 PM
249
points
23
comments
5
min read
LW
link
1
review
Curse of knowledge and Naive realism: Bias in Evaluating AGI X-Risks
Remmelt
and
flandry19
Dec 31, 2022, 1:33 PM
−7
points
1
comment
1
min read
LW
link
(www.lesswrong.com)
[Question]
What career advice do you give to software engineers?
Antb
Dec 31, 2022, 12:01 PM
15
points
4
comments
1
min read
LW
link
[Question]
Are Mixture-of-Experts Transformers More Interpretable Than Dense Transformers?
simeon_c
Dec 31, 2022, 11:34 AM
8
points
5
comments
1
min read
LW
link
[Question]
In which cases can ChatGPT be used as an aid for thesis or scientific paper writing?
Bob Guran
Dec 31, 2022, 10:50 AM
1
point
1
comment
1
min read
LW
link
Back to top
Next
N
W
F
A
C
D
E
F
G
H
I
Customize appearance
Current theme:
default
A
C
D
E
F
G
H
I
Less Wrong (text)
Less Wrong (link)
Invert colors
Reset to defaults
OK
Cancel
Hi, I’m Bobby the Basilisk! Click on the minimize button (
) to minimize the theme tweaker window, so that you can see what the page looks like with the current tweaked values. (But remember,
the changes won’t be saved until you click “OK”!
)
Theme tweaker help
Show Bobby the Basilisk
OK
Cancel