Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
Archive
Sequences
About
Search
Log In
All
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
2025
All
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
All
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
Page
2
The ultimate limits of alignment will determine the shape of the long term future
beren
Jan 2, 2023, 12:47 PM
34
points
2
comments
6
min read
LW
link
A kernel of Lie theory
Alok Singh
Jan 2, 2023, 9:20 AM
−1
points
8
comments
1
min read
LW
link
(alok.github.io)
Belief Bias: Bias in Evaluating AGI X-Risks
Remmelt
and
flandry19
Jan 2, 2023, 8:59 AM
−10
points
1
comment
1
min read
LW
link
Pacing: inexplicably good
KatjaGrace
Jan 2, 2023, 8:30 AM
39
points
7
comments
1
min read
LW
link
(worldspiritsockpuppet.com)
Alignment, Anger, and Love: Preparing for the Emergence of Superintelligent AI
tavurth
Jan 2, 2023, 6:16 AM
2
points
3
comments
1
min read
LW
link
[Question]
How can total world index fund growth outpace money supply growth over the long term?
pando
Jan 2, 2023, 5:33 AM
4
points
7
comments
1
min read
LW
link
My first year in AI alignment
Alex_Altair
Jan 2, 2023, 1:28 AM
61
points
10
comments
7
min read
LW
link
Sail Over Mountains of ICE...
AnthonyRepetto
Jan 2, 2023, 12:27 AM
26
points
51
comments
7
min read
LW
link
Fun math facts about 2023
Adam Scherlis
Jan 1, 2023, 11:38 PM
9
points
6
comments
1
min read
LW
link
The Thingness of Things
TsviBT
Jan 1, 2023, 10:19 PM
51
points
35
comments
10
min read
LW
link
Thoughts On Expanding the AI Safety Community: Benefits and Challenges of Outreach to Non-Technical Professionals
Yashvardhan Sharma
Jan 1, 2023, 7:21 PM
4
points
4
comments
7
min read
LW
link
[Question]
Would it be good or bad for the US military to get involved in AI risk?
Grant Demaree
Jan 1, 2023, 7:02 PM
50
points
12
comments
1
min read
LW
link
Better New Year’s Goals through Aligning the Elephant and the Rider
moridinamael
Jan 1, 2023, 5:54 PM
20
points
0
comments
2
min read
LW
link
(guildoftherose.org)
A Löbian argument pattern for implicit reasoning in natural language: Löbian party invitations
Andrew_Critch
Jan 1, 2023, 5:39 PM
23
points
8
comments
7
min read
LW
link
woke offline, anti-woke online
Yair Halberstadt
Jan 1, 2023, 8:24 AM
13
points
12
comments
1
min read
LW
link
Summary of 80k’s AI problem profile
JakubK
Jan 1, 2023, 7:30 AM
7
points
0
comments
5
min read
LW
link
(forum.effectivealtruism.org)
What percent of people work in moral mazes?
Raemon
Jan 1, 2023, 4:33 AM
21
points
9
comments
4
min read
LW
link
Recursive Middle Manager Hell
Raemon
Jan 1, 2023, 4:33 AM
224
points
46
comments
11
min read
LW
link
1
review
Challenge to the notion that anything is (maybe) possible with AGI
Remmelt
and
flandry19
Jan 1, 2023, 3:57 AM
−27
points
4
comments
1
min read
LW
link
(mflb.com)
The Roots of Progress’s 2022 in review
jasoncrawford
Jan 1, 2023, 2:54 AM
14
points
2
comments
15
min read
LW
link
(rootsofprogress.org)
Investing for a World Transformed by AI
PeterMcCluskey
Jan 1, 2023, 2:47 AM
70
points
24
comments
6
min read
LW
link
1
review
(bayesianinvestor.com)
Why Free Will is NOT an illusion
Akira Pyinya
Jan 1, 2023, 2:29 AM
0
points
16
comments
1
min read
LW
link
Localhost Security Messaging
jefftk
Jan 1, 2023, 2:20 AM
7
points
3
comments
1
min read
LW
link
(www.jefftk.com)
0 and 1 aren’t probabilities
Alok Singh
Jan 1, 2023, 12:09 AM
2
points
4
comments
2
min read
LW
link
(en.wikipedia.org)
‘simulator’ framing and confusions about LLMs
Beth Barnes
Dec 31, 2022, 11:38 PM
104
points
11
comments
4
min read
LW
link
Monitoring devices I have loved
Elizabeth
Dec 31, 2022, 10:51 PM
62
points
13
comments
3
min read
LW
link
1
review
Slack matters more than any outcome
Valentine
Dec 31, 2022, 8:11 PM
164
points
56
comments
19
min read
LW
link
1
review
To Be Particular About Morality
AGO
Dec 31, 2022, 7:58 PM
6
points
2
comments
7
min read
LW
link
200 COP in MI: Interpreting Algorithmic Problems
Neel Nanda
Dec 31, 2022, 7:55 PM
33
points
2
comments
10
min read
LW
link
The Feeling of Idea Scarcity
johnswentworth
Dec 31, 2022, 5:34 PM
251
points
23
comments
5
min read
LW
link
1
review
Curse of knowledge and Naive realism: Bias in Evaluating AGI X-Risks
Remmelt
and
flandry19
Dec 31, 2022, 1:33 PM
−7
points
1
comment
1
min read
LW
link
(www.lesswrong.com)
[Question]
What career advice do you give to software engineers?
Antb
Dec 31, 2022, 12:01 PM
15
points
4
comments
1
min read
LW
link
[Question]
Are Mixture-of-Experts Transformers More Interpretable Than Dense Transformers?
simeon_c
Dec 31, 2022, 11:34 AM
8
points
5
comments
1
min read
LW
link
[Question]
In which cases can ChatGPT be used as an aid for thesis or scientific paper writing?
Bob Guran
Dec 31, 2022, 10:50 AM
1
point
1
comment
1
min read
LW
link
Two Issues with Playing Chicken with the Universe
Chris_Leong
Dec 31, 2022, 6:47 AM
4
points
4
comments
2
min read
LW
link
Extreme risk neutrality isn’t always wrong
Grant Demaree
Dec 31, 2022, 4:05 AM
28
points
19
comments
4
min read
LW
link
Verbal parity: What is it and how to measure it? + an edited version of “Against John Searle, Gary Marcus, the Chinese Room thought experiment and its world”
philosophybear
Dec 31, 2022, 3:46 AM
2
points
0
comments
11
min read
LW
link
Should AI systems have to identify themselves?
Darren McKee
Dec 31, 2022, 2:57 AM
2
points
2
comments
1
min read
LW
link
[Question]
What do you imagine, when you imagine “taking over the world”?
johnswentworth
Dec 31, 2022, 1:04 AM
22
points
16
comments
1
min read
LW
link
A few thoughts on my self-study for alignment research
Thomas Kehrenberg
Dec 30, 2022, 10:05 PM
6
points
0
comments
2
min read
LW
link
Christmas Microscopy
jefftk
Dec 30, 2022, 9:10 PM
27
points
0
comments
1
min read
LW
link
(www.jefftk.com)
What “upside” of AI?
False Name
Dec 30, 2022, 8:58 PM
0
points
5
comments
4
min read
LW
link
Evidence on recursive self-improvement from current ML
beren
Dec 30, 2022, 8:53 PM
31
points
12
comments
6
min read
LW
link
[Question]
Is ChatGPT TAI?
Amal
30 Dec 2022 19:44 UTC
14
points
5
comments
1
min read
LW
link
My thoughts on OpenAI’s alignment plan
Orpheus16
30 Dec 2022 19:33 UTC
55
points
3
comments
20
min read
LW
link
Beyond Rewards and Values: A Non-dualistic Approach to Universal Intelligence
Akira Pyinya
30 Dec 2022 19:05 UTC
10
points
4
comments
14
min read
LW
link
10 Years of LessWrong
SebastianG
30 Dec 2022 17:15 UTC
73
points
2
comments
4
min read
LW
link
Chatbots as a Publication Format
derek shiller
30 Dec 2022 14:11 UTC
6
points
6
comments
4
min read
LW
link
Human sexuality as an interesting case study of alignment
beren
30 Dec 2022 13:37 UTC
39
points
26
comments
3
min read
LW
link
The Twitter Files: Covid Edition
Zvi
30 Dec 2022 13:30 UTC
32
points
2
comments
10
min read
LW
link
(thezvi.wordpress.com)
Previous
Back to top
Next
N
W
F
A
C
D
E
F
G
H
I
Customize appearance
Current theme:
default
A
C
D
E
F
G
H
I
Less Wrong (text)
Less Wrong (link)
Invert colors
Reset to defaults
OK
Cancel
Hi, I’m Bobby the Basilisk! Click on the minimize button (
) to minimize the theme tweaker window, so that you can see what the page looks like with the current tweaked values. (But remember,
the changes won’t be saved until you click “OK”!
)
Theme tweaker help
Show Bobby the Basilisk
OK
Cancel