Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
Archive
Sequences
About
Search
Log In
All
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
2025
All
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
All
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
Page
1
Outreach success: Intro to AI risk that has been successful
Michael Tontchev
Jun 1, 2023, 11:12 PM
83
points
8
comments
74
min read
LW
link
(medium.com)
Open Source LLMs Can Now Actively Lie
Josh Levy
Jun 1, 2023, 10:03 PM
6
points
0
comments
3
min read
LW
link
Safe AI and moral AI
William D'Alessandro
Jun 1, 2023, 9:36 PM
−3
points
0
comments
10
min read
LW
link
AI #14: A Very Good Sentence
Zvi
Jun 1, 2023, 9:30 PM
118
points
30
comments
65
min read
LW
link
(thezvi.wordpress.com)
Four levels of understanding decision theory
Max H
Jun 1, 2023, 8:55 PM
12
points
11
comments
4
min read
LW
link
Things I Learned by Spending Five Thousand Hours In Non-EA Charities
jenn
Jun 1, 2023, 8:48 PM
430
points
35
comments
8
min read
LW
link
1
review
(jenn.site)
self-improvement-executors are not goal-maximizers
bhauth
Jun 1, 2023, 8:46 PM
14
points
0
comments
1
min read
LW
link
Experimental Fat Loss
johnlawrenceaspden
Jun 1, 2023, 8:26 PM
23
points
5
comments
1
min read
LW
link
Yudkowsky vs Hanson on FOOM: Whose Predictions Were Better?
1a3orn
Jun 1, 2023, 7:36 PM
137
points
76
comments
24
min read
LW
link
2
reviews
Progress links and tweets, 2023-06-01
jasoncrawford
Jun 1, 2023, 7:03 PM
10
points
3
comments
1
min read
LW
link
(rootsofprogress.org)
[Question]
When does an AI become intelligent enough to become self-aware and power-seeking?
FinalFormal2
Jun 1, 2023, 6:09 PM
1
point
1
comment
1
min read
LW
link
Uncertainty about the future does not imply that AGI will go well
Lauro Langosco
Jun 1, 2023, 5:38 PM
62
points
11
comments
7
min read
LW
link
[Question]
What are the arguments for/against FOOM?
FinalFormal2
Jun 1, 2023, 5:23 PM
8
points
0
comments
1
min read
LW
link
Change my mind: Veganism entails trade-offs, and health is one of the axes
Elizabeth
Jun 1, 2023, 5:10 PM
160
points
85
comments
19
min read
LW
link
2
reviews
(acesounderglass.com)
The unspoken but ridiculous assumption of AI doom: the hidden doom assumption
Christopher King
Jun 1, 2023, 5:01 PM
−9
points
1
comment
3
min read
LW
link
Don’t waste your time meditating on meditation retreats!
EternallyBlissful
Jun 1, 2023, 4:56 PM
23
points
7
comments
11
min read
LW
link
[Request]: Use “Epilogenics” instead of “Eugenics” in most circumstances
GeneSmith
Jun 1, 2023, 3:36 PM
56
points
49
comments
1
min read
LW
link
Book Club: Thomas Schelling’s “The Strategy of Conflict”
Optimization Process
Jun 1, 2023, 3:29 PM
6
points
1
comment
1
min read
LW
link
Probably tell your friends when they make big mistakes
Chi Nguyen
Jun 1, 2023, 2:30 PM
15
points
1
comment
LW
link
Yes, avoiding extinction from AI *is* an urgent priority: a response to Seth Lazar, Jeremy Howard, and Arvind Narayanan.
Soroush Pour
Jun 1, 2023, 1:38 PM
17
points
0
comments
5
min read
LW
link
(www.soroushjp.com)
Work dumber not smarter
lemonhope
Jun 1, 2023, 12:40 PM
101
points
17
comments
3
min read
LW
link
Short Remark on the (subjective) mathematical ‘naturalness’ of the Nanda—Lieberum addition modulo 113 algorithm
carboniferous_umbraculum
Jun 1, 2023, 11:31 AM
104
points
12
comments
2
min read
LW
link
How will they feed us
meijer1973
Jun 1, 2023, 8:49 AM
4
points
3
comments
5
min read
LW
link
“LLMs Don’t Have a Coherent Model of the World”—What it Means, Why it Matters
Davidmanheim
Jun 1, 2023, 7:46 AM
32
points
2
comments
7
min read
LW
link
General intelligence: what is it, what makes it hard, and will we have it soon?
homeopathicsyzygy
Jun 1, 2023, 6:46 AM
2
points
0
comments
21
min read
LW
link
Maximal Sentience: A Sentience Spectrum and Test Foundation
Snowyiu
Jun 1, 2023, 6:45 AM
1
point
2
comments
4
min read
LW
link
Re: The Crux List
Logan Zoellner
Jun 1, 2023, 4:48 AM
11
points
0
comments
2
min read
LW
link
An explanation of decision theories
metachirality
Jun 1, 2023, 3:42 AM
20
points
4
comments
5
min read
LW
link
Dancing to Positional Calling
jefftk
Jun 1, 2023, 2:40 AM
11
points
2
comments
2
min read
LW
link
(www.jefftk.com)
Intrinsic vs. Extrinsic Alignment
Alfonso Pérez Escudero
Jun 1, 2023, 1:06 AM
1
point
1
comment
3
min read
LW
link
Limiting factors to predict AI take-off speed
Alfonso Pérez Escudero
May 31, 2023, 11:19 PM
1
point
0
comments
6
min read
LW
link
Unpredictability and the Increasing Difficulty of AI Alignment for Increasingly Intelligent AI
Max_He-Ho
May 31, 2023, 10:25 PM
5
points
2
comments
20
min read
LW
link
Shutdown-Seeking AI
Simon Goldstein
May 31, 2023, 10:19 PM
50
points
32
comments
15
min read
LW
link
Full Automation is Unlikely and Unnecessary for Explosive Growth
aog
May 31, 2023, 9:55 PM
28
points
3
comments
5
min read
LW
link
LessWrong Community Weekend 2023 Updates: Keynote Speaker Malcolm Ocean, Remaining Tickets and More
Henry Prowbell
May 31, 2023, 9:53 PM
23
points
0
comments
2
min read
LW
link
The Divine Move Paradox & Thinking as a Species
Christopher James Hart
May 31, 2023, 9:38 PM
9
points
8
comments
3
min read
LW
link
Intent-aligned AI systems deplete human agency: the need for agency foundations research in AI safety
catubc
May 31, 2023, 9:18 PM
26
points
4
comments
11
min read
LW
link
[Question]
How much overlap is there between the utility function of GPT-n and GPT-(n+1), assuming both are near AGI?
Phosphorous
May 31, 2023, 8:28 PM
2
points
0
comments
2
min read
LW
link
My AI-risk cartoon
pre
May 31, 2023, 7:46 PM
6
points
0
comments
1
min read
LW
link
Evaluation Evidence Reconstructions of Mock Crimes Submission 3
Alan E Dunne
May 31, 2023, 7:03 PM
−1
points
0
comments
3
min read
LW
link
Improving Mathematical Reasoning with-Process Supervision
p.b.
May 31, 2023, 7:00 PM
14
points
3
comments
1
min read
LW
link
(openai.com)
The Crux List
Zvi
May 31, 2023, 6:30 PM
72
points
19
comments
33
min read
LW
link
(thezvi.wordpress.com)
Stages of Survival
Zvi
May 31, 2023, 6:30 PM
44
points
0
comments
17
min read
LW
link
(thezvi.wordpress.com)
Types and Degrees of Alignment
Zvi
May 31, 2023, 6:30 PM
36
points
10
comments
8
min read
LW
link
(thezvi.wordpress.com)
To Predict What Happens, Ask What Happens
Zvi
May 31, 2023, 6:30 PM
81
points
0
comments
9
min read
LW
link
(thezvi.wordpress.com)
A push towards interactive transformer decoding
R0bk
May 31, 2023, 5:56 PM
3
points
0
comments
2
min read
LW
link
(github.com)
Neuroevolution, Social Intelligence, and Logic
vinnik.dmitry07
May 31, 2023, 5:54 PM
1
point
0
comments
10
min read
LW
link
Contrast Pairs Drive the Empirical Performance of Contrast Consistent Search (CCS)
Scott Emmons
May 31, 2023, 5:09 PM
97
points
1
comment
6
min read
LW
link
1
review
Cosmopolitan values don’t come free
So8res
May 31, 2023, 3:58 PM
137
points
85
comments
1
min read
LW
link
[Question]
Arguments Against Fossil Future?
Sable
May 31, 2023, 1:41 PM
13
points
29
comments
1
min read
LW
link
Back to top
Next
N
W
F
A
C
D
E
F
G
H
I
Customize appearance
Current theme:
default
A
C
D
E
F
G
H
I
Less Wrong (text)
Less Wrong (link)
Invert colors
Reset to defaults
OK
Cancel
Hi, I’m Bobby the Basilisk! Click on the minimize button (
) to minimize the theme tweaker window, so that you can see what the page looks like with the current tweaked values. (But remember,
the changes won’t be saved until you click “OK”!
)
Theme tweaker help
Show Bobby the Basilisk
OK
Cancel