Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
Archive
Sequences
About
Search
Log In
All
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
2025
All
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
All
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
Page
1
[Question]
How could AIs ‘see’ each other’s source code?
Kenny
Jun 2, 2023, 10:41 PM
29
points
45
comments
1
min read
LW
link
Proposal: labs should precommit to pausing if an AI argues for itself to be improved
NickGabs
Jun 2, 2023, 10:31 PM
3
points
3
comments
4
min read
LW
link
Inference from a Mathematical Description of an Existing Alignment Research: a proposal for an outer alignment research program
Christopher King
Jun 2, 2023, 9:54 PM
7
points
4
comments
16
min read
LW
link
Thoughts on Dancing the Whole Dance: Positional Calling for Contra
jefftk
Jun 2, 2023, 8:50 PM
10
points
0
comments
5
min read
LW
link
(www.jefftk.com)
Advice for Entering AI Safety Research
scasper
Jun 2, 2023, 8:46 PM
26
points
2
comments
5
min read
LW
link
AI should be used to find better morality
Jorterder
Jun 2, 2023, 8:38 PM
−21
points
1
comment
1
min read
LW
link
A mind needn’t be curious to reap the benefits of curiosity
So8res
Jun 2, 2023, 6:00 PM
78
points
14
comments
1
min read
LW
link
[Question]
Are computationally complex algorithms expensive to have, expensive to operate, or both?
Noosphere89
Jun 2, 2023, 5:50 PM
7
points
5
comments
1
min read
LW
link
[Replication] Conjecture’s Sparse Coding in Toy Models
Hoagy
and
Logan Riggs
Jun 2, 2023, 5:34 PM
24
points
0
comments
1
min read
LW
link
Limits to Learning: Rethinking AGI’s Path to Dominance
tangerine
Jun 2, 2023, 4:43 PM
10
points
4
comments
15
min read
LW
link
The Control Problem: Unsolved or Unsolvable?
Remmelt
Jun 2, 2023, 3:42 PM
55
points
46
comments
14
min read
LW
link
Hallucinating Suction
Johannes C. Mayer
Jun 2, 2023, 2:16 PM
6
points
0
comments
2
min read
LW
link
Winning doesn’t need to flow through increases in rationality
Michel
Jun 2, 2023, 12:05 PM
11
points
5
comments
1
min read
LW
link
Product Recommendation: LessWrong dialogues with Recast
Bart Bussmann
Jun 2, 2023, 8:05 AM
5
points
0
comments
1
min read
LW
link
Think carefully before calling RL policies “agents”
TurnTrout
Jun 2, 2023, 3:46 AM
134
points
38
comments
4
min read
LW
link
1
review
Dreams of “Mathopedia”
Nicholas / Heather Kross
Jun 2, 2023, 1:30 AM
40
points
16
comments
2
min read
LW
link
(www.thinkingmuchbetter.com)
Outreach success: Intro to AI risk that has been successful
Michael Tontchev
Jun 1, 2023, 11:12 PM
83
points
8
comments
74
min read
LW
link
(medium.com)
Open Source LLMs Can Now Actively Lie
Josh Levy
Jun 1, 2023, 10:03 PM
6
points
0
comments
3
min read
LW
link
Safe AI and moral AI
William D'Alessandro
Jun 1, 2023, 9:36 PM
−3
points
0
comments
10
min read
LW
link
AI #14: A Very Good Sentence
Zvi
Jun 1, 2023, 9:30 PM
118
points
30
comments
65
min read
LW
link
(thezvi.wordpress.com)
Four levels of understanding decision theory
Max H
Jun 1, 2023, 8:55 PM
12
points
11
comments
4
min read
LW
link
Things I Learned by Spending Five Thousand Hours In Non-EA Charities
jenn
Jun 1, 2023, 8:48 PM
430
points
35
comments
8
min read
LW
link
1
review
(jenn.site)
self-improvement-executors are not goal-maximizers
bhauth
Jun 1, 2023, 8:46 PM
14
points
0
comments
1
min read
LW
link
Experimental Fat Loss
johnlawrenceaspden
Jun 1, 2023, 8:26 PM
23
points
5
comments
1
min read
LW
link
Yudkowsky vs Hanson on FOOM: Whose Predictions Were Better?
1a3orn
Jun 1, 2023, 7:36 PM
137
points
76
comments
24
min read
LW
link
2
reviews
Progress links and tweets, 2023-06-01
jasoncrawford
Jun 1, 2023, 7:03 PM
10
points
3
comments
1
min read
LW
link
(rootsofprogress.org)
[Question]
When does an AI become intelligent enough to become self-aware and power-seeking?
FinalFormal2
Jun 1, 2023, 6:09 PM
1
point
1
comment
1
min read
LW
link
Uncertainty about the future does not imply that AGI will go well
Lauro Langosco
Jun 1, 2023, 5:38 PM
62
points
11
comments
7
min read
LW
link
[Question]
What are the arguments for/against FOOM?
FinalFormal2
Jun 1, 2023, 5:23 PM
8
points
0
comments
1
min read
LW
link
Change my mind: Veganism entails trade-offs, and health is one of the axes
Elizabeth
Jun 1, 2023, 5:10 PM
160
points
85
comments
19
min read
LW
link
2
reviews
(acesounderglass.com)
The unspoken but ridiculous assumption of AI doom: the hidden doom assumption
Christopher King
Jun 1, 2023, 5:01 PM
−9
points
1
comment
3
min read
LW
link
Don’t waste your time meditating on meditation retreats!
EternallyBlissful
Jun 1, 2023, 4:56 PM
23
points
7
comments
11
min read
LW
link
[Request]: Use “Epilogenics” instead of “Eugenics” in most circumstances
GeneSmith
Jun 1, 2023, 3:36 PM
56
points
49
comments
1
min read
LW
link
Book Club: Thomas Schelling’s “The Strategy of Conflict”
Optimization Process
Jun 1, 2023, 3:29 PM
6
points
1
comment
1
min read
LW
link
Probably tell your friends when they make big mistakes
Chi Nguyen
Jun 1, 2023, 2:30 PM
15
points
1
comment
LW
link
Yes, avoiding extinction from AI *is* an urgent priority: a response to Seth Lazar, Jeremy Howard, and Arvind Narayanan.
Soroush Pour
Jun 1, 2023, 1:38 PM
17
points
0
comments
5
min read
LW
link
(www.soroushjp.com)
Work dumber not smarter
lemonhope
Jun 1, 2023, 12:40 PM
101
points
17
comments
3
min read
LW
link
Short Remark on the (subjective) mathematical ‘naturalness’ of the Nanda—Lieberum addition modulo 113 algorithm
carboniferous_umbraculum
Jun 1, 2023, 11:31 AM
104
points
12
comments
2
min read
LW
link
How will they feed us
meijer1973
Jun 1, 2023, 8:49 AM
4
points
3
comments
5
min read
LW
link
“LLMs Don’t Have a Coherent Model of the World”—What it Means, Why it Matters
Davidmanheim
Jun 1, 2023, 7:46 AM
32
points
2
comments
7
min read
LW
link
General intelligence: what is it, what makes it hard, and will we have it soon?
homeopathicsyzygy
Jun 1, 2023, 6:46 AM
2
points
0
comments
21
min read
LW
link
Maximal Sentience: A Sentience Spectrum and Test Foundation
Snowyiu
Jun 1, 2023, 6:45 AM
1
point
2
comments
4
min read
LW
link
Re: The Crux List
Logan Zoellner
Jun 1, 2023, 4:48 AM
11
points
0
comments
2
min read
LW
link
An explanation of decision theories
metachirality
Jun 1, 2023, 3:42 AM
20
points
4
comments
5
min read
LW
link
Dancing to Positional Calling
jefftk
Jun 1, 2023, 2:40 AM
11
points
2
comments
2
min read
LW
link
(www.jefftk.com)
Intrinsic vs. Extrinsic Alignment
Alfonso Pérez Escudero
Jun 1, 2023, 1:06 AM
1
point
1
comment
3
min read
LW
link
Limiting factors to predict AI take-off speed
Alfonso Pérez Escudero
May 31, 2023, 11:19 PM
1
point
0
comments
6
min read
LW
link
Unpredictability and the Increasing Difficulty of AI Alignment for Increasingly Intelligent AI
Max_He-Ho
May 31, 2023, 10:25 PM
5
points
2
comments
20
min read
LW
link
Shutdown-Seeking AI
Simon Goldstein
May 31, 2023, 10:19 PM
50
points
32
comments
15
min read
LW
link
Full Automation is Unlikely and Unnecessary for Explosive Growth
aog
May 31, 2023, 9:55 PM
28
points
3
comments
5
min read
LW
link
Back to top
Next
N
W
F
A
C
D
E
F
G
H
I
Customize appearance
Current theme:
default
A
C
D
E
F
G
H
I
Less Wrong (text)
Less Wrong (link)
Invert colors
Reset to defaults
OK
Cancel
Hi, I’m Bobby the Basilisk! Click on the minimize button (
) to minimize the theme tweaker window, so that you can see what the page looks like with the current tweaked values. (But remember,
the changes won’t be saved until you click “OK”!
)
Theme tweaker help
Show Bobby the Basilisk
OK
Cancel