Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
Archive
Sequences
About
Search
Log In
All
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
2025
All
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
All
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
Page
2
The World in 2029
Nathan Young
Mar 2, 2024, 6:03 PM
74
points
37
comments
3
min read
LW
link
The Most Dangerous Idea
rogersbacon
Mar 2, 2024, 5:53 PM
−8
points
2
comments
26
min read
LW
link
(www.secretorum.life)
Future life
DavidMadsen
Mar 2, 2024, 3:41 PM
−12
points
2
comments
2
min read
LW
link
Ugo Conti’s Whistle-Controlled Synthesizer
jefftk
Mar 2, 2024, 2:50 AM
15
points
1
comment
2
min read
LW
link
(www.jefftk.com)
A one-sentence formulation of the AI X-Risk argument I try to make
tcelferact
Mar 2, 2024, 12:44 AM
3
points
0
comments
LW
link
If you weren’t such an idiot...
kave
and
Mark Xu
Mar 2, 2024, 12:01 AM
157
points
76
comments
2
min read
LW
link
(markxu.com)
Increasing IQ is trivial
George3d6
Mar 1, 2024, 10:43 PM
37
points
61
comments
6
min read
LW
link
(epistemink.substack.com)
self-fulfilling prophecies when applying for funding
Chipmonk
Mar 1, 2024, 7:01 PM
31
points
0
comments
1
min read
LW
link
(chipmonk.substack.com)
Antagonistic AI
Xybermancer
Mar 1, 2024, 6:50 PM
−8
points
1
comment
1
min read
LW
link
Against Augmentation of Intelligence, Human or Otherwise (An Anti-Natalist Argument)
Benjamin Bourlier
Mar 1, 2024, 6:45 PM
−31
points
18
comments
3
min read
LW
link
Elon files grave charges against OpenAI
mako yass
Mar 1, 2024, 5:42 PM
38
points
10
comments
1
min read
LW
link
(www.courthousenews.com)
Notes on Dwarkesh Patel’s Podcast with Demis Hassabis
Zvi
Mar 1, 2024, 4:30 PM
93
points
0
comments
8
min read
LW
link
(thezvi.wordpress.com)
What does your philosophy maximize?
Antb
Mar 1, 2024, 4:10 PM
0
points
1
comment
1
min read
LW
link
The Defence production act and AI policy
NathanBarnard
Mar 1, 2024, 2:26 PM
37
points
0
comments
2
min read
LW
link
Don’t Endorse the Idea of Market Failure
Maxwell Tabarrok
Mar 1, 2024, 2:04 PM
14
points
22
comments
4
min read
LW
link
(www.maximum-progress.com)
[Question]
Is it possible to make more specific bookmarks?
numpyNaN
Mar 1, 2024, 12:47 PM
1
point
0
comments
1
min read
LW
link
Wholesome Culture
owencb
Mar 1, 2024, 12:08 PM
29
points
3
comments
LW
link
Adding Sensors to Mandolin?
jefftk
Mar 1, 2024, 2:10 AM
6
points
1
comment
1
min read
LW
link
(www.jefftk.com)
The Parable Of The Fallen Pendulum—Part 1
johnswentworth
Mar 1, 2024, 12:25 AM
110
points
32
comments
2
min read
LW
link
Gradations of moral weight
MichaelStJules
Feb 29, 2024, 11:08 PM
1
point
0
comments
LW
link
Approaching Human-Level Forecasting with Language Models
Fred Zhang
,
dannyhalawi
and
jsteinhardt
Feb 29, 2024, 10:36 PM
60
points
6
comments
3
min read
LW
link
Paper review: “The Unreasonable Effectiveness of Easy Training Data for Hard Tasks”
Vassil Tashev
Feb 29, 2024, 6:44 PM
11
points
0
comments
4
min read
LW
link
What’s in the box?! – Towards interpretability by distinguishing niches of value within neural networks.
Joshua Clancy
Feb 29, 2024, 6:33 PM
3
points
4
comments
128
min read
LW
link
Short Post: Discerning Truth from Trash
FinalFormal2
Feb 29, 2024, 6:09 PM
−2
points
0
comments
1
min read
LW
link
AI #53: One More Leap
Zvi
Feb 29, 2024, 4:10 PM
45
points
0
comments
38
min read
LW
link
(thezvi.wordpress.com)
Cryonics p(success) estimates are only weakly associated with interest in pursuing cryonics in the LW 2023 Survey
Andy_McKenzie
Feb 29, 2024, 2:47 PM
28
points
6
comments
1
min read
LW
link
Bengio’s Alignment Proposal: “Towards a Cautious Scientist AI with Convergent Safety Bounds”
mattmacdermott
Feb 29, 2024, 1:59 PM
76
points
19
comments
14
min read
LW
link
(yoshuabengio.org)
Tips for Empirical Alignment Research
Ethan Perez
Feb 29, 2024, 6:04 AM
164
points
4
comments
23
min read
LW
link
[Question]
Supposing the 1bit LLM paper pans out
O O
Feb 29, 2024, 5:31 AM
27
points
11
comments
1
min read
LW
link
Can RLLMv3′s ability to defend against jailbreaks be attributed to datasets containing stories about Jung’s shadow integration theory?
MiguelDev
Feb 29, 2024, 5:13 AM
7
points
2
comments
11
min read
LW
link
Post series on “Liability Law for reducing Existential Risk from AI”
Nora_Ammann
Feb 29, 2024, 4:39 AM
42
points
1
comment
1
min read
LW
link
(forum.effectivealtruism.org)
Tour Retrospective February 2024
jefftk
Feb 29, 2024, 3:50 AM
10
points
0
comments
4
min read
LW
link
(www.jefftk.com)
Locating My Eyes (Part 3 of “The Sense of Physical Necessity”)
LoganStrohl
Feb 29, 2024, 3:09 AM
43
points
4
comments
22
min read
LW
link
Conspiracy Theorists Aren’t Ignorant. They’re Bad At Epistemology.
omnizoid
Feb 28, 2024, 11:39 PM
18
points
10
comments
5
min read
LW
link
Discovering alignment windfalls reduces AI risk
goodgravy
and
stuhlmueller
Feb 28, 2024, 9:23 PM
15
points
1
comment
8
min read
LW
link
(blog.elicit.com)
my theory of the industrial revolution
bhauth
Feb 28, 2024, 9:07 PM
23
points
7
comments
3
min read
LW
link
(www.bhauth.com)
Wholesomeness and Effective Altruism
owencb
Feb 28, 2024, 8:28 PM
42
points
3
comments
LW
link
timestamping through the Singularity
throwaway918119127
Feb 28, 2024, 7:09 PM
−2
points
4
comments
8
min read
LW
link
Evidential Cooperation in Large Worlds: Potential Objections & FAQ
Chi Nguyen
and
_will_
Feb 28, 2024, 6:58 PM
42
points
5
comments
LW
link
Timaeus’s First Four Months
Jesse Hoogland
,
Daniel Murfet
,
Stan van Wingerden
and
Alexander Gietelink Oldenziel
Feb 28, 2024, 5:01 PM
173
points
6
comments
6
min read
LW
link
Notes on control evaluations for safety cases
ryan_greenblatt
,
Buck
and
Fabien Roger
Feb 28, 2024, 4:15 PM
49
points
0
comments
32
min read
LW
link
Corporate Governance for Frontier AI Labs: A Research Agenda
Matthew Wearden
Feb 28, 2024, 11:29 AM
4
points
0
comments
16
min read
LW
link
(matthewwearden.co.uk)
How AI Will Change Education
robotelvis
Feb 28, 2024, 5:30 AM
6
points
3
comments
5
min read
LW
link
(messyprogress.substack.com)
Band Lessons?
jefftk
Feb 28, 2024, 3:00 AM
13
points
3
comments
1
min read
LW
link
(www.jefftk.com)
New LessWrong review winner UI (“The LeastWrong” section and full-art post pages)
kave
Feb 28, 2024, 2:42 AM
105
points
64
comments
1
min read
LW
link
Counting arguments provide no evidence for AI doom
Nora Belrose
and
Quintin Pope
Feb 27, 2024, 11:03 PM
101
points
188
comments
14
min read
LW
link
Which animals realize which types of subjective welfare?
MichaelStJules
Feb 27, 2024, 7:31 PM
4
points
0
comments
LW
link
Biosecurity and AI: Risks and Opportunities
Steve Newman
Feb 27, 2024, 6:45 PM
11
points
1
comment
7
min read
LW
link
(www.safe.ai)
The Gemini Incident Continues
Zvi
Feb 27, 2024, 4:00 PM
45
points
6
comments
48
min read
LW
link
(thezvi.wordpress.com)
How I internalized my achievements to better deal with negative feelings
Raymond Koopmanschap
Feb 27, 2024, 3:10 PM
42
points
7
comments
6
min read
LW
link
Previous
Back to top
Next
N
W
F
A
C
D
E
F
G
H
I
Customize appearance
Current theme:
default
A
C
D
E
F
G
H
I
Less Wrong (text)
Less Wrong (link)
Invert colors
Reset to defaults
OK
Cancel
Hi, I’m Bobby the Basilisk! Click on the minimize button (
) to minimize the theme tweaker window, so that you can see what the page looks like with the current tweaked values. (But remember,
the changes won’t be saved until you click “OK”!
)
Theme tweaker help
Show Bobby the Basilisk
OK
Cancel