Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
Archive
Sequences
About
Search
Log In
All
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
2025
All
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
All
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
Page
1
Bengio’s FAQ on Catastrophic AI Risks
Vaniver
Jun 29, 2023, 11:04 PM
39
points
0
comments
1
min read
LW
link
(yoshuabengio.org)
AGI & War
Calecute
Jun 29, 2023, 10:20 PM
9
points
1
comment
1
min read
LW
link
Biosafety Regulations (BMBL) and their relevance for AI
Štěpán Los
Jun 29, 2023, 7:22 PM
4
points
0
comments
4
min read
LW
link
Nature Releases A Stupid Editorial On AI Risk
Bentham's Bulldog
Jun 29, 2023, 7:00 PM
2
points
1
comment
3
min read
LW
link
AI Safety without Alignment: How humans can WIN against AI
vicchain
Jun 29, 2023, 5:53 PM
1
point
1
comment
2
min read
LW
link
Challenge proposal: smallest possible self-hardening backdoor for RLHF
Christopher King
Jun 29, 2023, 4:56 PM
7
points
0
comments
2
min read
LW
link
AI #18: The Great Debate Debate
Zvi
Jun 29, 2023, 4:20 PM
47
points
9
comments
52
min read
LW
link
(thezvi.wordpress.com)
Bruce Sterling on the AI mania of 2023
Mitchell_Porter
Jun 29, 2023, 5:00 AM
25
points
1
comment
1
min read
LW
link
(www.newsweek.com)
Cheat sheet of AI X-risk
momom2
Jun 29, 2023, 4:28 AM
19
points
1
comment
7
min read
LW
link
Anthropically Blind: the anthropic shadow is reflectively inconsistent
Christopher King
Jun 29, 2023, 2:36 AM
43
points
40
comments
10
min read
LW
link
One path to coherence: conditionalization
porby
Jun 29, 2023, 1:08 AM
28
points
4
comments
4
min read
LW
link
AXRP announcement: Survey, Store Closing, Patreon
DanielFilan
Jun 28, 2023, 11:40 PM
14
points
0
comments
1
min read
LW
link
Metaphors for AI, and why I don’t like them
boazbarak
Jun 28, 2023, 10:47 PM
38
points
18
comments
12
min read
LW
link
Transforming Democracy: A Unique Funding Opportunity for US Federal Approval Voting
Aaron Hamlin
Jun 28, 2023, 10:07 PM
25
points
6
comments
2
min read
LW
link
AGI x Animal Welfare: A High-EV Outreach Opportunity?
simeon_c
Jun 28, 2023, 8:44 PM
29
points
0
comments
LW
link
A “weak” AGI may attempt an unlikely-to-succeed takeover
RobertM
Jun 28, 2023, 8:31 PM
56
points
17
comments
3
min read
LW
link
Progress links and tweets, 2023-06-28: “We can do big things again in Pennsylvania”
jasoncrawford
Jun 28, 2023, 8:23 PM
14
points
1
comment
1
min read
LW
link
(rootsofprogress.org)
[Question]
What money-pumps exist, if any, for deontologists?
Daniel Kokotajlo
Jun 28, 2023, 7:08 PM
39
points
35
comments
1
min read
LW
link
[Question]
What is your financial portfolio?
Algon
Jun 28, 2023, 6:39 PM
11
points
11
comments
1
min read
LW
link
Levels of safety for AI and other technologies
jasoncrawford
Jun 28, 2023, 6:35 PM
16
points
0
comments
2
min read
LW
link
(rootsofprogress.org)
LeCun says making a utility function is intractable
Iknownothing
Jun 28, 2023, 6:02 PM
2
points
3
comments
1
min read
LW
link
My research agenda in agent foundations
Alex_Altair
Jun 28, 2023, 6:00 PM
75
points
9
comments
11
min read
LW
link
AI Incident Sharing—Best practices from other fields and a comprehensive list of existing platforms
Štěpán Los
Jun 28, 2023, 5:21 PM
20
points
0
comments
4
min read
LW
link
The Case for Overconfidence is Overstated
Kevin Dorst
Jun 28, 2023, 5:21 PM
50
points
13
comments
8
min read
LW
link
(kevindorst.substack.com)
When do “brains beat brawn” in Chess? An experiment
titotal
Jun 28, 2023, 1:33 PM
318
points
106
comments
7
min read
LW
link
2
reviews
(titotal.substack.com)
Giving an evolutionary explanation for Kahneman and Tversky’s insights on subjective satisfaction
Lionel
Jun 28, 2023, 12:17 PM
−7
points
1
comment
1
min read
LW
link
(lionelpage.substack.com)
Nature: “Stop talking about tomorrow’s AI doomsday when AI poses risks today”
Ben Smith
Jun 28, 2023, 5:59 AM
40
points
8
comments
2
min read
LW
link
(www.nature.com)
Request: Put Carl Shulman’s recent podcast into an organized written format
Aryeh Englander
Jun 28, 2023, 2:58 AM
19
points
4
comments
1
min read
LW
link
Prediction Market: Will I Pull “The One Ring To Rule Them All?”
Connor Tabarrok
Jun 28, 2023, 2:41 AM
1
point
0
comments
1
min read
LW
link
(manifold.markets)
Carl Shulman on The Lunar Society (7 hour, two-part podcast)
ESRogs
Jun 28, 2023, 1:23 AM
79
points
17
comments
1
min read
LW
link
(www.dwarkeshpatel.com)
Brief summary of ai-plans.com
Iknownothing
Jun 28, 2023, 12:33 AM
9
points
4
comments
2
min read
LW
link
(ai-plans.com)
Catastrophic Risks from AI #6: Discussion and FAQ
Dan H
,
Mantas Mazeika
and
TW123
Jun 27, 2023, 11:23 PM
24
points
1
comment
13
min read
LW
link
(arxiv.org)
Catastrophic Risks from AI #5: Rogue AIs
Dan H
,
Mantas Mazeika
and
TW123
Jun 27, 2023, 10:06 PM
15
points
0
comments
22
min read
LW
link
(arxiv.org)
AISN #12: Policy Proposals from NTIA’s Request for Comment and Reconsidering Instrumental Convergence
Dan H
Jun 27, 2023, 5:20 PM
6
points
0
comments
LW
link
The Weight of the Future (Why The Apocalypse Can Be A Relief)
Sable
Jun 27, 2023, 5:18 PM
18
points
14
comments
3
min read
LW
link
(affablyevil.substack.com)
Aligning AI by optimizing for “wisdom”
JustinShovelain
and
Elliot Mckernon
Jun 27, 2023, 3:20 PM
28
points
8
comments
12
min read
LW
link
Freedom under Naturalistic Dualism
Arturo Macias
Jun 27, 2023, 2:34 PM
1
point
36
comments
1
min read
LW
link
(www.jneurophilosophy.com)
Munk AI debate: confusions and possible cruxes
Steven Byrnes
Jun 27, 2023, 2:18 PM
244
points
21
comments
8
min read
LW
link
Ateliers: Motivation
Stephen Fowler
Jun 27, 2023, 1:07 PM
7
points
0
comments
2
min read
LW
link
Self-Blinded Caffeine RCT
niplav
Jun 27, 2023, 12:38 PM
45
points
9
comments
8
min read
LW
link
An overview of the points system
Iknownothing
Jun 27, 2023, 9:09 AM
3
points
4
comments
1
min read
LW
link
(ai-plans.com)
AISC team report: Soft-optimization, Bayes and Goodhart
Simon Fischer
,
benjaminko
,
jazcarretao
,
DFNaiff
and
Jeremy Gillen
Jun 27, 2023, 6:05 AM
38
points
2
comments
15
min read
LW
link
Epistemic spot checking one claim in The Precipice
Isaac King
Jun 27, 2023, 1:03 AM
33
points
3
comments
1
min read
LW
link
nuclear costs are inflation
bhauth
Jun 26, 2023, 10:30 PM
8
points
42
comments
5
min read
LW
link
(www.bhauth.com)
Man in the Arena
Richard_Ngo
Jun 26, 2023, 9:57 PM
66
points
6
comments
8
min read
LW
link
Catastrophic Risks from AI #4: Organizational Risks
Dan H
,
Mantas Mazeika
and
TW123
Jun 26, 2023, 7:36 PM
23
points
0
comments
21
min read
LW
link
(arxiv.org)
The fraught voyage of aligned novelty
TsviBT
Jun 26, 2023, 7:10 PM
13
points
0
comments
17
min read
LW
link
[Question]
Deceptive AI vs. shifting instrumental incentives
Aryeh Englander
Jun 26, 2023, 6:09 PM
7
points
2
comments
3
min read
LW
link
On the Cost of Thriving Index
Zvi
Jun 26, 2023, 3:30 PM
33
points
6
comments
9
min read
LW
link
(thezvi.wordpress.com)
“Safety Culture for AI” is important, but isn’t going to be easy
Davidmanheim
Jun 26, 2023, 12:52 PM
47
points
2
comments
2
min read
LW
link
(forum.effectivealtruism.org)
Back to top
Next
N
W
F
A
C
D
E
F
G
H
I
Customize appearance
Current theme:
default
A
C
D
E
F
G
H
I
Less Wrong (text)
Less Wrong (link)
Invert colors
Reset to defaults
OK
Cancel
Hi, I’m Bobby the Basilisk! Click on the minimize button (
) to minimize the theme tweaker window, so that you can see what the page looks like with the current tweaked values. (But remember,
the changes won’t be saved until you click “OK”!
)
Theme tweaker help
Show Bobby the Basilisk
OK
Cancel