Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
Archive
Sequences
About
Search
Log In
All
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
2025
All
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
All
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
Page
1
Epistemic Hell
rogersbacon
Jan 27, 2024, 5:13 PM
71
points
20
comments
14
min read
LW
link
David Burns Thinks Psychotherapy Is a Learnable Skill. Git Gud.
Morpheus
Jan 27, 2024, 1:21 PM
28
points
20
comments
11
min read
LW
link
(podcast.clearerthinking.org)
Aligned AI is dual use technology
lc
Jan 27, 2024, 6:50 AM
58
points
31
comments
2
min read
LW
link
Questions I’d Want to Ask an AGI+ to Test Its Understanding of Ethics
sweenesm
Jan 26, 2024, 11:40 PM
14
points
6
comments
4
min read
LW
link
An Invitation to Refrain from Downvoting Posts into Net-Negative Karma
MikkW
Jan 26, 2024, 8:13 PM
2
points
12
comments
1
min read
LW
link
The Good Balsamic Vinegar
jenn
Jan 26, 2024, 7:30 PM
52
points
4
comments
2
min read
LW
link
(jenn.site)
The Perspective-based Explanation to the Reflective Inconsistency Paradox
dadadarren
Jan 26, 2024, 7:00 PM
10
points
16
comments
8
min read
LW
link
To Boldly Code
StrivingForLegibility
Jan 26, 2024, 6:25 PM
25
points
4
comments
3
min read
LW
link
Incorporating Mechanism Design Into Decision Theory
StrivingForLegibility
Jan 26, 2024, 6:25 PM
17
points
4
comments
4
min read
LW
link
Making every researcher seek grants is a broken model
jasoncrawford
Jan 26, 2024, 4:06 PM
159
points
41
comments
4
min read
LW
link
(rootsofprogress.org)
Notes on Innocence
David Gross
Jan 26, 2024, 2:45 PM
13
points
21
comments
18
min read
LW
link
Stacked Laptop Monitor
jefftk
Jan 26, 2024, 2:10 PM
22
points
5
comments
1
min read
LW
link
(www.jefftk.com)
Surgery Works Well Without The FDA
Maxwell Tabarrok
Jan 26, 2024, 1:31 PM
43
points
28
comments
4
min read
LW
link
(maximumprogress.substack.com)
[Question]
Workshop (hackathon, residence program, etc.) about for-profit AI Safety projects?
Roman Leventov
Jan 26, 2024, 9:49 AM
21
points
5
comments
1
min read
LW
link
Without fundamental advances, misalignment and catastrophe are the default outcomes of training powerful AI
Jeremy Gillen
and
peterbarnett
Jan 26, 2024, 7:22 AM
161
points
60
comments
57
min read
LW
link
Approximately Bayesian Reasoning: Knightian Uncertainty, Goodhart, and the Look-Elsewhere Effect
RogerDearnaley
Jan 26, 2024, 3:58 AM
16
points
2
comments
11
min read
LW
link
Musings on Cargo Cult Consciousness
Gareth Davidson
Jan 25, 2024, 11:00 PM
−13
points
11
comments
17
min read
LW
link
RAND report finds no effect of current LLMs on viability of bioterrorism attacks
StellaAthena
Jan 25, 2024, 7:17 PM
94
points
14
comments
1
min read
LW
link
(www.rand.org)
[Question]
Bayesian Reflection Principles and Ignorance of the Future
crickets
Jan 25, 2024, 7:00 PM
5
points
3
comments
1
min read
LW
link
“Does your paradigm beget new, good, paradigms?”
Raemon
Jan 25, 2024, 6:23 PM
40
points
6
comments
2
min read
LW
link
AI #48: The Talk of Davos
Zvi
Jan 25, 2024, 4:20 PM
38
points
9
comments
36
min read
LW
link
(thezvi.wordpress.com)
Importing a Python File by Name
jefftk
Jan 25, 2024, 4:00 PM
12
points
7
comments
1
min read
LW
link
(www.jefftk.com)
[Repost] The Copenhagen Interpretation of Ethics
mesaoptimizer
Jan 25, 2024, 3:20 PM
77
points
4
comments
5
min read
LW
link
(web.archive.org)
Nash Bargaining between Subagents doesn’t solve the Shutdown Problem
A.H.
Jan 25, 2024, 10:47 AM
22
points
1
comment
9
min read
LW
link
Status-oriented spending
Adam Zerner
Jan 25, 2024, 6:46 AM
14
points
19
comments
4
min read
LW
link
Protecting agent boundaries
Chipmonk
Jan 25, 2024, 4:13 AM
11
points
6
comments
2
min read
LW
link
[Question]
Is a random box of gas predictable after 20 seconds?
Thomas Kwa
and
habryka
Jan 24, 2024, 11:00 PM
37
points
35
comments
1
min read
LW
link
[Question]
Will quantum randomness affect the 2028 election?
Thomas Kwa
and
habryka
Jan 24, 2024, 10:54 PM
66
points
52
comments
1
min read
LW
link
AISN #30: Investments in Compute and Military AI Plus, Japan and Singapore’s National AI Safety Institutes
Dan H
and
Corin Katzke
Jan 24, 2024, 7:38 PM
27
points
1
comment
6
min read
LW
link
(newsletter.safe.ai)
Krueger Lab AI Safety Internship 2024
Joey Bream
Jan 24, 2024, 7:17 PM
3
points
0
comments
1
min read
LW
link
Agents that act for reasons: a thought experiment
Michele Campolo
Jan 24, 2024, 4:47 PM
3
points
0
comments
3
min read
LW
link
Impact Assessment of AI Safety Camp (Arb Research)
Samuel Holton
Jan 24, 2024, 4:19 PM
10
points
0
comments
11
min read
LW
link
(forum.effectivealtruism.org)
The case for ensuring that powerful AIs are controlled
ryan_greenblatt
and
Buck
Jan 24, 2024, 4:11 PM
276
points
73
comments
28
min read
LW
link
LLMs can strategically deceive while doing gain-of-function research
Igor Ivanov
Jan 24, 2024, 3:45 PM
36
points
4
comments
11
min read
LW
link
Monthly Roundup #14: January 2024
Zvi
Jan 24, 2024, 12:50 PM
38
points
22
comments
44
min read
LW
link
(thezvi.wordpress.com)
This might be the last AI Safety Camp
Remmelt
and
Linda Linsefors
Jan 24, 2024, 9:33 AM
196
points
34
comments
1
min read
LW
link
Global LessWrong/AC10 Meetup on VRChat
Tomás B.
and
the gears to ascension
Jan 24, 2024, 5:44 AM
15
points
2
comments
1
min read
LW
link
Humans aren’t fleeb.
Charlie Steiner
Jan 24, 2024, 5:31 AM
37
points
5
comments
2
min read
LW
link
A Paradigm Shift in Sustainability
Jose Miguel Cruz y Celis
Jan 23, 2024, 11:34 PM
5
points
0
comments
18
min read
LW
link
From Finite Factors to Bayes Nets
J Bostock
Jan 23, 2024, 8:03 PM
38
points
7
comments
8
min read
LW
link
Institutional economics through the lens of scale-free regulative development, morphogenesis, and cognitive science
Roman Leventov
Jan 23, 2024, 7:42 PM
8
points
0
comments
14
min read
LW
link
Making a Secular Solstice Songbook
jefftk
Jan 23, 2024, 7:40 PM
38
points
6
comments
1
min read
LW
link
(www.jefftk.com)
Simple Appreciations
Jonathan Moregård
23 Jan 2024 16:23 UTC
17
points
11
comments
4
min read
LW
link
(open.substack.com)
[Question]
What environmental cues had you not seen them would have ended in disaster?
koratkar
23 Jan 2024 14:59 UTC
11
points
1
comment
1
min read
LW
link
Loneliness and suicide mitigation for students using GPT3-enabled chatbots (survey of Replika users in Nature)
Kaj_Sotala
23 Jan 2024 14:05 UTC
45
points
2
comments
2
min read
LW
link
(www.nature.com)
“Safety as a Scientific Pursuit” (2024)
technicalities
23 Jan 2024 12:40 UTC
17
points
3
comments
2
min read
LW
link
(banburismus.substack.com)
Brainstorming: Slow Takeoff
David Piepgrass
23 Jan 2024 6:58 UTC
3
points
0
comments
51
min read
LW
link
Reframing Acausal Trolling as Acausal Patronage
StrivingForLegibility
23 Jan 2024 3:04 UTC
14
points
0
comments
2
min read
LW
link
Orthogonality or the “Human Worth Hypothesis”?
Jeffs
23 Jan 2024 0:57 UTC
21
points
31
comments
3
min read
LW
link
the subreddit size threshold
bhauth
23 Jan 2024 0:38 UTC
32
points
3
comments
4
min read
LW
link
(www.bhauth.com)
Back to top
Next
N
W
F
A
C
D
E
F
G
H
I
Customize appearance
Current theme:
default
A
C
D
E
F
G
H
I
Less Wrong (text)
Less Wrong (link)
Invert colors
Reset to defaults
OK
Cancel
Hi, I’m Bobby the Basilisk! Click on the minimize button (
) to minimize the theme tweaker window, so that you can see what the page looks like with the current tweaked values. (But remember,
the changes won’t be saved until you click “OK”!
)
Theme tweaker help
Show Bobby the Basilisk
OK
Cancel