Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
Archive
Sequences
About
Search
Log In
All
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
2025
All
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
All
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
Page
1
Inositol Non-Results
Elizabeth
Nov 29, 2023, 9:40 PM
20
points
2
comments
1
min read
LW
link
(acesounderglass.com)
Losing Metaphors: Zip and Paste
jefftk
Nov 29, 2023, 8:31 PM
26
points
6
comments
1
min read
LW
link
(www.jefftk.com)
Preserving our heritage: Building a movement and a knowledge ark for current and future generations
rnk8
Nov 29, 2023, 7:20 PM
0
points
5
comments
12
min read
LW
link
AGI Alignment is Absurd
Youssef Mohamed
Nov 29, 2023, 7:11 PM
−9
points
4
comments
3
min read
LW
link
The origins of the steam engine: An essay with interactive animated diagrams
jasoncrawford
Nov 29, 2023, 6:30 PM
30
points
1
comment
1
min read
LW
link
(rootsofprogress.org)
ChatGPT 4 solved all the gotcha problems I posed that tripped ChatGPT 3.5
VipulNaik
Nov 29, 2023, 6:11 PM
33
points
16
comments
14
min read
LW
link
“Clean” vs. “messy” goal-directedness (Section 2.2.3 of “Scheming AIs”)
Joe Carlsmith
Nov 29, 2023, 4:32 PM
29
points
1
comment
11
min read
LW
link
Lying Alignment Chart
Zack_M_Davis
Nov 29, 2023, 4:15 PM
77
points
17
comments
1
min read
LW
link
Rethink Priorities: Seeking Expressions of Interest for Special Projects Next Year
kierangreig
Nov 29, 2023, 1:59 PM
4
points
0
comments
5
min read
LW
link
[Question]
Thoughts on teletransportation with copies?
titotal
Nov 29, 2023, 12:56 PM
15
points
13
comments
1
min read
LW
link
Interpretability with Sparse Autoencoders (Colab exercises)
CallumMcDougall
Nov 29, 2023, 12:56 PM
76
points
9
comments
4
min read
LW
link
The 101 Space You Will Always Have With You
Screwtape
Nov 29, 2023, 4:56 AM
275
points
23
comments
6
min read
LW
link
1
review
Trust your intuition—Kahneman’s book misses the forest for the trees
mnvr
Nov 29, 2023, 4:37 AM
−2
points
2
comments
2
min read
LW
link
Process Substitution Without Shell?
jefftk
Nov 29, 2023, 3:20 AM
19
points
18
comments
2
min read
LW
link
(www.jefftk.com)
Deception Chess: Game #2
Zane
Nov 29, 2023, 2:43 AM
29
points
17
comments
2
min read
LW
link
Black Box Biology
GeneSmith
Nov 29, 2023, 2:27 AM
65
points
30
comments
2
min read
LW
link
[Question]
What would be the shelf life of nuclear weapon-secrecy if nuclear weapons had not immediately been used in combat?
Gram Stone
Nov 29, 2023, 12:53 AM
7
points
2
comments
1
min read
LW
link
Scaling laws for dominant assurance contracts
jessicata
Nov 28, 2023, 11:11 PM
36
points
5
comments
7
min read
LW
link
(unstableontology.com)
I’m confused about innate smell neuroanatomy
Steven Byrnes
Nov 28, 2023, 8:49 PM
40
points
2
comments
9
min read
LW
link
How to Control an LLM’s Behavior (why my P(DOOM) went down)
RogerDearnaley
Nov 28, 2023, 7:56 PM
65
points
30
comments
11
min read
LW
link
[Question]
Is there a word for discrimination against A.I.?
Aaron Bohannon
Nov 28, 2023, 7:03 PM
1
point
4
comments
1
min read
LW
link
Update #2 to “Dominant Assurance Contract Platform”: EnsureDone
moyamo
Nov 28, 2023, 6:02 PM
33
points
2
comments
1
min read
LW
link
Ethicophysics II: Politics is the Mind-Savior
MadHatter
Nov 28, 2023, 4:27 PM
−9
points
9
comments
4
min read
LW
link
(bittertruths.substack.com)
Neither EA nor e/acc is what we need to build the future
jasoncrawford
Nov 28, 2023, 4:04 PM
7
points
22
comments
3
min read
LW
link
(rootsofprogress.org)
Agentic Growth
Logan Kieller
Nov 28, 2023, 3:45 PM
1
point
0
comments
3
min read
LW
link
(logankieller.substack.com)
AISC project: How promising is automating alignment research? (literature review)
Bogdan Ionut Cirstea
Nov 28, 2023, 2:47 PM
4
points
1
comment
1
min read
LW
link
(docs.google.com)
A day in the life of a mechanistic interpretability researcher
Bill Benzon
Nov 28, 2023, 2:45 PM
3
points
3
comments
1
min read
LW
link
Two sources of beyond-episode goals (Section 2.2.2 of “Scheming AIs”)
Joe Carlsmith
Nov 28, 2023, 1:49 PM
11
points
1
comment
15
min read
LW
link
Self-Referential Probabilistic Logic Admits the Payor’s Lemma
Yudhister Kumar
Nov 28, 2023, 10:27 AM
80
points
14
comments
6
min read
LW
link
[Question]
How can I use AI without increasing AI-risk?
Yoav Ravid
Nov 28, 2023, 10:05 AM
18
points
6
comments
1
min read
LW
link
A Reading From The Book Of Sequences
Screwtape
Nov 28, 2023, 6:45 AM
8
points
0
comments
4
min read
LW
link
Anthropic Fall 2023 Debate Progress Update
Ansh Radhakrishnan
Nov 28, 2023, 5:37 AM
76
points
9
comments
12
min read
LW
link
Apocalypse insurance, and the hardline libertarian take on AI risk
So8res
Nov 28, 2023, 2:09 AM
134
points
40
comments
7
min read
LW
link
1
review
My techno-optimism [By Vitalik Buterin]
habryka
Nov 27, 2023, 11:53 PM
107
points
17
comments
2
min read
LW
link
(www.lesswrong.com)
[Question]
Could Germany have won World War I with high probability given the benefit of hindsight?
Roko
Nov 27, 2023, 10:52 PM
10
points
18
comments
1
min read
LW
link
[Question]
Could World War I have been prevented given the benefit of hindsight?
Roko
Nov 27, 2023, 10:39 PM
16
points
8
comments
1
min read
LW
link
AISC 2024 - Project Summaries
NickyP
Nov 27, 2023, 10:32 PM
48
points
3
comments
18
min read
LW
link
“Epistemic range of motion” and LessWrong moderation
habryka
and
Gabriel Alfour
Nov 27, 2023, 9:58 PM
65
points
3
comments
12
min read
LW
link
Apply to the Conceptual Boundaries Workshop for AI Safety
Chipmonk
Nov 27, 2023, 9:04 PM
50
points
0
comments
3
min read
LW
link
There is no IQ for AI
Gabriel Alfour
Nov 27, 2023, 6:21 PM
30
points
10
comments
9
min read
LW
link
(cognition.cafe)
Two concepts of an “episode” (Section 2.2.1 of “Scheming AIs”)
Joe Carlsmith
Nov 27, 2023, 6:01 PM
19
points
1
comment
13
min read
LW
link
[Linkpost] George Mack’s Razors
trevor
Nov 27, 2023, 5:53 PM
38
points
8
comments
3
min read
LW
link
(twitter.com)
On possible cross-fertilization between AI and neuroscience [Creativity]
Bill Benzon
Nov 27, 2023, 4:50 PM
15
points
22
comments
7
min read
LW
link
Ethicophysics I
MadHatter
Nov 27, 2023, 3:44 PM
−1
points
16
comments
1
min read
LW
link
(open.substack.com)
Sentience Institute 2023 End of Year Summary
michael_dello
Nov 27, 2023, 12:11 PM
11
points
0
comments
5
min read
LW
link
(www.sentienceinstitute.org)
[Question]
A Question about Corrigibility (2015)
A.H.
Nov 27, 2023, 12:05 PM
4
points
2
comments
1
min read
LW
link
Appendices to the live agendas
technicalities
and
Stag
Nov 27, 2023, 11:10 AM
16
points
4
comments
1
min read
LW
link
Shallow review of live agendas in alignment & safety
technicalities
and
Stag
Nov 27, 2023, 11:10 AM
348
points
73
comments
29
min read
LW
link
1
review
Napoleon stole the Roman Inquisition archives and investigated the Galileo case
Meow P
Nov 27, 2023, 9:41 AM
−3
points
0
comments
1
min read
LW
link
(www.cricetuscricetus.co.uk)
Found Paper: “FDT in an evolutionary environment”
the gears to ascension
Nov 27, 2023, 5:27 AM
30
points
47
comments
1
min read
LW
link
(arxiv.org)
Back to top
Next
N
W
F
A
C
D
E
F
G
H
I
Customize appearance
Current theme:
default
A
C
D
E
F
G
H
I
Less Wrong (text)
Less Wrong (link)
Invert colors
Reset to defaults
OK
Cancel
Hi, I’m Bobby the Basilisk! Click on the minimize button (
) to minimize the theme tweaker window, so that you can see what the page looks like with the current tweaked values. (But remember,
the changes won’t be saved until you click “OK”!
)
Theme tweaker help
Show Bobby the Basilisk
OK
Cancel