Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
Archive
Sequences
About
Search
Log In
All
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
2025
All
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
All
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
Page
1
TinyStories: Small Language Models That Still Speak Coherent English
Ulisse Mini
May 28, 2023, 10:23 PM
66
points
8
comments
2
min read
LW
link
(arxiv.org)
“Membranes” is better terminology than “boundaries” alone
Chipmonk
and
the gears to ascension
May 28, 2023, 10:16 PM
30
points
12
comments
3
min read
LW
link
The king token
p.b.
May 28, 2023, 7:18 PM
17
points
0
comments
4
min read
LW
link
Language Agents Reduce the Risk of Existential Catastrophe
cdkg
and
Simon Goldstein
May 28, 2023, 7:10 PM
39
points
14
comments
26
min read
LW
link
Devil’s Advocate: Adverse Selection Against Conscientiousness
lionhearted (Sebastian Marshall)
May 28, 2023, 5:53 PM
10
points
2
comments
1
min read
LW
link
Reacts now enabled on 100% of posts, though still just experimenting
Ruby
May 28, 2023, 5:36 AM
88
points
73
comments
2
min read
LW
link
My AI Alignment Research Agenda and Threat Model, right now (May 2023)
Nicholas / Heather Kross
May 28, 2023, 3:23 AM
25
points
0
comments
6
min read
LW
link
(www.thinkingmuchbetter.com)
Kelly betting vs expectation maximization
MorgneticField
May 28, 2023, 1:54 AM
35
points
33
comments
5
min read
LW
link
Why and When Interpretability Work is Dangerous
Nicholas / Heather Kross
May 28, 2023, 12:27 AM
20
points
9
comments
8
min read
LW
link
(www.thinkingmuchbetter.com)
Twin Cities ACX Meetup—June 2023
Timothy M.
May 27, 2023, 8:11 PM
1
point
1
comment
1
min read
LW
link
Project Idea: Challenge Groups for Alignment Researchers
Adam Zerner
May 27, 2023, 8:10 PM
13
points
0
comments
1
min read
LW
link
Introspective Bayes
False Name
May 27, 2023, 7:35 PM
−3
points
2
comments
16
min read
LW
link
Should Rational Animations invite viewers to read content on LessWrong?
Writer
May 27, 2023, 7:26 PM
40
points
9
comments
3
min read
LW
link
Who are the Experts on Cryonics?
Mati_Roy
May 27, 2023, 7:24 PM
30
points
9
comments
1
min read
LW
link
(biostasis.substack.com)
AI and Planet Earth are incompatible.
archeon
May 27, 2023, 6:59 PM
−4
points
2
comments
1
min read
LW
link
South Bay ACX/LW Meetup
IS
May 27, 2023, 5:25 PM
2
points
0
comments
1
min read
LW
link
Hands-On Experience Is Not Magic
Thane Ruthenis
May 27, 2023, 4:57 PM
22
points
14
comments
5
min read
LW
link
Is Deontological AI Safe? [Feedback Draft]
Dan H
and
William D'Alessandro
May 27, 2023, 4:39 PM
19
points
15
comments
20
min read
LW
link
San Francisco ACX Meetup “First Saturday” June 3, 1 pm
guenael
May 27, 2023, 1:58 PM
1
point
0
comments
1
min read
LW
link
Papers on protein design
alexlyzhov
May 27, 2023, 1:18 AM
9
points
0
comments
3
min read
LW
link
D&D.Sci 5E: Return of the League of Defenders
aphyer
May 26, 2023, 8:39 PM
42
points
11
comments
3
min read
LW
link
Seeking (Paid) Case Studies on Standards
HoldenKarnofsky
May 26, 2023, 5:58 PM
69
points
9
comments
11
min read
LW
link
Conditional Prediction with Zero-Sum Training Solves Self-Fulfilling Prophecies
Rubi J. Hudson
and
Johannes Treutlein
May 26, 2023, 5:44 PM
88
points
13
comments
24
min read
LW
link
Request: stop advancing AI capabilities
So8res
May 26, 2023, 5:42 PM
154
points
24
comments
1
min read
LW
link
Bandgaps, Brains, and Bioweapons: The limitations of computational science and what it means for AGI
titotal
May 26, 2023, 3:57 PM
36
points
20
comments
LW
link
The American Information Revolution in Global Perspective
jasoncrawford
May 26, 2023, 12:39 PM
16
points
1
comment
5
min read
LW
link
(rootsofprogress.org)
Helio-Selenic Laser Telescope (in SPACE!?)
Alexander Gietelink Oldenziel
May 26, 2023, 11:24 AM
8
points
2
comments
4
min read
LW
link
[Question]
Why is violence against AI labs a taboo?
ArisC
May 26, 2023, 8:00 AM
−21
points
63
comments
1
min read
LW
link
Where do you lie on two axes of world manipulability?
Max H
May 26, 2023, 3:04 AM
31
points
15
comments
3
min read
LW
link
Some thoughts on automating alignment research
Lukas Finnveden
May 26, 2023, 1:50 AM
30
points
4
comments
6
min read
LW
link
[Question]
What’s your viewpoint on the likelihood of GPT-5 being able to autonomously create, train, and implement an AI superior to GPT-5?
Super AGI
May 26, 2023, 1:43 AM
7
points
15
comments
1
min read
LW
link
Before smart AI, there will be many mediocre or specialized AIs
Lukas Finnveden
May 26, 2023, 1:38 AM
58
points
14
comments
9
min read
LW
link
1
review
how humans are aligned
bhauth
May 26, 2023, 12:09 AM
14
points
3
comments
1
min read
LW
link
[Question]
What vegan food resources have you found useful?
Elizabeth
May 25, 2023, 10:46 PM
29
points
6
comments
LW
link
Mob and Bailey
Screwtape
May 25, 2023, 10:14 PM
82
points
17
comments
7
min read
LW
link
1
review
Look At What’s In Front Of You (Conclusion to The Nuts and Bolts of Naturalism)
LoganStrohl
May 25, 2023, 7:00 PM
50
points
1
comment
2
min read
LW
link
[Market] Will AI xrisk seem to be handled seriously by the end of 2026?
tailcalled
May 25, 2023, 6:51 PM
15
points
2
comments
1
min read
LW
link
(manifold.markets)
[Question]
What should my college major be if I want to do AI alignment research?
metachirality
May 25, 2023, 6:23 PM
8
points
7
comments
1
min read
LW
link
Is behavioral safety “solved” in non-adversarial conditions?
Robert_AIZI
May 25, 2023, 5:56 PM
26
points
8
comments
2
min read
LW
link
(aizi.substack.com)
Book Review: How Minds Change
bc4026bd4aaa5b7fe
May 25, 2023, 5:55 PM
313
points
52
comments
15
min read
LW
link
Self-administered EMDR without a therapist is very useful for a lot of things!
EternallyBlissful
May 25, 2023, 5:54 PM
49
points
12
comments
11
min read
LW
link
RecurrentGPT: a loom-type tool with a twist
mishka
May 25, 2023, 5:09 PM
10
points
0
comments
3
min read
LW
link
(arxiv.org)
The Genie in the Bottle: An Introduction to AI Alignment and Risk
Snorkelfarsan
May 25, 2023, 4:30 PM
5
points
1
comment
25
min read
LW
link
AI #13: Potential Algorithmic Improvements
Zvi
May 25, 2023, 3:40 PM
45
points
4
comments
67
min read
LW
link
(thezvi.wordpress.com)
Solving the Mechanistic Interpretability challenges: EIS VII Challenge 2
StefanHex
and
Marius Hobbhahn
25 May 2023 15:37 UTC
71
points
1
comment
13
min read
LW
link
Malthusian Competition (not as bad as it seems)
Logan Zoellner
25 May 2023 15:30 UTC
6
points
11
comments
2
min read
LW
link
You Don’t Always Need Indexes
jefftk
25 May 2023 14:20 UTC
22
points
6
comments
1
min read
LW
link
(www.jefftk.com)
Theories of Biological Inspiration
Eric Zhang
25 May 2023 13:07 UTC
7
points
3
comments
1
min read
LW
link
Evaluating strategic reasoning in GPT models
phelps-sg
25 May 2023 11:51 UTC
4
points
1
comment
8
min read
LW
link
Requirements for a STEM-capable AGI Value Learner (my Case for Less Doom)
RogerDearnaley
25 May 2023 9:26 UTC
33
points
3
comments
15
min read
LW
link
Back to top
Next
N
W
F
A
C
D
E
F
G
H
I
Customize appearance
Current theme:
default
A
C
D
E
F
G
H
I
Less Wrong (text)
Less Wrong (link)
Invert colors
Reset to defaults
OK
Cancel
Hi, I’m Bobby the Basilisk! Click on the minimize button (
) to minimize the theme tweaker window, so that you can see what the page looks like with the current tweaked values. (But remember,
the changes won’t be saved until you click “OK”!
)
Theme tweaker help
Show Bobby the Basilisk
OK
Cancel