Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
Archive
Sequences
About
Search
Log In
All
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
2025
All
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
All
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
Page
1
D&D.Sci 5E: Return of the League of Defenders
aphyer
May 26, 2023, 8:39 PM
42
points
11
comments
3
min read
LW
link
Seeking (Paid) Case Studies on Standards
HoldenKarnofsky
May 26, 2023, 5:58 PM
69
points
9
comments
11
min read
LW
link
Conditional Prediction with Zero-Sum Training Solves Self-Fulfilling Prophecies
Rubi J. Hudson
and
Johannes Treutlein
May 26, 2023, 5:44 PM
88
points
13
comments
24
min read
LW
link
Request: stop advancing AI capabilities
So8res
May 26, 2023, 5:42 PM
154
points
24
comments
1
min read
LW
link
Bandgaps, Brains, and Bioweapons: The limitations of computational science and what it means for AGI
titotal
May 26, 2023, 3:57 PM
36
points
20
comments
LW
link
The American Information Revolution in Global Perspective
jasoncrawford
May 26, 2023, 12:39 PM
16
points
1
comment
5
min read
LW
link
(rootsofprogress.org)
Helio-Selenic Laser Telescope (in SPACE!?)
Alexander Gietelink Oldenziel
May 26, 2023, 11:24 AM
8
points
2
comments
4
min read
LW
link
[Question]
Why is violence against AI labs a taboo?
ArisC
May 26, 2023, 8:00 AM
−21
points
63
comments
1
min read
LW
link
Where do you lie on two axes of world manipulability?
Max H
May 26, 2023, 3:04 AM
31
points
15
comments
3
min read
LW
link
Some thoughts on automating alignment research
Lukas Finnveden
May 26, 2023, 1:50 AM
30
points
4
comments
6
min read
LW
link
[Question]
What’s your viewpoint on the likelihood of GPT-5 being able to autonomously create, train, and implement an AI superior to GPT-5?
Super AGI
May 26, 2023, 1:43 AM
7
points
15
comments
1
min read
LW
link
Before smart AI, there will be many mediocre or specialized AIs
Lukas Finnveden
May 26, 2023, 1:38 AM
58
points
14
comments
9
min read
LW
link
1
review
how humans are aligned
bhauth
May 26, 2023, 12:09 AM
14
points
3
comments
1
min read
LW
link
[Question]
What vegan food resources have you found useful?
Elizabeth
May 25, 2023, 10:46 PM
29
points
6
comments
LW
link
Mob and Bailey
Screwtape
May 25, 2023, 10:14 PM
82
points
17
comments
7
min read
LW
link
1
review
Look At What’s In Front Of You (Conclusion to The Nuts and Bolts of Naturalism)
LoganStrohl
May 25, 2023, 7:00 PM
50
points
1
comment
2
min read
LW
link
[Market] Will AI xrisk seem to be handled seriously by the end of 2026?
tailcalled
May 25, 2023, 6:51 PM
15
points
2
comments
1
min read
LW
link
(manifold.markets)
[Question]
What should my college major be if I want to do AI alignment research?
metachirality
May 25, 2023, 6:23 PM
8
points
7
comments
1
min read
LW
link
Is behavioral safety “solved” in non-adversarial conditions?
Robert_AIZI
May 25, 2023, 5:56 PM
26
points
8
comments
2
min read
LW
link
(aizi.substack.com)
Book Review: How Minds Change
bc4026bd4aaa5b7fe
May 25, 2023, 5:55 PM
313
points
52
comments
15
min read
LW
link
Self-administered EMDR without a therapist is very useful for a lot of things!
EternallyBlissful
May 25, 2023, 5:54 PM
49
points
12
comments
11
min read
LW
link
RecurrentGPT: a loom-type tool with a twist
mishka
May 25, 2023, 5:09 PM
10
points
0
comments
3
min read
LW
link
(arxiv.org)
The Genie in the Bottle: An Introduction to AI Alignment and Risk
Snorkelfarsan
May 25, 2023, 4:30 PM
5
points
1
comment
25
min read
LW
link
AI #13: Potential Algorithmic Improvements
Zvi
May 25, 2023, 3:40 PM
45
points
4
comments
67
min read
LW
link
(thezvi.wordpress.com)
Solving the Mechanistic Interpretability challenges: EIS VII Challenge 2
StefanHex
and
Marius Hobbhahn
May 25, 2023, 3:37 PM
71
points
1
comment
13
min read
LW
link
Malthusian Competition (not as bad as it seems)
Logan Zoellner
May 25, 2023, 3:30 PM
6
points
11
comments
2
min read
LW
link
You Don’t Always Need Indexes
jefftk
May 25, 2023, 2:20 PM
22
points
6
comments
1
min read
LW
link
(www.jefftk.com)
Theories of Biological Inspiration
Eric Zhang
May 25, 2023, 1:07 PM
7
points
3
comments
1
min read
LW
link
Evaluating strategic reasoning in GPT models
phelps-sg
May 25, 2023, 11:51 AM
4
points
1
comment
8
min read
LW
link
Requirements for a STEM-capable AGI Value Learner (my Case for Less Doom)
RogerDearnaley
May 25, 2023, 9:26 AM
33
points
3
comments
15
min read
LW
link
Alignment solutions for weak AI don’t (necessarily) scale to strong AI
Michael Tontchev
May 25, 2023, 8:26 AM
6
points
0
comments
5
min read
LW
link
[Question]
What features would you like to see in a personal forcasting / prediction tracking app?
regnarg
May 25, 2023, 8:18 AM
9
points
0
comments
1
min read
LW
link
Announcing the Confido app: bringing forecasting to everyone
regnarg
May 25, 2023, 8:18 AM
6
points
2
comments
10
min read
LW
link
(forum.effectivealtruism.org)
But What If We Actually Want To Maximize Paperclips?
snerx
May 25, 2023, 7:13 AM
−17
points
6
comments
7
min read
LW
link
DeepMind: Model evaluation for extreme risks
Zach Stein-Perlman
May 25, 2023, 3:00 AM
94
points
12
comments
1
min read
LW
link
1
review
(arxiv.org)
Why I’m Not (Yet) A Full-Time Technical Alignment Researcher
Nicholas / Heather Kross
May 25, 2023, 1:26 AM
41
points
21
comments
4
min read
LW
link
(www.thinkingmuchbetter.com)
Two ideas for alignment, perpetual mutual distrust and induction
APaleBlueDot
May 25, 2023, 12:56 AM
1
point
2
comments
4
min read
LW
link
Evaluating Evidence Reconstructions of Mock Crimes -Submission 2
Alan E Dunne
May 24, 2023, 10:17 PM
−1
points
1
comment
3
min read
LW
link
[Linkpost] Interpretability Dreams
DanielFilan
May 24, 2023, 9:08 PM
39
points
2
comments
2
min read
LW
link
(transformer-circuits.pub)
Rishi Sunak mentions “existential threats” in talk with OpenAI, DeepMind, Anthropic CEOs
Arjun Panickssery
,
Baldassare Castiglione
and
Cleo Nardo
May 24, 2023, 9:06 PM
34
points
1
comment
1
min read
LW
link
(www.gov.uk)
If you’re not a morning person, consider quitting allergy pills
Brendan Long
May 24, 2023, 8:11 PM
8
points
3
comments
1
min read
LW
link
Adumbrations on AGI from an outsider
nicholashalden
May 24, 2023, 5:41 PM
57
points
44
comments
8
min read
LW
link
(nicholashalden.home.blog)
Open Thread With Experimental Feature: Reactions
jimrandomh
May 24, 2023, 4:46 PM
101
points
189
comments
3
min read
LW
link
A rejection of the Orthogonality Thesis
ArisC
May 24, 2023, 4:37 PM
−2
points
11
comments
2
min read
LW
link
(medium.com)
Aligned AI via monitoring objectives in AutoGPT-like systems
Paul Colognese
May 24, 2023, 3:59 PM
27
points
4
comments
4
min read
LW
link
The Office of Science and Technology Policy put out a request for information on A.I.
HiroSakuraba
May 24, 2023, 1:33 PM
59
points
4
comments
1
min read
LW
link
(www.whitehouse.gov)
ChatGPT (May 2023) on Designing Friendly Superintelligence
Mitchell_Porter
May 24, 2023, 10:47 AM
5
points
0
comments
1
min read
LW
link
(singularitypolitics.wordpress.com)
No—AI is just as energy-efficient as your brain.
Maxwell Clarke
May 24, 2023, 2:30 AM
11
points
7
comments
1
min read
LW
link
[Question]
What projects and efforts are there to promote AI safety research?
Christopher King
May 24, 2023, 12:33 AM
4
points
0
comments
1
min read
LW
link
My May 2023 priorities for AI x-safety: more empathy, more unification of concerns, and less vilification of OpenAI
Andrew_Critch
May 24, 2023, 12:02 AM
268
points
39
comments
8
min read
LW
link
Back to top
Next
N
W
F
A
C
D
E
F
G
H
I
Customize appearance
Current theme:
default
A
C
D
E
F
G
H
I
Less Wrong (text)
Less Wrong (link)
Invert colors
Reset to defaults
OK
Cancel
Hi, I’m Bobby the Basilisk! Click on the minimize button (
) to minimize the theme tweaker window, so that you can see what the page looks like with the current tweaked values. (But remember,
the changes won’t be saved until you click “OK”!
)
Theme tweaker help
Show Bobby the Basilisk
OK
Cancel