Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
Archive
Sequences
About
Search
Log In
All
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
2025
All
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
All
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
Page
3
Leveling Up Or Leveling Off? Understanding The Science Behind Skill Plateaus
lynettebye
Jun 16, 2023, 12:18 AM
45
points
9
comments
18
min read
LW
link
If you are too stressed, walk away from the front lines
Neil
Jun 12, 2023, 2:26 PM
44
points
14
comments
5
min read
LW
link
How tall is the Shard, really?
philh
Jun 23, 2023, 8:10 AM
44
points
10
comments
9
min read
LW
link
(reasonableapproximation.net)
A summary of current work in AI governance
constructive
Jun 17, 2023, 6:41 PM
44
points
1
comment
11
min read
LW
link
(forum.effectivealtruism.org)
resolving some neural network mysteries
bhauth
Jun 19, 2023, 12:09 AM
44
points
6
comments
2
min read
LW
link
(www.bhauth.com)
On the Apple Vision Pro
Zvi
Jun 14, 2023, 5:50 PM
44
points
17
comments
11
min read
LW
link
(thezvi.wordpress.com)
Anthropically Blind: the anthropic shadow is reflectively inconsistent
Christopher King
Jun 29, 2023, 2:36 AM
43
points
40
comments
10
min read
LW
link
One implementation of regulatory GPU restrictions
porby
Jun 4, 2023, 8:34 PM
42
points
6
comments
5
min read
LW
link
The (local) unit of intelligence is FLOPs
boazbarak
Jun 5, 2023, 6:23 PM
42
points
7
comments
5
min read
LW
link
Unfaithful Explanations in Chain-of-Thought Prompting
Miles Turpin
Jun 3, 2023, 12:22 AM
42
points
8
comments
7
min read
LW
link
Cryonics Career Survey (more jobs than you think)
Mati_Roy
Jun 18, 2023, 2:13 AM
41
points
1
comment
2
min read
LW
link
Nature: “Stop talking about tomorrow’s AI doomsday when AI poses risks today”
Ben Smith
Jun 28, 2023, 5:59 AM
40
points
8
comments
2
min read
LW
link
(www.nature.com)
Dreams of “Mathopedia”
Nicholas / Heather Kross
Jun 2, 2023, 1:30 AM
40
points
16
comments
2
min read
LW
link
(www.thinkingmuchbetter.com)
Catastrophic Risks from AI #1: Introduction
Dan H
,
Mantas Mazeika
and
TW123
Jun 22, 2023, 5:09 PM
40
points
1
comment
5
min read
LW
link
(arxiv.org)
AI-Plans.com—a contributable compendium
Iknownothing
Jun 25, 2023, 2:40 PM
39
points
7
comments
4
min read
LW
link
(ai-plans.com)
[Question]
What money-pumps exist, if any, for deontologists?
Daniel Kokotajlo
Jun 28, 2023, 7:08 PM
39
points
35
comments
1
min read
LW
link
Bengio’s FAQ on Catastrophic AI Risks
Vaniver
Jun 29, 2023, 11:04 PM
39
points
0
comments
1
min read
LW
link
(yoshuabengio.org)
AISC team report: Soft-optimization, Bayes and Goodhart
Simon Fischer
,
benjaminko
,
jazcarretao
,
DFNaiff
and
Jeremy Gillen
Jun 27, 2023, 6:05 AM
38
points
2
comments
15
min read
LW
link
Metaphors for AI, and why I don’t like them
boazbarak
Jun 28, 2023, 10:47 PM
38
points
18
comments
12
min read
LW
link
Catastrophic Risks from AI #2: Malicious Use
Dan H
,
Mantas Mazeika
and
TW123
Jun 22, 2023, 5:10 PM
38
points
1
comment
17
min read
LW
link
(arxiv.org)
Correctly Calibrated Trust
habryka
Jun 24, 2023, 7:48 PM
38
points
3
comments
11
min read
LW
link
(forum.effectivealtruism.org)
Solomonoff induction still works if the universe is uncomputable, and its usefulness doesn’t require knowing Occam’s razor
Christopher King
Jun 18, 2023, 1:52 AM
38
points
28
comments
4
min read
LW
link
The Sharp Right Turn: sudden deceptive alignment as a convergent goal
avturchin
Jun 6, 2023, 9:59 AM
38
points
5
comments
1
min read
LW
link
Wildfire of strategicness
TsviBT
Jun 5, 2023, 1:59 PM
38
points
19
comments
1
min read
LW
link
Why I am not a longtermist (May 2022)
boazbarak
Jun 6, 2023, 8:36 PM
38
points
19
comments
9
min read
LW
link
(windowsontheory.org)
Aura as a proprioceptive glitch
pchvykov
Jun 12, 2023, 7:30 PM
37
points
4
comments
4
min read
LW
link
<$750k grants for General Purpose AI Assurance/Safety Research
Phosphorous
Jun 13, 2023, 4:45 AM
37
points
1
comment
1
min read
LW
link
(cset.georgetown.edu)
Society Library seeking contributions for canonical AI Safety debate map
Jarred Filmer
Jun 6, 2023, 6:15 PM
36
points
0
comments
1
min read
LW
link
(www.societylibrary.org)
Why libertarians are advocating for regulation on AI
RobertM
Jun 14, 2023, 8:59 PM
36
points
13
comments
4
min read
LW
link
[Linkpost] Large Language Models Converge on Brain-Like Word Representations
Bogdan Ionut Cirstea
Jun 11, 2023, 11:20 AM
36
points
12
comments
1
min read
LW
link
“Natural is better” is a valuable heuristic
Neil
Jun 20, 2023, 10:25 PM
35
points
16
comments
4
min read
LW
link
The Dictatorship Problem
alyssavance
Jun 11, 2023, 2:45 AM
35
points
145
comments
11
min read
LW
link
10 quick takes about AGI
Max H
Jun 20, 2023, 2:22 AM
35
points
17
comments
7
min read
LW
link
Scaffolded LLMs: Less Obvious Concerns
Stephen Fowler
Jun 16, 2023, 10:39 AM
34
points
15
comments
14
min read
LW
link
Anthropic | Charting a Path to AI Accountability
Gabe M
Jun 14, 2023, 4:43 AM
34
points
2
comments
3
min read
LW
link
(www.anthropic.com)
Experiments in Evaluating Steering Vectors
Gytis Daujotas
Jun 19, 2023, 3:11 PM
34
points
4
comments
4
min read
LW
link
Meta-conversation shouldn’t be taboo
Adam Zerner
Jun 5, 2023, 12:19 AM
34
points
36
comments
4
min read
LW
link
The AGI Race Between the US and China Doesn’t Exist.
Eva_B
Jun 3, 2023, 12:22 AM
33
points
15
comments
7
min read
LW
link
(evabehrens.substack.com)
Epistemic spot checking one claim in The Precipice
Isaac King
Jun 27, 2023, 1:03 AM
33
points
3
comments
1
min read
LW
link
Announcing AISafety.info’s Write-a-thon (June 16-18) and Second Distillation Fellowship (July 3-October 2)
steven0461
Jun 3, 2023, 2:03 AM
33
points
1
comment
2
min read
LW
link
Intelligence Officials Say U.S. Has Retrieved Craft of Non-Human Origin
lc
Jun 6, 2023, 3:54 AM
33
points
151
comments
1
min read
LW
link
(thedebrief.org)
Multiple stages of fallacy—justifications and non-justifications for the multiple stage fallacy
AronT
13 Jun 2023 17:37 UTC
33
points
2
comments
5
min read
LW
link
(coordinationishard.substack.com)
Transformative AGI by 2043 is <1% likely
Ted Sanders
6 Jun 2023 17:36 UTC
33
points
117
comments
5
min read
LW
link
(arxiv.org)
On the Cost of Thriving Index
Zvi
26 Jun 2023 15:30 UTC
33
points
6
comments
9
min read
LW
link
(thezvi.wordpress.com)
Never Fight The Last War
ChristianKl
20 Jun 2023 12:35 UTC
32
points
4
comments
1
min read
LW
link
“LLMs Don’t Have a Coherent Model of the World”—What it Means, Why it Matters
Davidmanheim
1 Jun 2023 7:46 UTC
32
points
2
comments
7
min read
LW
link
Andrew Ng wants to have a conversation about extinction risk from AI
Leon Lang
5 Jun 2023 22:29 UTC
32
points
2
comments
1
min read
LW
link
(twitter.com)
UK PM: $125M for AI safety
Hauke Hillebrandt
12 Jun 2023 12:33 UTC
31
points
11
comments
1
min read
LW
link
(twitter.com)
Park Toys
jefftk
23 Jun 2023 16:00 UTC
31
points
5
comments
1
min read
LW
link
(www.jefftk.com)
Philosophical Cyborg (Part 1)
ukc10014
,
Roman Leventov
and
NicholasKees
14 Jun 2023 16:20 UTC
31
points
4
comments
13
min read
LW
link
Back to first
Previous
Back to top
Next
N
W
F
A
C
D
E
F
G
H
I
Customize appearance
Current theme:
default
A
C
D
E
F
G
H
I
Less Wrong (text)
Less Wrong (link)
Invert colors
Reset to defaults
OK
Cancel
Hi, I’m Bobby the Basilisk! Click on the minimize button (
) to minimize the theme tweaker window, so that you can see what the page looks like with the current tweaked values. (But remember,
the changes won’t be saved until you click “OK”!
)
Theme tweaker help
Show Bobby the Basilisk
OK
Cancel