Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
Archive
Sequences
About
Search
Log In
All
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
2025
All
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
All
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
Page
2
Alignment is not enough
Alan Chan
Jan 12, 2023, 12:33 AM
12
points
6
comments
11
min read
LW
link
(coordination.substack.com)
How it feels to have your mind hacked by an AI
blaked
Jan 12, 2023, 12:33 AM
367
points
222
comments
17
min read
LW
link
Categorical-measure-theoretic approach to optimal policies tending to seek power
jacek
Jan 12, 2023, 12:32 AM
31
points
3
comments
6
min read
LW
link
Any person/mind should have the right to suicide
askofa
Jan 12, 2023, 12:32 AM
14
points
13
comments
2
min read
LW
link
Have we really forsaken natural selection?
KatjaGrace
Jan 12, 2023, 12:10 AM
34
points
7
comments
2
min read
LW
link
(worldspiritsockpuppet.com)
[Question]
Using Finite Factored Sets for Causal Representation Learning?
David Reber
Jan 11, 2023, 10:06 PM
2
points
3
comments
1
min read
LW
link
GWWC’s Handling of Conflicting Funding Bars
jefftk
Jan 11, 2023, 8:30 PM
19
points
0
comments
3
min read
LW
link
(www.jefftk.com)
How to write a big cartesian product symbol in MathJax
Matthias G. Mayer
Jan 11, 2023, 8:21 PM
8
points
1
comment
1
min read
LW
link
What’s the deal with AI consciousness?
TW123
Jan 11, 2023, 4:37 PM
6
points
13
comments
9
min read
LW
link
(aiwatchtower.substack.com)
[Question]
Any significant updates on long covid risk analysis?
Randomized, Controlled
Jan 11, 2023, 2:31 PM
23
points
11
comments
1
min read
LW
link
internal in nonstandard analysis
Alok Singh
Jan 11, 2023, 9:58 AM
9
points
1
comment
1
min read
LW
link
Compounding Resource X
Raemon
Jan 11, 2023, 3:14 AM
77
points
6
comments
9
min read
LW
link
Running With a Backpack
jefftk
Jan 11, 2023, 3:00 AM
19
points
11
comments
1
min read
LW
link
(www.jefftk.com)
A simple thought experiment showing why recessions are an unnecessary bug in our economic system
skogsnisse
Jan 11, 2023, 12:43 AM
1
point
1
comment
1
min read
LW
link
We don’t trade with ants
KatjaGrace
Jan 10, 2023, 11:50 PM
272
points
109
comments
7
min read
LW
link
1
review
(worldspiritsockpuppet.com)
[Question]
Who are the people who are currently profiting from inflation?
skogsnisse
Jan 10, 2023, 9:39 PM
1
point
2
comments
1
min read
LW
link
Is Progress Real?
rogersbacon
Jan 10, 2023, 5:42 PM
5
points
14
comments
14
min read
LW
link
(www.secretorum.life)
200 COP in MI: Interpreting Reinforcement Learning
Neel Nanda
Jan 10, 2023, 5:37 PM
25
points
1
comment
10
min read
LW
link
AGI and the EMH: markets are not expecting aligned or unaligned AI in the next 30 years
basil.halperin
,
J. Zachary Mazlish
and
tmychow
Jan 10, 2023, 4:06 PM
119
points
44
comments
26
min read
LW
link
The Alignment Problem from a Deep Learning Perspective (major rewrite)
SoerenMind
,
Richard_Ngo
and
LawrenceC
Jan 10, 2023, 4:06 PM
84
points
8
comments
39
min read
LW
link
(arxiv.org)
Against using stock prices to forecast AI timelines
basil.halperin
,
tmychow
and
J. Zachary Mazlish
Jan 10, 2023, 4:03 PM
23
points
2
comments
2
min read
LW
link
Sorting Pebbles Into Correct Heaps: The Animation
Writer
Jan 10, 2023, 3:58 PM
26
points
2
comments
1
min read
LW
link
(youtu.be)
Escape Velocity from Bullshit Jobs
Zvi
Jan 10, 2023, 2:30 PM
61
points
18
comments
5
min read
LW
link
(thezvi.wordpress.com)
Scaling laws vs individual differences
beren
Jan 10, 2023, 1:22 PM
45
points
21
comments
7
min read
LW
link
Notes on writing
RP
Jan 10, 2023, 4:01 AM
35
points
11
comments
3
min read
LW
link
Idea: Learning How To Move Towards The Metagame
Algon
Jan 10, 2023, 12:58 AM
10
points
3
comments
1
min read
LW
link
Review AI Alignment posts to help figure out how to make a proper AI Alignment review
habryka
and
Raemon
Jan 10, 2023, 12:19 AM
85
points
31
comments
2
min read
LW
link
Against the paradox of tolerance
pchvykov
Jan 10, 2023, 12:12 AM
1
point
11
comments
3
min read
LW
link
Increased Scam Quality/Quantity (Hypothesis in need of data)?
Beeblebrox
Jan 9, 2023, 10:57 PM
9
points
6
comments
1
min read
LW
link
Wentworth and Larsen on buying time
Orpheus16
,
Thomas Larsen
and
johnswentworth
Jan 9, 2023, 9:31 PM
74
points
6
comments
12
min read
LW
link
EA & LW Forum Summaries—Holiday Edition (19th Dec − 8th Jan)
Zoe Williams
Jan 9, 2023, 9:06 PM
11
points
0
comments
LW
link
GWWC Should Require Public Charity Evaluations
jefftk
Jan 9, 2023, 8:10 PM
28
points
0
comments
4
min read
LW
link
(www.jefftk.com)
[MLSN #7]: an example of an emergent internal optimizer
joshc
and
Dan H
Jan 9, 2023, 7:39 PM
28
points
0
comments
6
min read
LW
link
Trying to isolate objectives: approaches toward high-level interpretability
Jozdien
Jan 9, 2023, 6:33 PM
49
points
14
comments
8
min read
LW
link
The special nature of special relativity
adamShimi
Jan 9, 2023, 5:30 PM
37
points
1
comment
3
min read
LW
link
(epistemologicalvigilance.substack.com)
Pierre Menard, pixel art, and entropy
Joey Marcellino
Jan 9, 2023, 4:34 PM
1
point
1
comment
6
min read
LW
link
Forecasting extreme outcomes
AidanGoth
Jan 9, 2023, 4:34 PM
4
points
1
comment
2
min read
LW
link
(docs.google.com)
Evidence under Adversarial Conditions
PeterMcCluskey
Jan 9, 2023, 4:21 PM
57
points
1
comment
3
min read
LW
link
(bayesianinvestor.com)
How to Bounded Distrust
Zvi
Jan 9, 2023, 1:10 PM
122
points
17
comments
4
min read
LW
link
1
review
(thezvi.wordpress.com)
Reification bias
adamShimi
and
Gabriel Alfour
Jan 9, 2023, 12:22 PM
25
points
6
comments
2
min read
LW
link
Big list of AI safety videos
JakubK
Jan 9, 2023, 6:12 AM
11
points
2
comments
1
min read
LW
link
(docs.google.com)
Rationality Practice: Self-Deception
Darmani
9 Jan 2023 4:07 UTC
6
points
0
comments
1
min read
LW
link
Wolf Incident Postmortem
jefftk
9 Jan 2023 3:20 UTC
137
points
13
comments
1
min read
LW
link
(www.jefftk.com)
You’re Not One “You”—How Decision Theories Are Talking Past Each Other
keith_wynroe
9 Jan 2023 1:21 UTC
28
points
11
comments
8
min read
LW
link
On Blogging and Podcasting
DanielFilan
9 Jan 2023 0:40 UTC
18
points
6
comments
11
min read
LW
link
(danielfilan.com)
ChatGPT tells stories about XP-708-DQ, Eliezer, dragons, dark sorceresses, and unaligned robots becoming aligned
Bill Benzon
8 Jan 2023 23:21 UTC
6
points
2
comments
18
min read
LW
link
Simulacra are Things
janus
8 Jan 2023 23:03 UTC
63
points
7
comments
2
min read
LW
link
[Question]
GPT learning from smarter texts?
Viliam
8 Jan 2023 22:23 UTC
26
points
7
comments
1
min read
LW
link
Latent variable prediction markets mockup + designer request
tailcalled
8 Jan 2023 22:18 UTC
25
points
4
comments
1
min read
LW
link
Citability of Lesswrong and the Alignment Forum
Leon Lang
8 Jan 2023 22:12 UTC
48
points
2
comments
1
min read
LW
link
Previous
Back to top
Next
N
W
F
A
C
D
E
F
G
H
I
Customize appearance
Current theme:
default
A
C
D
E
F
G
H
I
Less Wrong (text)
Less Wrong (link)
Invert colors
Reset to defaults
OK
Cancel
Hi, I’m Bobby the Basilisk! Click on the minimize button (
) to minimize the theme tweaker window, so that you can see what the page looks like with the current tweaked values. (But remember,
the changes won’t be saved until you click “OK”!
)
Theme tweaker help
Show Bobby the Basilisk
OK
Cancel