Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
Archive
Sequences
About
Search
Log In
All
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
2025
All
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
All
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
Page
2
internal in nonstandard analysis
Alok Singh
Jan 11, 2023, 9:58 AM
9
points
1
comment
1
min read
LW
link
Compounding Resource X
Raemon
Jan 11, 2023, 3:14 AM
77
points
6
comments
9
min read
LW
link
Running With a Backpack
jefftk
Jan 11, 2023, 3:00 AM
19
points
11
comments
1
min read
LW
link
(www.jefftk.com)
A simple thought experiment showing why recessions are an unnecessary bug in our economic system
skogsnisse
Jan 11, 2023, 12:43 AM
1
point
1
comment
1
min read
LW
link
We don’t trade with ants
KatjaGrace
Jan 10, 2023, 11:50 PM
272
points
109
comments
7
min read
LW
link
1
review
(worldspiritsockpuppet.com)
[Question]
Who are the people who are currently profiting from inflation?
skogsnisse
Jan 10, 2023, 9:39 PM
1
point
2
comments
1
min read
LW
link
Is Progress Real?
rogersbacon
Jan 10, 2023, 5:42 PM
5
points
14
comments
14
min read
LW
link
(www.secretorum.life)
200 COP in MI: Interpreting Reinforcement Learning
Neel Nanda
Jan 10, 2023, 5:37 PM
25
points
1
comment
10
min read
LW
link
AGI and the EMH: markets are not expecting aligned or unaligned AI in the next 30 years
basil.halperin
,
J. Zachary Mazlish
and
tmychow
Jan 10, 2023, 4:06 PM
119
points
44
comments
26
min read
LW
link
The Alignment Problem from a Deep Learning Perspective (major rewrite)
SoerenMind
,
Richard_Ngo
and
LawrenceC
Jan 10, 2023, 4:06 PM
84
points
8
comments
39
min read
LW
link
(arxiv.org)
Against using stock prices to forecast AI timelines
basil.halperin
,
tmychow
and
J. Zachary Mazlish
Jan 10, 2023, 4:03 PM
23
points
2
comments
2
min read
LW
link
Sorting Pebbles Into Correct Heaps: The Animation
Writer
Jan 10, 2023, 3:58 PM
26
points
2
comments
1
min read
LW
link
(youtu.be)
Escape Velocity from Bullshit Jobs
Zvi
Jan 10, 2023, 2:30 PM
61
points
18
comments
5
min read
LW
link
(thezvi.wordpress.com)
Scaling laws vs individual differences
beren
Jan 10, 2023, 1:22 PM
45
points
21
comments
7
min read
LW
link
Notes on writing
RP
Jan 10, 2023, 4:01 AM
35
points
11
comments
3
min read
LW
link
Idea: Learning How To Move Towards The Metagame
Algon
Jan 10, 2023, 12:58 AM
10
points
3
comments
1
min read
LW
link
Review AI Alignment posts to help figure out how to make a proper AI Alignment review
habryka
and
Raemon
Jan 10, 2023, 12:19 AM
85
points
31
comments
2
min read
LW
link
Against the paradox of tolerance
pchvykov
Jan 10, 2023, 12:12 AM
1
point
11
comments
3
min read
LW
link
Increased Scam Quality/Quantity (Hypothesis in need of data)?
Beeblebrox
Jan 9, 2023, 10:57 PM
9
points
6
comments
1
min read
LW
link
Wentworth and Larsen on buying time
Orpheus16
,
Thomas Larsen
and
johnswentworth
Jan 9, 2023, 9:31 PM
74
points
6
comments
12
min read
LW
link
EA & LW Forum Summaries—Holiday Edition (19th Dec − 8th Jan)
Zoe Williams
Jan 9, 2023, 9:06 PM
11
points
0
comments
LW
link
GWWC Should Require Public Charity Evaluations
jefftk
Jan 9, 2023, 8:10 PM
28
points
0
comments
4
min read
LW
link
(www.jefftk.com)
[MLSN #7]: an example of an emergent internal optimizer
joshc
and
Dan H
Jan 9, 2023, 7:39 PM
28
points
0
comments
6
min read
LW
link
Trying to isolate objectives: approaches toward high-level interpretability
Jozdien
Jan 9, 2023, 6:33 PM
49
points
14
comments
8
min read
LW
link
The special nature of special relativity
adamShimi
Jan 9, 2023, 5:30 PM
37
points
1
comment
3
min read
LW
link
(epistemologicalvigilance.substack.com)
Pierre Menard, pixel art, and entropy
Joey Marcellino
Jan 9, 2023, 4:34 PM
1
point
1
comment
6
min read
LW
link
Forecasting extreme outcomes
AidanGoth
Jan 9, 2023, 4:34 PM
4
points
1
comment
2
min read
LW
link
(docs.google.com)
Evidence under Adversarial Conditions
PeterMcCluskey
Jan 9, 2023, 4:21 PM
57
points
1
comment
3
min read
LW
link
(bayesianinvestor.com)
How to Bounded Distrust
Zvi
Jan 9, 2023, 1:10 PM
122
points
17
comments
4
min read
LW
link
1
review
(thezvi.wordpress.com)
Reification bias
adamShimi
and
Gabriel Alfour
Jan 9, 2023, 12:22 PM
25
points
6
comments
2
min read
LW
link
Big list of AI safety videos
JakubK
Jan 9, 2023, 6:12 AM
11
points
2
comments
1
min read
LW
link
(docs.google.com)
Rationality Practice: Self-Deception
Darmani
Jan 9, 2023, 4:07 AM
6
points
0
comments
1
min read
LW
link
Wolf Incident Postmortem
jefftk
Jan 9, 2023, 3:20 AM
137
points
13
comments
1
min read
LW
link
(www.jefftk.com)
You’re Not One “You”—How Decision Theories Are Talking Past Each Other
keith_wynroe
Jan 9, 2023, 1:21 AM
28
points
11
comments
8
min read
LW
link
On Blogging and Podcasting
DanielFilan
Jan 9, 2023, 12:40 AM
18
points
6
comments
11
min read
LW
link
(danielfilan.com)
ChatGPT tells stories about XP-708-DQ, Eliezer, dragons, dark sorceresses, and unaligned robots becoming aligned
Bill Benzon
Jan 8, 2023, 11:21 PM
6
points
2
comments
18
min read
LW
link
Simulacra are Things
janus
Jan 8, 2023, 11:03 PM
63
points
7
comments
2
min read
LW
link
[Question]
GPT learning from smarter texts?
Viliam
Jan 8, 2023, 10:23 PM
26
points
7
comments
1
min read
LW
link
Latent variable prediction markets mockup + designer request
tailcalled
Jan 8, 2023, 10:18 PM
25
points
4
comments
1
min read
LW
link
Citability of Lesswrong and the Alignment Forum
Leon Lang
Jan 8, 2023, 10:12 PM
48
points
2
comments
1
min read
LW
link
I tried to learn as much Deep Learning math as I could in 24 hours
Phosphorous
Jan 8, 2023, 9:07 PM
31
points
2
comments
7
min read
LW
link
[Question]
What specific thing would you do with AI Alignment Research Assistant GPT?
quetzal_rainbow
Jan 8, 2023, 7:24 PM
47
points
9
comments
1
min read
LW
link
[Question]
Research ideas (AI Interpretability & Neurosciences) for a 2-months project
flux
8 Jan 2023 15:36 UTC
3
points
1
comment
1
min read
LW
link
200 COP in MI: Image Model Interpretability
Neel Nanda
8 Jan 2023 14:53 UTC
18
points
3
comments
6
min read
LW
link
Halifax Monthly Meetup: Moloch in the HRM
Ideopunk
8 Jan 2023 14:49 UTC
10
points
0
comments
1
min read
LW
link
Dangers of deference
TsviBT
8 Jan 2023 14:36 UTC
62
points
5
comments
2
min read
LW
link
Could evolution produce something truly aligned with its own optimization standards? What would an answer to this mean for AI alignment?
No77e
8 Jan 2023 11:04 UTC
3
points
4
comments
1
min read
LW
link
AI psychology should ground the theories of AI consciousness and inform human-AI ethical interaction design
Roman Leventov
8 Jan 2023 6:37 UTC
20
points
8
comments
2
min read
LW
link
Stop Talking to Each Other and Start Buying Things: Three Decades of Survival in the Desert of Social Media
the gears to ascension
8 Jan 2023 4:45 UTC
1
point
14
comments
1
min read
LW
link
(catvalente.substack.com)
Can Ads be GDPR Compliant?
jefftk
8 Jan 2023 2:50 UTC
39
points
10
comments
7
min read
LW
link
(www.jefftk.com)
Previous
Back to top
Next
N
W
F
A
C
D
E
F
G
H
I
Customize appearance
Current theme:
default
A
C
D
E
F
G
H
I
Less Wrong (text)
Less Wrong (link)
Invert colors
Reset to defaults
OK
Cancel
Hi, I’m Bobby the Basilisk! Click on the minimize button (
) to minimize the theme tweaker window, so that you can see what the page looks like with the current tweaked values. (But remember,
the changes won’t be saved until you click “OK”!
)
Theme tweaker help
Show Bobby the Basilisk
OK
Cancel