Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
Archive
Sequences
About
Search
Log In
All
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
2025
All
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
All
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
Page
2
Halifax Monthly Meetup: Moloch in the HRM
Ideopunk
Jan 8, 2023, 2:49 PM
10
points
0
comments
1
min read
LW
link
Dangers of deference
TsviBT
Jan 8, 2023, 2:36 PM
62
points
5
comments
2
min read
LW
link
Could evolution produce something truly aligned with its own optimization standards? What would an answer to this mean for AI alignment?
No77e
Jan 8, 2023, 11:04 AM
3
points
4
comments
1
min read
LW
link
AI psychology should ground the theories of AI consciousness and inform human-AI ethical interaction design
Roman Leventov
Jan 8, 2023, 6:37 AM
20
points
8
comments
2
min read
LW
link
Stop Talking to Each Other and Start Buying Things: Three Decades of Survival in the Desert of Social Media
the gears to ascension
Jan 8, 2023, 4:45 AM
1
point
14
comments
1
min read
LW
link
(catvalente.substack.com)
Can Ads be GDPR Compliant?
jefftk
Jan 8, 2023, 2:50 AM
39
points
10
comments
7
min read
LW
link
(www.jefftk.com)
Feature suggestion: add a ‘clarity score’ to posts
LVSN
Jan 8, 2023, 1:00 AM
17
points
5
comments
1
min read
LW
link
[Question]
How do I better stick to a morning schedule?
Randomized, Controlled
Jan 8, 2023, 12:52 AM
8
points
8
comments
1
min read
LW
link
Protectionism will Slow the Deployment of AI
Ben Goldhaber
Jan 7, 2023, 8:57 PM
30
points
6
comments
2
min read
LW
link
David Krueger on AI Alignment in Academia, Coordination and Testing Intuitions
Michaël Trazzi
Jan 7, 2023, 7:59 PM
13
points
0
comments
4
min read
LW
link
(theinsideview.ai)
Looking for Spanish AI Alignment Researchers
Antb
Jan 7, 2023, 6:52 PM
7
points
3
comments
1
min read
LW
link
Nothing New: Productive Reframing
adamShimi
Jan 7, 2023, 6:43 PM
44
points
7
comments
3
min read
LW
link
(epistemologicalvigilance.substack.com)
[Question]
Asking for a name for a symptom of rationalization
metachirality
Jan 7, 2023, 6:34 PM
6
points
5
comments
1
min read
LW
link
The Fountain of Health: a First Principles Guide to Rejuvenation
PhilJackson
Jan 7, 2023, 6:34 PM
115
points
39
comments
41
min read
LW
link
What’s wrong with the paperclips scenario?
No77e
Jan 7, 2023, 5:58 PM
31
points
11
comments
1
min read
LW
link
Building a Rosetta stone for reductionism and telism (WIP)
mrcbarbier
Jan 7, 2023, 4:22 PM
5
points
0
comments
8
min read
LW
link
What should a telic science look like?
mrcbarbier
Jan 7, 2023, 4:13 PM
10
points
0
comments
11
min read
LW
link
Open & Welcome Thread—January 2023
DragonGod
Jan 7, 2023, 11:16 AM
15
points
37
comments
1
min read
LW
link
Anchoring focalism and the Identifiable victim effect: Bias in Evaluating AGI X-Risks
Remmelt
Jan 7, 2023, 9:59 AM
1
point
2
comments
LW
link
Can ChatGPT count?
p.b.
Jan 7, 2023, 7:57 AM
13
points
11
comments
2
min read
LW
link
Benevolent AI and mental health
peter schwarz
Jan 7, 2023, 1:30 AM
−31
points
2
comments
1
min read
LW
link
An Ignorant View on Ineffectiveness of AI Safety
Iknownothing
Jan 7, 2023, 1:29 AM
14
points
7
comments
3
min read
LW
link
Optimizing Human Collective Intelligence to Align AI
Shoshannah Tekofsky
Jan 7, 2023, 1:21 AM
12
points
5
comments
6
min read
LW
link
[Question]
[Discussion] How Broad is the Human Cognitive Spectrum?
DragonGod
Jan 7, 2023, 12:56 AM
29
points
51
comments
2
min read
LW
link
Implications of simulators
TW123
Jan 7, 2023, 12:37 AM
17
points
0
comments
12
min read
LW
link
[Linkpost] Jan Leike on three kinds of alignment taxes
Orpheus16
Jan 6, 2023, 11:57 PM
27
points
2
comments
3
min read
LW
link
(aligned.substack.com)
The Limit of Language Models
DragonGod
Jan 6, 2023, 11:53 PM
44
points
26
comments
4
min read
LW
link
Why didn’t we get the four-hour workday?
jasoncrawford
Jan 6, 2023, 9:29 PM
141
points
34
comments
6
min read
LW
link
(rootsofprogress.org)
AI security might be helpful for AI alignment
Igor Ivanov
Jan 6, 2023, 8:16 PM
36
points
1
comment
2
min read
LW
link
Categorizing failures as “outer” or “inner” misalignment is often confused
Rohin Shah
Jan 6, 2023, 3:48 PM
93
points
21
comments
8
min read
LW
link
Definitions of “objective” should be Probable and Predictive
Rohin Shah
Jan 6, 2023, 3:40 PM
43
points
27
comments
12
min read
LW
link
200 COP in MI: Techniques, Tooling and Automation
Neel Nanda
Jan 6, 2023, 3:08 PM
13
points
0
comments
15
min read
LW
link
Ball Square Station and Ridership Maximization
jefftk
Jan 6, 2023, 1:20 PM
13
points
0
comments
1
min read
LW
link
(www.jefftk.com)
Childhood Roundup #1
Zvi
Jan 6, 2023, 1:00 PM
84
points
27
comments
8
min read
LW
link
(thezvi.wordpress.com)
AI improving AI [MLAISU W01!]
Esben Kran
Jan 6, 2023, 11:13 AM
5
points
0
comments
4
min read
LW
link
(newsletter.apartresearch.com)
AI Safety Camp, Virtual Edition 2023
Linda Linsefors
Jan 6, 2023, 11:09 AM
40
points
10
comments
3
min read
LW
link
(aisafety.camp)
Kakistocuriosity
LVSN
Jan 6, 2023, 7:38 AM
7
points
3
comments
1
min read
LW
link
AI Safety Camp: Machine Learning for Scientific Discovery
Eleni Angelou
Jan 6, 2023, 3:21 AM
3
points
0
comments
1
min read
LW
link
Metaculus Year in Review: 2022
ChristianWilliams
Jan 6, 2023, 1:23 AM
6
points
0
comments
LW
link
UDASSA
Jacob Falkovich
Jan 6, 2023, 1:07 AM
27
points
8
comments
10
min read
LW
link
The Involuntary Pacifists
Capybasilisk
Jan 6, 2023, 12:28 AM
11
points
3
comments
2
min read
LW
link
Get an Electric Toothbrush.
Cervera
Jan 5, 2023, 9:08 PM
21
points
4
comments
1
min read
LW
link
Discursive Competence in ChatGPT, Part 1: Talking with Dragons
Bill Benzon
Jan 5, 2023, 9:01 PM
2
points
0
comments
6
min read
LW
link
Transformative AI issues (not just misalignment): an overview
HoldenKarnofsky
Jan 5, 2023, 8:20 PM
34
points
6
comments
18
min read
LW
link
(www.cold-takes.com)
How to slow down scientific progress, according to Leo Szilard
jasoncrawford
Jan 5, 2023, 6:26 PM
134
points
18
comments
2
min read
LW
link
(rootsofprogress.org)
Paper: Superposition, Memorization, and Double Descent (Anthropic)
LawrenceC
Jan 5, 2023, 5:54 PM
53
points
11
comments
1
min read
LW
link
(transformer-circuits.pub)
Collapse Might Not Be Desirable
Dzoldzaya
Jan 5, 2023, 5:29 PM
−2
points
9
comments
2
min read
LW
link
Singapore—Small casual dinner in Chinatown #6
Joe Rocca
Jan 5, 2023, 5:00 PM
2
points
1
comment
1
min read
LW
link
[Question]
Image generation and alignment
rpglover64
Jan 5, 2023, 4:05 PM
3
points
3
comments
1
min read
LW
link
[Question]
Machine Learning vs Differential Privacy
Ilio
Jan 5, 2023, 3:14 PM
10
points
10
comments
1
min read
LW
link
Previous
Back to top
Next
N
W
F
A
C
D
E
F
G
H
I
Customize appearance
Current theme:
default
A
C
D
E
F
G
H
I
Less Wrong (text)
Less Wrong (link)
Invert colors
Reset to defaults
OK
Cancel
Hi, I’m Bobby the Basilisk! Click on the minimize button (
) to minimize the theme tweaker window, so that you can see what the page looks like with the current tweaked values. (But remember,
the changes won’t be saved until you click “OK”!
)
Theme tweaker help
Show Bobby the Basilisk
OK
Cancel