Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
Archive
Sequences
About
Search
Log In
All
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
2025
All
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
All
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
Page
1
ChatGPT tells stories about XP-708-DQ, Eliezer, dragons, dark sorceresses, and unaligned robots becoming aligned
Bill Benzon
Jan 8, 2023, 11:21 PM
6
points
2
comments
18
min read
LW
link
Simulacra are Things
janus
Jan 8, 2023, 11:03 PM
63
points
7
comments
2
min read
LW
link
[Question]
GPT learning from smarter texts?
Viliam
Jan 8, 2023, 10:23 PM
26
points
7
comments
1
min read
LW
link
Latent variable prediction markets mockup + designer request
tailcalled
Jan 8, 2023, 10:18 PM
25
points
4
comments
1
min read
LW
link
Citability of Lesswrong and the Alignment Forum
Leon Lang
Jan 8, 2023, 10:12 PM
48
points
2
comments
1
min read
LW
link
I tried to learn as much Deep Learning math as I could in 24 hours
Phosphorous
Jan 8, 2023, 9:07 PM
31
points
2
comments
7
min read
LW
link
[Question]
What specific thing would you do with AI Alignment Research Assistant GPT?
quetzal_rainbow
Jan 8, 2023, 7:24 PM
47
points
9
comments
1
min read
LW
link
[Question]
Research ideas (AI Interpretability & Neurosciences) for a 2-months project
flux
Jan 8, 2023, 3:36 PM
3
points
1
comment
1
min read
LW
link
200 COP in MI: Image Model Interpretability
Neel Nanda
Jan 8, 2023, 2:53 PM
18
points
3
comments
6
min read
LW
link
Halifax Monthly Meetup: Moloch in the HRM
Ideopunk
Jan 8, 2023, 2:49 PM
10
points
0
comments
1
min read
LW
link
Dangers of deference
TsviBT
Jan 8, 2023, 2:36 PM
62
points
5
comments
2
min read
LW
link
Could evolution produce something truly aligned with its own optimization standards? What would an answer to this mean for AI alignment?
No77e
Jan 8, 2023, 11:04 AM
3
points
4
comments
1
min read
LW
link
AI psychology should ground the theories of AI consciousness and inform human-AI ethical interaction design
Roman Leventov
Jan 8, 2023, 6:37 AM
20
points
8
comments
2
min read
LW
link
Stop Talking to Each Other and Start Buying Things: Three Decades of Survival in the Desert of Social Media
the gears to ascension
Jan 8, 2023, 4:45 AM
1
point
14
comments
1
min read
LW
link
(catvalente.substack.com)
Can Ads be GDPR Compliant?
jefftk
Jan 8, 2023, 2:50 AM
39
points
10
comments
7
min read
LW
link
(www.jefftk.com)
Feature suggestion: add a ‘clarity score’ to posts
LVSN
Jan 8, 2023, 1:00 AM
17
points
5
comments
1
min read
LW
link
[Question]
How do I better stick to a morning schedule?
Randomized, Controlled
Jan 8, 2023, 12:52 AM
8
points
8
comments
1
min read
LW
link
Protectionism will Slow the Deployment of AI
Ben Goldhaber
Jan 7, 2023, 8:57 PM
30
points
6
comments
2
min read
LW
link
David Krueger on AI Alignment in Academia, Coordination and Testing Intuitions
Michaël Trazzi
Jan 7, 2023, 7:59 PM
13
points
0
comments
4
min read
LW
link
(theinsideview.ai)
Looking for Spanish AI Alignment Researchers
Antb
Jan 7, 2023, 6:52 PM
7
points
3
comments
1
min read
LW
link
Nothing New: Productive Reframing
adamShimi
Jan 7, 2023, 6:43 PM
44
points
7
comments
3
min read
LW
link
(epistemologicalvigilance.substack.com)
[Question]
Asking for a name for a symptom of rationalization
metachirality
Jan 7, 2023, 6:34 PM
6
points
5
comments
1
min read
LW
link
The Fountain of Health: a First Principles Guide to Rejuvenation
PhilJackson
Jan 7, 2023, 6:34 PM
115
points
39
comments
41
min read
LW
link
What’s wrong with the paperclips scenario?
No77e
Jan 7, 2023, 5:58 PM
31
points
11
comments
1
min read
LW
link
Building a Rosetta stone for reductionism and telism (WIP)
mrcbarbier
Jan 7, 2023, 4:22 PM
5
points
0
comments
8
min read
LW
link
What should a telic science look like?
mrcbarbier
Jan 7, 2023, 4:13 PM
10
points
0
comments
11
min read
LW
link
Open & Welcome Thread—January 2023
DragonGod
Jan 7, 2023, 11:16 AM
15
points
37
comments
1
min read
LW
link
Anchoring focalism and the Identifiable victim effect: Bias in Evaluating AGI X-Risks
Remmelt
Jan 7, 2023, 9:59 AM
1
point
2
comments
LW
link
Can ChatGPT count?
p.b.
Jan 7, 2023, 7:57 AM
13
points
11
comments
2
min read
LW
link
Benevolent AI and mental health
peter schwarz
Jan 7, 2023, 1:30 AM
−31
points
2
comments
1
min read
LW
link
An Ignorant View on Ineffectiveness of AI Safety
Iknownothing
Jan 7, 2023, 1:29 AM
14
points
7
comments
3
min read
LW
link
Optimizing Human Collective Intelligence to Align AI
Shoshannah Tekofsky
Jan 7, 2023, 1:21 AM
12
points
5
comments
6
min read
LW
link
[Question]
[Discussion] How Broad is the Human Cognitive Spectrum?
DragonGod
Jan 7, 2023, 12:56 AM
29
points
51
comments
2
min read
LW
link
Implications of simulators
TW123
Jan 7, 2023, 12:37 AM
17
points
0
comments
12
min read
LW
link
[Linkpost] Jan Leike on three kinds of alignment taxes
Orpheus16
Jan 6, 2023, 11:57 PM
27
points
2
comments
3
min read
LW
link
(aligned.substack.com)
The Limit of Language Models
DragonGod
Jan 6, 2023, 11:53 PM
44
points
26
comments
4
min read
LW
link
Why didn’t we get the four-hour workday?
jasoncrawford
Jan 6, 2023, 9:29 PM
141
points
34
comments
6
min read
LW
link
(rootsofprogress.org)
AI security might be helpful for AI alignment
Igor Ivanov
Jan 6, 2023, 8:16 PM
36
points
1
comment
2
min read
LW
link
Categorizing failures as “outer” or “inner” misalignment is often confused
Rohin Shah
Jan 6, 2023, 3:48 PM
93
points
21
comments
8
min read
LW
link
Definitions of “objective” should be Probable and Predictive
Rohin Shah
Jan 6, 2023, 3:40 PM
43
points
27
comments
12
min read
LW
link
200 COP in MI: Techniques, Tooling and Automation
Neel Nanda
Jan 6, 2023, 3:08 PM
13
points
0
comments
15
min read
LW
link
Ball Square Station and Ridership Maximization
jefftk
Jan 6, 2023, 1:20 PM
13
points
0
comments
1
min read
LW
link
(www.jefftk.com)
Childhood Roundup #1
Zvi
Jan 6, 2023, 1:00 PM
84
points
27
comments
8
min read
LW
link
(thezvi.wordpress.com)
AI improving AI [MLAISU W01!]
Esben Kran
6 Jan 2023 11:13 UTC
5
points
0
comments
4
min read
LW
link
(newsletter.apartresearch.com)
AI Safety Camp, Virtual Edition 2023
Linda Linsefors
6 Jan 2023 11:09 UTC
40
points
10
comments
3
min read
LW
link
(aisafety.camp)
Kakistocuriosity
LVSN
6 Jan 2023 7:38 UTC
7
points
3
comments
1
min read
LW
link
AI Safety Camp: Machine Learning for Scientific Discovery
Eleni Angelou
6 Jan 2023 3:21 UTC
3
points
0
comments
1
min read
LW
link
Metaculus Year in Review: 2022
ChristianWilliams
6 Jan 2023 1:23 UTC
6
points
0
comments
LW
link
UDASSA
Jacob Falkovich
6 Jan 2023 1:07 UTC
27
points
8
comments
10
min read
LW
link
The Involuntary Pacifists
Capybasilisk
6 Jan 2023 0:28 UTC
11
points
3
comments
2
min read
LW
link
Back to top
Next
N
W
F
A
C
D
E
F
G
H
I
Customize appearance
Current theme:
default
A
C
D
E
F
G
H
I
Less Wrong (text)
Less Wrong (link)
Invert colors
Reset to defaults
OK
Cancel
Hi, I’m Bobby the Basilisk! Click on the minimize button (
) to minimize the theme tweaker window, so that you can see what the page looks like with the current tweaked values. (But remember,
the changes won’t be saved until you click “OK”!
)
Theme tweaker help
Show Bobby the Basilisk
OK
Cancel