Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
Archive
Sequences
About
Search
Log In
All
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
2025
All
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
All
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
Page
1
Air Conditioner Test Results & Discussion
johnswentworth
Jun 22, 2022, 10:26 PM
82
points
42
comments
6
min read
LW
link
Announcing the LessWrong Curated Podcast
Ben Pace
and
Solenoid_Entity
Jun 22, 2022, 10:16 PM
137
points
27
comments
1
min read
LW
link
Google’s new text-to-image model—Parti, a demonstration of scaling benefits
Kayden
Jun 22, 2022, 8:00 PM
32
points
4
comments
1
min read
LW
link
Building an Epistemic Status Tracker
rcu
Jun 22, 2022, 6:57 PM
7
points
8
comments
1
min read
LW
link
Confusion about neuroscience/cognitive science as a danger for AI Alignment
Samuel Nellessen
Jun 22, 2022, 5:59 PM
3
points
1
comment
3
min read
LW
link
(snellessen.com)
[Question]
How do I use caffeine optimally?
randomstring
Jun 22, 2022, 5:59 PM
18
points
31
comments
1
min read
LW
link
Make learning a reality
Dalton Mabery
Jun 22, 2022, 3:58 PM
13
points
2
comments
1
min read
LW
link
Reflection Mechanisms as an Alignment target: A survey
Marius Hobbhahn
,
elandgre
and
Beth Barnes
Jun 22, 2022, 3:05 PM
32
points
1
comment
14
min read
LW
link
House Phone
jefftk
Jun 22, 2022, 2:20 PM
15
points
2
comments
1
min read
LW
link
(www.jefftk.com)
How to Visualize Bayesianism
David Udell
Jun 22, 2022, 1:57 PM
9
points
2
comments
3
min read
LW
link
[Question]
Are there spaces for extremely short-form rationality content?
Aleksi Liimatainen
Jun 22, 2022, 10:39 AM
5
points
1
comment
1
min read
LW
link
Solstice Movie Review: Summer Wars
SebastianG
Jun 22, 2022, 1:09 AM
22
points
6
comments
1
min read
LW
link
Security Mindset: Lessons from 20+ years of Software Security Failures Relevant to AGI Alignment
elspood
Jun 21, 2022, 11:55 PM
362
points
42
comments
7
min read
LW
link
1
review
A Quick List of Some Problems in AI Alignment As A Field
Nicholas / Heather Kross
Jun 21, 2022, 11:23 PM
75
points
12
comments
6
min read
LW
link
(www.thinkingmuchbetter.com)
[Question]
What is the difference between AI misalignment and bad programming?
puzzleGuzzle
Jun 21, 2022, 9:52 PM
6
points
2
comments
1
min read
LW
link
What I mean by the phrase “getting intimate with reality”
Luise
Jun 21, 2022, 7:42 PM
6
points
0
comments
2
min read
LW
link
(forum.effectivealtruism.org)
What I mean by the phrase “taking ideas seriously”
Luise
Jun 21, 2022, 7:42 PM
5
points
2
comments
1
min read
LW
link
(forum.effectivealtruism.org)
Hydrophobic Glasses Coating Review
jefftk
Jun 21, 2022, 6:00 PM
16
points
6
comments
1
min read
LW
link
(www.jefftk.com)
Progress links and tweets, 2022-06-20
jasoncrawford
Jun 21, 2022, 5:12 PM
12
points
2
comments
1
min read
LW
link
(rootsofprogress.org)
Debating Whether AI is Conscious Is A Distraction from Real Problems
sidhe_they
Jun 21, 2022, 4:56 PM
2
points
10
comments
1
min read
LW
link
(techpolicy.press)
Mitigating the damage from unaligned ASI by cooperating with aliens that don’t exist yet
MSRayne
Jun 21, 2022, 4:12 PM
−8
points
7
comments
6
min read
LW
link
The inordinately slow spread of good AGI conversations in ML
Rob Bensinger
Jun 21, 2022, 4:09 PM
173
points
62
comments
8
min read
LW
link
Getting from an unaligned AGI to an aligned AGI?
Tor Økland Barstad
Jun 21, 2022, 12:36 PM
13
points
7
comments
9
min read
LW
link
Common but neglected risk factors that may let you get Paxlovid
DirectedEvolution
Jun 21, 2022, 7:34 AM
29
points
8
comments
4
min read
LW
link
Dagger of Detect Evil
lsusr
Jun 21, 2022, 6:23 AM
45
points
22
comments
3
min read
LW
link
[Question]
How easy/fast is it for a AGI to hack computers/a human brain?
Noosphere89
Jun 21, 2022, 12:34 AM
0
points
1
comment
1
min read
LW
link
[Question]
What is the most probable AI?
Zeruel017
Jun 20, 2022, 11:26 PM
−2
points
0
comments
3
min read
LW
link
Evaluating a Corsi-Rosenthal Filter Cube
jefftk
Jun 20, 2022, 7:40 PM
13
points
4
comments
1
min read
LW
link
(www.jefftk.com)
Survey re AIS/LTism office in NYC
RyanCarey
Jun 20, 2022, 7:21 PM
7
points
0
comments
1
min read
LW
link
Is This Thing Sentient, Y/N?
Thane Ruthenis
Jun 20, 2022, 6:37 PM
4
points
10
comments
7
min read
LW
link
Steam
abramdemski
Jun 20, 2022, 5:38 PM
149
points
13
comments
5
min read
LW
link
1
review
Parable: The Bomb that doesn’t Explode
Lone Pine
Jun 20, 2022, 4:41 PM
14
points
5
comments
2
min read
LW
link
On corrigibility and its basin
Donald Hobson
Jun 20, 2022, 4:33 PM
16
points
3
comments
2
min read
LW
link
Announcing the DWATV Discord
Zvi
Jun 20, 2022, 3:50 PM
20
points
9
comments
1
min read
LW
link
(thezvi.wordpress.com)
Key Papers in Language Model Safety
aog
Jun 20, 2022, 3:00 PM
40
points
1
comment
22
min read
LW
link
Relationship Advice Repository
Ruby
Jun 20, 2022, 2:39 PM
109
points
36
comments
38
min read
LW
link
Adaptation Executors and the Telos Margin
Plinthist
Jun 20, 2022, 1:06 PM
2
points
8
comments
5
min read
LW
link
Are we there yet?
theflowerpot
Jun 20, 2022, 11:19 AM
2
points
2
comments
1
min read
LW
link
Causal confusion as an argument against the scaling hypothesis
RobertKirk
and
David Scott Krueger (formerly: capybaralet)
Jun 20, 2022, 10:54 AM
86
points
30
comments
15
min read
LW
link
An AI defense-offense symmetry thesis
Chris van Merwijk
Jun 20, 2022, 10:01 AM
10
points
9
comments
3
min read
LW
link
Let’s See You Write That Corrigibility Tag
Eliezer Yudkowsky
Jun 19, 2022, 9:11 PM
125
points
70
comments
1
min read
LW
link
Half-baked alignment idea: training to generalize
Aaron Bergman
Jun 19, 2022, 8:16 PM
10
points
2
comments
4
min read
LW
link
Where I agree and disagree with Eliezer
paulfchristiano
Jun 19, 2022, 7:15 PM
901
points
224
comments
18
min read
LW
link
2
reviews
[Question]
AI misalignment risk from GPT-like systems?
fiso64
Jun 19, 2022, 5:35 PM
10
points
8
comments
1
min read
LW
link
[Link-post] On Deference and Yudkowsky’s AI Risk Estimates
bmg
Jun 19, 2022, 5:25 PM
29
points
8
comments
1
min read
LW
link
Hebbian Learning Is More Common Than You Think
Aleksi Liimatainen
Jun 19, 2022, 3:57 PM
8
points
2
comments
1
min read
LW
link
The Malthusian Trap: An Extremely Short Introduction
Davis Kedrosky
Jun 19, 2022, 3:25 PM
5
points
0
comments
6
min read
LW
link
(daviskedrosky.substack.com)
Parliaments without the Parties
Yair Halberstadt
Jun 19, 2022, 2:06 PM
18
points
18
comments
2
min read
LW
link
Lamda is not an LLM
Kevin
Jun 19, 2022, 11:13 AM
7
points
10
comments
1
min read
LW
link
(www.wired.com)
Getting stuck in local minima
louis030195
Jun 19, 2022, 8:50 AM
3
points
1
comment
1
min read
LW
link
(brain.louis030195.com)
Back to top
Next
N
W
F
A
C
D
E
F
G
H
I
Customize appearance
Current theme:
default
A
C
D
E
F
G
H
I
Less Wrong (text)
Less Wrong (link)
Invert colors
Reset to defaults
OK
Cancel
Hi, I’m Bobby the Basilisk! Click on the minimize button (
) to minimize the theme tweaker window, so that you can see what the page looks like with the current tweaked values. (But remember,
the changes won’t be saved until you click “OK”!
)
Theme tweaker help
Show Bobby the Basilisk
OK
Cancel