Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
Archive
Sequences
About
Search
Log In
All
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
2025
All
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
All
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
Page
2
OpenAI Launches Superalignment Taskforce
Zvi
Jul 11, 2023, 1:00 PM
150
points
40
comments
49
min read
LW
link
(thezvi.wordpress.com)
Critiquing Risks From Learned Optimization, and Avoiding Cached Theories
ProofBySonnet
Jul 11, 2023, 11:38 AM
1
point
0
comments
6
min read
LW
link
[UPDATE: deadline extended to July 24!] New wind in rationality’s sails: Applications for Epistea Residency 2023 are now open
Jana Meixnerová
and
Irena Kotíková
Jul 11, 2023, 11:02 AM
80
points
7
comments
3
min read
LW
link
Two Hot Takes about Quine
Charlie Steiner
Jul 11, 2023, 6:42 AM
17
points
0
comments
2
min read
LW
link
Disincentivizing deception in mesa optimizers with Model Tampering
martinkunev
Jul 11, 2023, 12:44 AM
3
points
0
comments
2
min read
LW
link
Drawn Out: a story
Richard_Ngo
Jul 11, 2023, 12:08 AM
82
points
2
comments
8
min read
LW
link
Definitions are about efficiency and consistency with common language.
Nacruno96
Jul 10, 2023, 11:46 PM
1
point
0
comments
4
min read
LW
link
Reframing Evolution—An information wavefront traveling through time
Joshua Clancy
Jul 10, 2023, 10:36 PM
1
point
0
comments
5
min read
LW
link
(midflip.org)
GPT-7: The Tale of the Big Computer (An Experimental Story)
Justin Bullock
Jul 10, 2023, 8:22 PM
4
points
4
comments
5
min read
LW
link
Cost-effectiveness of professional field-building programs for AI safety research
Dan H
Jul 10, 2023, 6:28 PM
8
points
5
comments
18
min read
LW
link
Cost-effectiveness of student programs for AI safety research
Dan H
Jul 10, 2023, 6:28 PM
15
points
2
comments
15
min read
LW
link
Modeling the impact of AI safety field-building programs
Dan H
Jul 10, 2023, 6:27 PM
21
points
0
comments
7
min read
LW
link
I think Michael Bailey’s dismissal of my autogynephilia questions for Scott Alexander and Aella makes very little sense
tailcalled
Jul 10, 2023, 5:39 PM
46
points
45
comments
2
min read
LW
link
Incentives from a causal perspective
tom4everitt
,
James Fox
,
RyanCarey
,
mattmacdermott
,
sbenthall
and
Jonathan Richens
Jul 10, 2023, 5:16 PM
27
points
0
comments
6
min read
LW
link
Is the Endowment Effect Due to Incomparability?
Kevin Dorst
Jul 10, 2023, 4:26 PM
21
points
10
comments
7
min read
LW
link
(kevindorst.substack.com)
Frontier AI Regulation
Zach Stein-Perlman
Jul 10, 2023, 2:30 PM
21
points
4
comments
8
min read
LW
link
(arxiv.org)
Why is it so hard to change people’s minds? Well, imagine if it wasn’t...
Celarix
Jul 10, 2023, 1:55 PM
6
points
9
comments
6
min read
LW
link
Consider Joining the UK Foundation Model Taskforce
Zvi
Jul 10, 2023, 1:50 PM
105
points
12
comments
1
min read
LW
link
(thezvi.wordpress.com)
“Reframing Superintelligence” + LLMs + 4 years
Eric Drexler
Jul 10, 2023, 1:42 PM
118
points
9
comments
12
min read
LW
link
Open-minded updatelessness
Nicolas Macé
,
JesseClifton
and
SMK
Jul 10, 2023, 11:08 AM
66
points
21
comments
12
min read
LW
link
Consciousness as a conflationary alliance term for intrinsically valued internal experiences
Andrew_Critch
Jul 10, 2023, 8:09 AM
212
points
54
comments
11
min read
LW
link
2
reviews
The world where LLMs are possible
Ape in the coat
Jul 10, 2023, 8:00 AM
20
points
10
comments
3
min read
LW
link
The virtue of determination
Richard_Ngo
Jul 10, 2023, 5:11 AM
73
points
6
comments
4
min read
LW
link
Some reasons to not say “Doomer”
Ruby
Jul 9, 2023, 9:05 PM
46
points
18
comments
4
min read
LW
link
The Seeker’s Game – Vignettes from the Bay
Yulia
Jul 9, 2023, 7:32 PM
141
points
19
comments
16
min read
LW
link
[Question]
Why have exposure notification apps been (mostly) discontinued?
VipulNaik
Jul 9, 2023, 7:07 PM
10
points
5
comments
2
min read
LW
link
[Question]
The Necessity of Privacy: A Condition for Social Change and Experimentation?
Blake
Jul 9, 2023, 6:42 PM
−8
points
1
comment
1
min read
LW
link
Attempting to Deconstruct “Real”
herschel
Jul 9, 2023, 4:40 PM
21
points
23
comments
2
min read
LW
link
Quick proposal: Decision market regrantor using manifund (please improve)
Nathan Young
Jul 9, 2023, 12:49 PM
10
points
5
comments
5
min read
LW
link
[Question]
Where are the people building AGI in the non-dumb way?
Johannes C. Mayer
Jul 9, 2023, 11:39 AM
10
points
19
comments
2
min read
LW
link
[Question]
What to read on the “informal multi-world model”?
mishka
Jul 9, 2023, 4:48 AM
13
points
23
comments
1
min read
LW
link
Whether LLMs “understand” anything is mostly a terminological dispute
RobertM
Jul 9, 2023, 3:31 AM
10
points
1
comment
1
min read
LW
link
“View”
herschel
Jul 8, 2023, 11:19 PM
6
points
0
comments
2
min read
LW
link
[Question]
H5N1. Just how bad is the situation?
Q Home
Jul 8, 2023, 10:09 PM
16
points
8
comments
1
min read
LW
link
A Two-Part System for Practical Self-Care
Jonathan Moregård
Jul 8, 2023, 9:23 PM
11
points
0
comments
3
min read
LW
link
(honestliving.substack.com)
Really Strong Features Found in Residual Stream
Logan Riggs
Jul 8, 2023, 7:40 PM
69
points
6
comments
2
min read
LW
link
Eight Strategies for Tackling the Hard Part of the Alignment Problem
scasper
Jul 8, 2023, 6:55 PM
42
points
11
comments
7
min read
LW
link
“Concepts of Agency in Biology” (Okasha, 2023) - Brief Paper Summary
Nora_Ammann
Jul 8, 2023, 6:22 PM
40
points
3
comments
7
min read
LW
link
Blanchard’s Dangerous Idea and the Plight of the Lucid Crossdreamer
Zack_M_Davis
Jul 8, 2023, 6:03 PM
38
points
135
comments
72
min read
LW
link
(unremediatedgender.space)
Continuous Adversarial Quality Assurance: Extending RLHF and Constitutional AI
Benaya Koren
Jul 8, 2023, 5:32 PM
6
points
0
comments
9
min read
LW
link
Commentless downvoting is not a good way to fight infohazards
DirectedEvolution
Jul 8, 2023, 5:29 PM
6
points
9
comments
3
min read
LW
link
[Question]
Why does anxiety (?) make me dumb?
TeaTieAndHat
Jul 8, 2023, 4:13 PM
18
points
14
comments
3
min read
LW
link
Economic Time Bomb: An Overlooked Employment Bubble Threatening the US Economy
Glenn Clayton
Jul 8, 2023, 3:19 PM
4
points
10
comments
6
min read
LW
link
What is everyone doing in AI governance
Igor Ivanov
Jul 8, 2023, 3:16 PM
12
points
0
comments
5
min read
LW
link
LLM misalignment can probably be found without manual prompt engineering
ProgramCrafter
Jul 8, 2023, 2:35 PM
1
point
0
comments
1
min read
LW
link
You must not fool yourself, and you are the easiest person to fool
Richard_Ngo
Jul 8, 2023, 2:05 PM
35
points
5
comments
4
min read
LW
link
Fixed Point: a love story
Richard_Ngo
Jul 8, 2023, 1:56 PM
100
points
2
comments
7
min read
LW
link
Announcing AI Alignment workshop at the ALIFE 2023 conference
rorygreig
Jul 8, 2023, 1:52 PM
16
points
0
comments
1
min read
LW
link
(humanvaluesandartificialagency.com)
3D Printed Talkbox Cap
jefftk
Jul 8, 2023, 1:00 PM
9
points
0
comments
1
min read
LW
link
(www.jefftk.com)
Writing this post as rationality case study
Ben Amitay
Jul 8, 2023, 12:24 PM
10
points
8
comments
2
min read
LW
link
Previous
Back to top
Next
N
W
F
A
C
D
E
F
G
H
I
Customize appearance
Current theme:
default
A
C
D
E
F
G
H
I
Less Wrong (text)
Less Wrong (link)
Invert colors
Reset to defaults
OK
Cancel
Hi, I’m Bobby the Basilisk! Click on the minimize button (
) to minimize the theme tweaker window, so that you can see what the page looks like with the current tweaked values. (But remember,
the changes won’t be saved until you click “OK”!
)
Theme tweaker help
Show Bobby the Basilisk
OK
Cancel