Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
Archive
Sequences
About
Search
Log In
All
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
2025
All
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
All
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
Page
2
$20K In Bounties for AI Safety Public Materials
Dan H
,
TW123
and
ozhang
Aug 5, 2022, 2:52 AM
71
points
9
comments
6
min read
LW
link
Paper reading as a Cargo Cult
jem-mosig
Aug 7, 2022, 7:50 AM
70
points
10
comments
5
min read
LW
link
Building a Bugs List prompts
CFAR!Duncan
Aug 13, 2022, 8:00 AM
69
points
9
comments
2
min read
LW
link
Jack Clark on the realities of AI policy
Kaj_Sotala
Aug 7, 2022, 8:44 AM
68
points
3
comments
3
min read
LW
link
(threadreaderapp.com)
The Expanding Moral Cinematic Universe
Raemon
Aug 28, 2022, 6:42 PM
67
points
9
comments
14
min read
LW
link
In Defense Of Making Money
George3d6
Aug 18, 2022, 2:10 PM
65
points
13
comments
7
min read
LW
link
(www.epistem.ink)
AI art isn’t “about to shake things up”. It’s already here.
Davis_Kingsley
Aug 22, 2022, 11:17 AM
65
points
19
comments
3
min read
LW
link
Vingean Agency
abramdemski
Aug 24, 2022, 8:08 PM
63
points
14
comments
3
min read
LW
link
ACX Meetups Everywhere List
Scott Alexander
Aug 26, 2022, 6:12 PM
63
points
1
comment
41
min read
LW
link
Encultured AI Pre-planning, Part 1: Enabling New Benchmarks
Andrew_Critch
and
Nick Hay
Aug 8, 2022, 10:44 PM
63
points
2
comments
6
min read
LW
link
Steganography in Chain of Thought Reasoning
A Ray
Aug 8, 2022, 3:47 AM
62
points
13
comments
6
min read
LW
link
Oops It’s Time To Overthrow the Organizer Day!
Screwtape
Aug 18, 2022, 4:40 PM
62
points
5
comments
4
min read
LW
link
Seeking PCK (Pedagogical Content Knowledge)
CFAR!Duncan
Aug 12, 2022, 4:15 AM
62
points
11
comments
5
min read
LW
link
Seeking Interns/RAs for Mechanistic Interpretability Projects
Neel Nanda
Aug 15, 2022, 7:11 AM
61
points
0
comments
2
min read
LW
link
Autonomy as taking responsibility for reference maintenance
Ramana Kumar
Aug 17, 2022, 12:50 PM
61
points
3
comments
5
min read
LW
link
An Introduction to Current Theories of Consciousness
hohenheim
Aug 28, 2022, 5:55 PM
60
points
43
comments
49
min read
LW
link
OpenAI’s Alignment Plans
dkirmani
Aug 24, 2022, 7:39 PM
60
points
17
comments
5
min read
LW
link
(openai.com)
Anti-squatted AI x-risk domains index
plex
Aug 12, 2022, 12:01 PM
59
points
6
comments
1
min read
LW
link
Finding Goals in the World Model
Jeremy Gillen
,
JamesH
and
Thomas Larsen
Aug 22, 2022, 6:06 PM
59
points
8
comments
13
min read
LW
link
The Pragmascope Idea
johnswentworth
Aug 4, 2022, 9:52 PM
59
points
20
comments
3
min read
LW
link
My thoughts on direct work (and joining LessWrong)
RobertM
Aug 16, 2022, 6:53 PM
58
points
4
comments
6
min read
LW
link
How to plan for a radically uncertain future?
Kerry
Aug 30, 2022, 2:14 AM
57
points
35
comments
1
min read
LW
link
EA & LW Forums Weekly Summary (21 Aug − 27 Aug 22′)
Zoe Williams
Aug 30, 2022, 1:42 AM
57
points
4
comments
12
min read
LW
link
How and why to turn everything into audio
KatWoods
and
AmberDawn
Aug 11, 2022, 8:55 AM
57
points
20
comments
5
min read
LW
link
Refine’s First Blog Post Day
adamShimi
Aug 13, 2022, 10:23 AM
55
points
3
comments
1
min read
LW
link
[Question]
How to bet against civilizational adequacy?
Wei Dai
12 Aug 2022 23:33 UTC
54
points
20
comments
1
min read
LW
link
All the posts I will never write
Alexander Gietelink Oldenziel
14 Aug 2022 18:29 UTC
54
points
8
comments
8
min read
LW
link
Brain-like AGI project “aintelope”
Gunnar_Zarncke
14 Aug 2022 16:33 UTC
54
points
2
comments
1
min read
LW
link
Transformer language models are doing something more general
Numendil
3 Aug 2022 21:13 UTC
53
points
6
comments
2
min read
LW
link
I missed the crux of the alignment problem the whole time
zeshen
13 Aug 2022 10:11 UTC
53
points
7
comments
3
min read
LW
link
Using GPT-3 to augment human intelligence
Henrik Karlsson
10 Aug 2022 15:54 UTC
52
points
8
comments
18
min read
LW
link
(escapingflatland.substack.com)
Variational Bayesian methods
Ege Erdil
25 Aug 2022 20:49 UTC
52
points
2
comments
9
min read
LW
link
A Data limited future
Donald Hobson
6 Aug 2022 14:56 UTC
52
points
25
comments
2
min read
LW
link
Turbocharging
CFAR!Duncan
2 Aug 2022 0:01 UTC
52
points
5
comments
9
min read
LW
link
Announcing Squiggle: Early Access
ozziegooen
3 Aug 2022 19:48 UTC
51
points
7
comments
7
min read
LW
link
(forum.effectivealtruism.org)
General alignment properties
TurnTrout
8 Aug 2022 23:40 UTC
51
points
2
comments
1
min read
LW
link
Againstness
CFAR!Duncan
2 Aug 2022 19:29 UTC
50
points
8
comments
9
min read
LW
link
Polaris, Five-Second Versions, and Thought Lengths
CFAR!Duncan
1 Aug 2022 7:14 UTC
50
points
12
comments
8
min read
LW
link
On Car Seats as Contraception
Zvi
22 Aug 2022 14:10 UTC
49
points
15
comments
35
min read
LW
link
(thezvi.wordpress.com)
Six weeks doesn’t make a habit
lynettebye
6 Aug 2022 8:54 UTC
48
points
1
comment
3
min read
LW
link
AGI Timelines Are Mostly Not Strategically Relevant To Alignment
johnswentworth
23 Aug 2022 20:15 UTC
48
points
34
comments
1
min read
LW
link
The Shard Theory Alignment Scheme
David Udell
25 Aug 2022 4:52 UTC
47
points
32
comments
2
min read
LW
link
Gradient descent doesn’t select for inner search
Ivan Vendrov
13 Aug 2022 4:15 UTC
47
points
23
comments
4
min read
LW
link
Covid 8/18/22: CDC Admits Mistakes
Zvi
18 Aug 2022 14:30 UTC
46
points
9
comments
17
min read
LW
link
(thezvi.wordpress.com)
Proposal: Consider not using distance-direction-dimension words in abstract discussions
moridinamael
9 Aug 2022 20:44 UTC
46
points
18
comments
5
min read
LW
link
The Falling Drill
Screwtape
5 Aug 2022 0:08 UTC
46
points
3
comments
2
min read
LW
link
Review: Amusing Ourselves to Death
L Rudolf L
20 Aug 2022 21:13 UTC
44
points
7
comments
16
min read
LW
link
1
review
(www.strataoftheworld.com)
Volunteer to host a meetup!
mingyuan
21 Aug 2022 22:43 UTC
44
points
1
comment
1
min read
LW
link
The Dumbest Possible Gets There First
Artaxerxes
13 Aug 2022 10:20 UTC
44
points
7
comments
2
min read
LW
link
The Solomonoff prior is malign. It’s not a big deal.
Charlie Steiner
25 Aug 2022 8:25 UTC
43
points
9
comments
7
min read
LW
link
Previous
Back to top
Next
N
W
F
A
C
D
E
F
G
H
I
Customize appearance
Current theme:
default
A
C
D
E
F
G
H
I
Less Wrong (text)
Less Wrong (link)
Invert colors
Reset to defaults
OK
Cancel
Hi, I’m Bobby the Basilisk! Click on the minimize button (
) to minimize the theme tweaker window, so that you can see what the page looks like with the current tweaked values. (But remember,
the changes won’t be saved until you click “OK”!
)
Theme tweaker help
Show Bobby the Basilisk
OK
Cancel