Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
Archive
Sequences
About
Search
Log In
All
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
2025
All
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
All
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
Page
1
AI coordination needs clear wins
evhub
Sep 1, 2022, 11:41 PM
147
points
16
comments
2
min read
LW
link
1
review
Short story speculating on possible ramifications of AI on the art world
Yitz
Sep 1, 2022, 9:15 PM
30
points
8
comments
3
min read
LW
link
(archiveofourown.org)
Why was progress so slow in the past?
jasoncrawford
Sep 1, 2022, 8:26 PM
54
points
31
comments
6
min read
LW
link
(rootsofprogress.org)
AI Safety and Neighboring Communities: A Quick-Start Guide, as of Summer 2022
Sam Bowman
Sep 1, 2022, 7:15 PM
76
points
2
comments
7
min read
LW
link
Gradient Hacker Design Principles From Biology
johnswentworth
Sep 1, 2022, 7:03 PM
60
points
13
comments
3
min read
LW
link
Book review: Put Your Ass Where Your Heart Wants to Be
Ruhul
Sep 1, 2022, 6:21 PM
1
point
2
comments
10
min read
LW
link
A Survey of Foundational Methods in Inverse Reinforcement Learning
adamk
Sep 1, 2022, 6:21 PM
27
points
0
comments
12
min read
LW
link
I Tripped and Became GPT! (And How This Updated My Timelines)
Frankophone
Sep 1, 2022, 5:56 PM
31
points
0
comments
4
min read
LW
link
[Question]
Fixed point theory (locally (α,β,ψ) dominated contractive condition)
muzammil
Sep 1, 2022, 5:56 PM
0
points
3
comments
1
min read
LW
link
Alignment is hard. Communicating that, might be harder
Eleni Angelou
Sep 1, 2022, 4:57 PM
7
points
8
comments
3
min read
LW
link
Covid 9/1/22: Meet the New Booster
Zvi
Sep 1, 2022, 2:00 PM
41
points
6
comments
14
min read
LW
link
(thezvi.wordpress.com)
A Starter-kit for Rationality Space
Jesse Hoogland
Sep 1, 2022, 1:04 PM
43
points
0
comments
1
min read
LW
link
(github.com)
Pondering the paucity of volcanic profanity post Pompeii perusal
CraigMichael
Sep 1, 2022, 9:29 AM
21
points
2
comments
15
min read
LW
link
Infra-Exercises, Part 1
Diffractor
,
Jack Parker
and
Connall Garrod
Sep 1, 2022, 5:06 AM
62
points
10
comments
1
min read
LW
link
Strategy For Conditioning Generative Models
james.lucassen
and
evhub
Sep 1, 2022, 4:34 AM
31
points
4
comments
18
min read
LW
link
Safety Committee Resources
jefftk
Sep 1, 2022, 2:30 AM
22
points
2
comments
1
min read
LW
link
(www.jefftk.com)
Progress links and tweets, 2022-08-31
jasoncrawford
Aug 31, 2022, 9:54 PM
13
points
4
comments
1
min read
LW
link
(rootsofprogress.org)
Enantiodromia
ChristianKl
Aug 31, 2022, 9:13 PM
38
points
7
comments
3
min read
LW
link
[Question]
Supposing Europe is headed for a serious energy crisis this winter, what can/should one do as an individual to prepare?
Erich_Grunewald
Aug 31, 2022, 7:28 PM
18
points
13
comments
1
min read
LW
link
New 80,000 Hours problem profile on existential risks from AI
Benjamin Hilton
Aug 31, 2022, 5:36 PM
28
points
6
comments
7
min read
LW
link
(80000hours.org)
Grand Theft Education
Zvi
Aug 31, 2022, 11:50 AM
66
points
18
comments
20
min read
LW
link
(thezvi.wordpress.com)
How much impact can any one man have?
GregorDeVillain
Aug 31, 2022, 10:26 AM
9
points
3
comments
4
min read
LW
link
[Question]
How might we make better use of AI capabilities research for alignment purposes?
Jemal Young
Aug 31, 2022, 4:19 AM
11
points
4
comments
1
min read
LW
link
[Question]
AI Box Experiment: Are people still interested?
Double
Aug 31, 2022, 3:04 AM
30
points
13
comments
1
min read
LW
link
OC ACX/LW in Newport Beach
Michael Michalchik
Aug 31, 2022, 2:56 AM
1
point
1
comment
1
min read
LW
link
Survey of NLP Researchers: NLP is contributing to AGI progress; major catastrophe plausible
Sam Bowman
Aug 31, 2022, 1:39 AM
91
points
6
comments
2
min read
LW
link
And the word was “God”
pchvykov
Aug 30, 2022, 9:13 PM
−22
points
4
comments
3
min read
LW
link
Worlds Where Iterative Design Fails
johnswentworth
Aug 30, 2022, 8:48 PM
209
points
30
comments
10
min read
LW
link
1
review
Inner Alignment via Superpowers
JamesH
,
Thomas Larsen
and
Jeremy Gillen
Aug 30, 2022, 8:01 PM
37
points
13
comments
4
min read
LW
link
ML Model Attribution Challenge [Linkpost]
aog
Aug 30, 2022, 7:34 PM
11
points
0
comments
1
min read
LW
link
(mlmac.io)
How likely is deceptive alignment?
evhub
Aug 30, 2022, 7:34 PM
105
points
28
comments
60
min read
LW
link
Built-In Bundling For Faster Loading
jefftk
Aug 30, 2022, 7:20 PM
15
points
0
comments
2
min read
LW
link
(www.jefftk.com)
[Question]
A bayesian updating on expert opinions
amarai
Aug 30, 2022, 11:56 AM
1
point
1
comment
1
min read
LW
link
Any Utilitarianism Makes Sense As Policy
George3d6
Aug 30, 2022, 9:55 AM
6
points
6
comments
7
min read
LW
link
(www.epistem.ink)
A gentle primer on caring, including in strange senses, with applications
Kaarel
Aug 30, 2022, 8:05 AM
10
points
4
comments
18
min read
LW
link
Modified Guess Culture
konstell
Aug 30, 2022, 2:30 AM
5
points
5
comments
1
min read
LW
link
(konstell.com)
[Question]
What is the best critique of AI existential risk arguments?
joshc
Aug 30, 2022, 2:18 AM
6
points
11
comments
1
min read
LW
link
How to plan for a radically uncertain future?
Kerry
Aug 30, 2022, 2:14 AM
57
points
35
comments
1
min read
LW
link
EA & LW Forums Weekly Summary (21 Aug − 27 Aug 22′)
Zoe Williams
Aug 30, 2022, 1:42 AM
57
points
4
comments
12
min read
LW
link
Can We Align a Self-Improving AGI?
Peter S. Park
Aug 30, 2022, 12:14 AM
8
points
5
comments
11
min read
LW
link
On the nature of help—a framework for helping
Faustify
Aug 29, 2022, 8:42 PM
3
points
2
comments
13
min read
LW
link
Fundamental Uncertainty: Chapter 4 - Why don’t we do what we think we should?
Gordon Seidoh Worley
Aug 29, 2022, 7:25 PM
15
points
6
comments
13
min read
LW
link
[Question]
How can I reconcile the two most likely requirements for humanities near-term survival.
Erlja Jkdf.
Aug 29, 2022, 6:46 PM
1
point
6
comments
1
min read
LW
link
*New* Canada AI Safety & Governance community
Wyatt Tessari L'Allié
Aug 29, 2022, 6:45 PM
21
points
0
comments
1
min read
LW
link
Are Generative World Models a Mesa-Optimization Risk?
Thane Ruthenis
Aug 29, 2022, 6:37 PM
14
points
2
comments
3
min read
LW
link
Sequencing Intro
jefftk
Aug 29, 2022, 5:50 PM
39
points
3
comments
5
min read
LW
link
(www.jefftk.com)
How Do AI Timelines Affect Existential Risk?
Stephen McAleese
Aug 29, 2022, 4:57 PM
7
points
9
comments
23
min read
LW
link
How might we align transformative AI if it’s developed very soon?
HoldenKarnofsky
Aug 29, 2022, 3:42 PM
140
points
55
comments
45
min read
LW
link
1
review
An Audio Introduction to Nick Bostrom
PeterH
Aug 29, 2022, 8:50 AM
12
points
0
comments
1
min read
LW
link
(forum.effectivealtruism.org)
Please Do Fight the Hypothetical
Lone Pine
Aug 29, 2022, 8:35 AM
18
points
6
comments
3
min read
LW
link
Back to top
Next
N
W
F
A
C
D
E
F
G
H
I
Customize appearance
Current theme:
default
A
C
D
E
F
G
H
I
Less Wrong (text)
Less Wrong (link)
Invert colors
Reset to defaults
OK
Cancel
Hi, I’m Bobby the Basilisk! Click on the minimize button (
) to minimize the theme tweaker window, so that you can see what the page looks like with the current tweaked values. (But remember,
the changes won’t be saved until you click “OK”!
)
Theme tweaker help
Show Bobby the Basilisk
OK
Cancel