Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
Archive
Sequences
About
Search
Log In
All
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
2025
All
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
All
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
Page
1
[Question]
Where’s the economic incentive for wokism coming from?
Valentine
Dec 8, 2022, 11:28 PM
12
points
105
comments
1
min read
LW
link
I Believe we are in a Hardware Overhang
nem
Dec 8, 2022, 11:18 PM
8
points
0
comments
1
min read
LW
link
Of pumpkins, the Falcon Heavy, and Groucho Marx: High-Level discourse structure in ChatGPT
Bill Benzon
Dec 8, 2022, 10:25 PM
2
points
0
comments
8
min read
LW
link
How Many Lives Does X-Risk Work Save From Nonexistence On Average?
Jordan Arel
Dec 8, 2022, 9:57 PM
4
points
5
comments
14
min read
LW
link
AI Safety Seems Hard to Measure
HoldenKarnofsky
Dec 8, 2022, 7:50 PM
71
points
6
comments
14
min read
LW
link
(www.cold-takes.com)
Playing shell games with definitions
weverka
Dec 8, 2022, 7:35 PM
9
points
3
comments
1
min read
LW
link
Notes on OpenAI’s alignment plan
Alex Flint
Dec 8, 2022, 7:13 PM
40
points
5
comments
7
min read
LW
link
Relevant to natural abstractions: Euclidean Symmetry Equivariant Machine Learning—Overview, Applications, and Open Questions
the gears to ascension
Dec 8, 2022, 6:01 PM
8
points
0
comments
1
min read
LW
link
(youtu.be)
I’ve started publishing the novel I wrote to promote EA
Timothy Underwood
Dec 8, 2022, 5:30 PM
10
points
3
comments
1
min read
LW
link
Neural networks biased towards geometrically simple functions?
DavidHolmes
Dec 8, 2022, 4:16 PM
16
points
2
comments
3
min read
LW
link
If Wentworth is right about natural abstractions, it would be bad for alignment
Wuschel Schulz
Dec 8, 2022, 3:19 PM
29
points
5
comments
4
min read
LW
link
Covid 12/8/22: Another Winter Wave
Zvi
Dec 8, 2022, 2:40 PM
23
points
8
comments
11
min read
LW
link
(thezvi.wordpress.com)
Why I’m Sceptical of Foom
DragonGod
Dec 8, 2022, 10:01 AM
20
points
36
comments
3
min read
LW
link
Take 7: You should talk about “the human’s utility function” less.
Charlie Steiner
Dec 8, 2022, 8:14 AM
50
points
22
comments
2
min read
LW
link
Machine Learning Consent
jefftk
Dec 8, 2022, 3:50 AM
38
points
14
comments
3
min read
LW
link
(www.jefftk.com)
Riffing on the agent type
Quinn
Dec 8, 2022, 12:19 AM
21
points
3
comments
4
min read
LW
link
[Question]
Looking for ideas of public assets (stocks, funds, ETFs) that I can invest in to have a chance at profiting from the mass adoption and commercialization of AI technology
Annapurna
Dec 7, 2022, 10:35 PM
15
points
9
comments
1
min read
LW
link
A Fallibilist Wordview
Toni MUENDEL
Dec 7, 2022, 8:59 PM
−13
points
2
comments
13
min read
LW
link
Thoughts on AGI organizations and capabilities work
Rob Bensinger
and
So8res
Dec 7, 2022, 7:46 PM
102
points
17
comments
5
min read
LW
link
How to Think About Climate Models and How to Improve Them
clans
Dec 7, 2022, 7:37 PM
7
points
0
comments
2
min read
LW
link
(locationtbd.home.blog)
The novelty quotient
River Lewis
Dec 7, 2022, 5:16 PM
4
points
7
comments
2
min read
LW
link
(heytraveler.substack.com)
ChatGPT: “An error occurred. If this issue persists...”
Bill Benzon
Dec 7, 2022, 3:41 PM
5
points
11
comments
3
min read
LW
link
Take 6: CAIS is actually Orwellian.
Charlie Steiner
Dec 7, 2022, 1:50 PM
12
points
8
comments
2
min read
LW
link
Peter Thiel on Technological Stagnation and Out of Touch Rationalists
Matt Goldenberg
Dec 7, 2022, 1:15 PM
9
points
26
comments
1
min read
LW
link
(youtu.be)
[Link] Wavefunctions: from Linear Algebra to Spinors
sen
Dec 7, 2022, 12:44 PM
11
points
12
comments
1
min read
LW
link
(paperclip.substack.com)
Why I like Zulip instead of Slack or Discord
Alok Singh
Dec 7, 2022, 9:28 AM
31
points
10
comments
1
min read
LW
link
Bioweapons, and ChatGPT (another vulnerability story)
Beeblebrox
Dec 7, 2022, 7:27 AM
−5
points
0
comments
2
min read
LW
link
Where to be an AI Safety Professor
scasper
Dec 7, 2022, 7:09 AM
31
points
12
comments
2
min read
LW
link
[Question]
Are there any tools to convert LW sequences to PDF or any other file format?
quetzal_rainbow
Dec 7, 2022, 5:28 AM
2
points
2
comments
1
min read
LW
link
Manifold Markets community meetup
Sinclair Chen
Dec 7, 2022, 3:25 AM
4
points
0
comments
1
min read
LW
link
“Attention Passengers”: not for Signs
jefftk
Dec 7, 2022, 2:00 AM
27
points
10
comments
1
min read
LW
link
(www.jefftk.com)
[ASoT] Probability Infects Concepts it Touches
Ulisse Mini
Dec 7, 2022, 1:48 AM
10
points
4
comments
1
min read
LW
link
Simple Way to Prevent Power-Seeking AI
research_prime_space
Dec 7, 2022, 12:26 AM
12
points
1
comment
1
min read
LW
link
In defense of probably wrong mechanistic models
evhub
Dec 6, 2022, 11:24 PM
55
points
10
comments
2
min read
LW
link
AI Safety in a Vulnerable World: Requesting Feedback on Preliminary Thoughts
Jordan Arel
Dec 6, 2022, 10:35 PM
4
points
2
comments
3
min read
LW
link
ChatGPT and the Human Race
Ben Reilly
Dec 6, 2022, 9:38 PM
6
points
1
comment
3
min read
LW
link
[Question]
How do finite factored sets compare with phase space?
Alex_Altair
Dec 6, 2022, 8:05 PM
15
points
1
comment
1
min read
LW
link
Mesa-Optimizers via Grokking
orthonormal
Dec 6, 2022, 8:05 PM
36
points
4
comments
6
min read
LW
link
Using GPT-Eliezer against ChatGPT Jailbreaking
Stuart_Armstrong
and
rgorman
Dec 6, 2022, 7:54 PM
170
points
85
comments
9
min read
LW
link
The Parable of the Crimp
Phosphorous
Dec 6, 2022, 6:41 PM
11
points
3
comments
3
min read
LW
link
The Categorical Imperative Obscures
Gordon Seidoh Worley
Dec 6, 2022, 5:48 PM
17
points
17
comments
2
min read
LW
link
MIRI’s “Death with Dignity” in 60 seconds.
Cleo Nardo
Dec 6, 2022, 5:18 PM
58
points
4
comments
1
min read
LW
link
Things roll downhill
awenonian
Dec 6, 2022, 3:27 PM
19
points
0
comments
1
min read
LW
link
EA & LW Forums Weekly Summary (28th Nov − 4th Dec 22′)
Zoe Williams
Dec 6, 2022, 9:38 AM
10
points
1
comment
LW
link
Take 5: Another problem for natural abstractions is laziness.
Charlie Steiner
Dec 6, 2022, 7:00 AM
31
points
4
comments
3
min read
LW
link
Verification Is Not Easier Than Generation In General
johnswentworth
Dec 6, 2022, 5:20 AM
73
points
27
comments
1
min read
LW
link
Shh, don’t tell the AI it’s likely to be evil
naterush
Dec 6, 2022, 3:35 AM
19
points
9
comments
1
min read
LW
link
[Question]
What are the major underlying divisions in AI safety?
Chris_Leong
6 Dec 2022 3:28 UTC
5
points
2
comments
1
min read
LW
link
[Link] Why I’m optimistic about OpenAI’s alignment approach
janleike
5 Dec 2022 22:51 UTC
98
points
15
comments
1
min read
LW
link
(aligned.substack.com)
The No Free Lunch theorem for dummies
Steven Byrnes
5 Dec 2022 21:46 UTC
37
points
16
comments
3
min read
LW
link
Back to top
Next
N
W
F
A
C
D
E
F
G
H
I
Customize appearance
Current theme:
default
A
C
D
E
F
G
H
I
Less Wrong (text)
Less Wrong (link)
Invert colors
Reset to defaults
OK
Cancel
Hi, I’m Bobby the Basilisk! Click on the minimize button (
) to minimize the theme tweaker window, so that you can see what the page looks like with the current tweaked values. (But remember,
the changes won’t be saved until you click “OK”!
)
Theme tweaker help
Show Bobby the Basilisk
OK
Cancel