Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
Archive
Sequences
About
Search
Log In
All
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
2025
All
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
All
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
Page
1
[Question]
Is there a fundamental distinction between simulating a mind and simulating *being* a mind? Is this a useful and important distinction?
Thoth Hermes
Apr 8, 2023, 11:44 PM
−17
points
8
comments
2
min read
LW
link
“warning about ai doom” is also “announcing capabilities progress to noobs”
the gears to ascension
Apr 8, 2023, 11:42 PM
23
points
5
comments
3
min read
LW
link
Feature Request: Right Click to Copy LaTeX
DragonGod
Apr 8, 2023, 11:27 PM
18
points
4
comments
1
min read
LW
link
ELCK might require nontrivial scalable alignment progress, and seems tractable enough to try
Alex Lawsen
Apr 8, 2023, 9:49 PM
17
points
0
comments
2
min read
LW
link
GPTs are Predictors, not Imitators
Eliezer Yudkowsky
Apr 8, 2023, 7:59 PM
416
points
100
comments
3
min read
LW
link
3
reviews
4 generations of alignment
qbolec
Apr 8, 2023, 7:59 PM
1
point
0
comments
3
min read
LW
link
The surprising parameter efficiency of vision models
beren
Apr 8, 2023, 7:44 PM
81
points
28
comments
4
min read
LW
link
Random Observation on AI goals
FTPickle
Apr 8, 2023, 7:28 PM
−11
points
2
comments
1
min read
LW
link
Can we evaluate the “tool versus agent” AGI prediction?
Xodarap
Apr 8, 2023, 6:40 PM
16
points
7
comments
LW
link
Relative Abstracted Agency
Audere
Apr 8, 2023, 4:57 PM
14
points
6
comments
5
min read
LW
link
The benevolence of the butcher
dr_s
Apr 8, 2023, 4:29 PM
84
points
33
comments
6
min read
LW
link
1
review
SERI MATS—Summer 2023 Cohort
Aris
,
Ryan Kidd
and
Christian Smith
Apr 8, 2023, 3:32 PM
71
points
25
comments
4
min read
LW
link
AI Proposals at ‘Two Sessions’: AGI as ‘Two Bombs, One Satellite’?
Derek M. Jones
Apr 8, 2023, 11:31 AM
5
points
0
comments
1
min read
LW
link
(www.chinatalk.media)
All images from the WaitButWhy sequence on AI
trevor
Apr 8, 2023, 7:36 AM
73
points
5
comments
2
min read
LW
link
Guidelines for productive discussions
ambigram
Apr 8, 2023, 6:00 AM
38
points
0
comments
5
min read
LW
link
All AGI Safety questions welcome (especially basic ones) [April 2023]
steven0461
Apr 8, 2023, 4:21 AM
57
points
89
comments
2
min read
LW
link
Bringing Agency Into AGI Extinction Is Superfluous
George3d6
Apr 8, 2023, 4:02 AM
28
points
18
comments
5
min read
LW
link
Lagos, Nigeria—ACX Meetups Everywhere 2023
damola
Apr 8, 2023, 3:55 AM
1
point
0
comments
1
min read
LW
link
Upcoming Changes in Large Language Models
Andrew Keenan Richardson
Apr 8, 2023, 3:41 AM
43
points
8
comments
4
min read
LW
link
(mechanisticmind.com)
Consider The Hand Axe
ymeskhout
Apr 8, 2023, 1:31 AM
142
points
16
comments
6
min read
LW
link
AGI as a new data point
Will Rodgers
Apr 8, 2023, 1:01 AM
−1
points
0
comments
1
min read
LW
link
Parametrize Priority Evaluations
SilverFlame
Apr 8, 2023, 12:39 AM
2
points
2
comments
6
min read
LW
link
Pausing AI Developments Isn’t Enough. We Need to Shut it All Down
Eliezer Yudkowsky
Apr 8, 2023, 12:36 AM
271
points
44
comments
12
min read
LW
link
1
review
Humanitarian Phase Transition needed before Technological Singularity
Dr_What
Apr 7, 2023, 11:17 PM
−9
points
5
comments
2
min read
LW
link
[Question]
Thoughts about Hugging Face?
kwiat.dev
Apr 7, 2023, 11:17 PM
7
points
0
comments
1
min read
LW
link
[Question]
Is it correct to frame alignment as “programming a good philosophy of meaning”?
Util
Apr 7, 2023, 11:16 PM
2
points
3
comments
1
min read
LW
link
Select Agent Specifications as Natural Abstractions
lukemarks
Apr 7, 2023, 11:16 PM
19
points
3
comments
5
min read
LW
link
n=3 AI Risk Quick Math and Reasoning
lionhearted (Sebastian Marshall)
Apr 7, 2023, 8:27 PM
6
points
3
comments
4
min read
LW
link
[Question]
What are good alternatives to Predictionbook for personal prediction tracking? Edited: I originally thought it was down but it was just 500 until I though of clearing cookies.
sortega
Apr 7, 2023, 7:18 PM
4
points
4
comments
1
min read
LW
link
Environments for Measuring Deception, Resource Acquisition, and Ethical Violations
Dan H
Apr 7, 2023, 6:40 PM
51
points
2
comments
2
min read
LW
link
(arxiv.org)
Superintelligence Is Not Omniscience
Jeffrey Heninger
Apr 7, 2023, 4:30 PM
16
points
21
comments
7
min read
LW
link
(aiimpacts.org)
An ‘AGI Emergency Eject Criteria’ consensus could be really useful.
tcelferact
Apr 7, 2023, 4:21 PM
5
points
0
comments
LW
link
Reliability, Security, and AI risk: Notes from infosec textbook chapter 1
Orpheus16
Apr 7, 2023, 3:47 PM
34
points
1
comment
4
min read
LW
link
Pre-registering a study
Robert_AIZI
Apr 7, 2023, 3:46 PM
10
points
0
comments
6
min read
LW
link
(aizi.substack.com)
Live discussion at Eastercon
Douglas_Reay
Apr 7, 2023, 3:25 PM
5
points
0
comments
1
min read
LW
link
[Question]
ChatGTP “Writing ” News Stories for The Guardian?
jmh
Apr 7, 2023, 12:16 PM
1
point
4
comments
1
min read
LW
link
Storyteller’s convention, 2223 A.D.
plex
Apr 7, 2023, 11:54 AM
8
points
0
comments
2
min read
LW
link
Stampy’s AI Safety Info—New Distillations #1 [March 2023]
markov
Apr 7, 2023, 11:06 AM
42
points
0
comments
2
min read
LW
link
(aisafety.info)
Beren’s “Deconfusing Direct vs Amortised Optimisation”
DragonGod
Apr 7, 2023, 8:57 AM
52
points
10
comments
3
min read
LW
link
Goal alignment without alignment on epistemology, ethics, and science is futile
Roman Leventov
Apr 7, 2023, 8:22 AM
20
points
2
comments
2
min read
LW
link
Polio Lab Leak Caught with Wastewater Sampling
Cullen
Apr 7, 2023, 1:06 AM
82
points
3
comments
LW
link
Catching the Eye of Sauron
Casey_
Apr 7, 2023, 12:40 AM
221
points
68
comments
4
min read
LW
link
[Question]
How to parallelize “inherently” serial theory work?
Nicholas / Heather Kross
Apr 7, 2023, 12:08 AM
16
points
6
comments
1
min read
LW
link
If Alignment is Hard, then so is Self-Improvement
PavleMiha
Apr 7, 2023, 12:08 AM
21
points
20
comments
1
min read
LW
link
Anthropic is further accelerating the Arms Race?
sapphire
Apr 6, 2023, 11:29 PM
82
points
22
comments
1
min read
LW
link
(techcrunch.com)
Suggestion for safe AI structure (Curated Transparent Decisions)
Kane Gregory
Apr 6, 2023, 10:00 PM
5
points
6
comments
3
min read
LW
link
10 reasons why lists of 10 reasons might be a winning strategy
trevor
6 Apr 2023 21:24 UTC
110
points
7
comments
1
min read
LW
link
A Defense of Utilitarianism
Pareto Optimal
6 Apr 2023 21:09 UTC
−3
points
2
comments
5
min read
LW
link
(paretooptimal.substack.com)
One Does Not Simply Replace the Humans
JerkyTreats
6 Apr 2023 20:56 UTC
9
points
3
comments
4
min read
LW
link
(www.lesswrong.com)
[Question]
Where to begin in ML/AI?
Jake the Student
6 Apr 2023 20:45 UTC
9
points
4
comments
1
min read
LW
link
Back to top
Next
N
W
F
A
C
D
E
F
G
H
I
Customize appearance
Current theme:
default
A
C
D
E
F
G
H
I
Less Wrong (text)
Less Wrong (link)
Invert colors
Reset to defaults
OK
Cancel
Hi, I’m Bobby the Basilisk! Click on the minimize button (
) to minimize the theme tweaker window, so that you can see what the page looks like with the current tweaked values. (But remember,
the changes won’t be saved until you click “OK”!
)
Theme tweaker help
Show Bobby the Basilisk
OK
Cancel