Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
Archive
Sequences
About
Search
Log In
All
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
2025
All
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
All
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
Page
1
[Question]
Near-mode cryonics: A thought experiment
Mati_Roy
Apr 9, 2023, 10:21 PM
3
points
2
comments
1
min read
LW
link
3 entirely different things we call “Time”
Vasyl Dotsenko
Apr 9, 2023, 8:08 PM
0
points
6
comments
2
min read
LW
link
(medium.com)
Foom seems unlikely in the current LLM training paradigm
Ocracoke
Apr 9, 2023, 7:41 PM
18
points
9
comments
1
min read
LW
link
What Piles Up Must Pile Down
silentbob
Apr 9, 2023, 6:37 PM
35
points
4
comments
6
min read
LW
link
[Question]
What games are using the concept of a Schelling point?
Mati_Roy
Apr 9, 2023, 5:21 PM
9
points
13
comments
1
min read
LW
link
[Question]
Review & rebuttal of “Why machines will never rule the world: artificial intelligence without fear”
mikbp
Apr 9, 2023, 3:06 PM
4
points
4
comments
1
min read
LW
link
Being at peace with Doom
Johannes C. Mayer
Apr 9, 2023, 2:53 PM
23
points
14
comments
4
min read
LW
link
1
review
Expanding the domain of discourse reveals structure already there but hidden
TsviBT
Apr 9, 2023, 1:36 PM
30
points
4
comments
6
min read
LW
link
Rooms Available in Downtown Berkeley Group House
Arjun Panickssery
Apr 9, 2023, 10:15 AM
3
points
0
comments
LW
link
AGI Safety Fundamentals 2023 Notes
Lisa Wang
Apr 9, 2023, 7:28 AM
3
points
0
comments
1
min read
LW
link
(lisacontemplates.blogspot.com)
Ng and LeCun on the 6-Month Pause (Transcript)
Stephen Fowler
Apr 9, 2023, 6:14 AM
29
points
7
comments
16
min read
LW
link
Agentized LLMs will change the alignment landscape
Seth Herd
Apr 9, 2023, 2:29 AM
160
points
102
comments
3
min read
LW
link
1
review
A decade of lurking, a month of posting
Max H
Apr 9, 2023, 12:21 AM
70
points
4
comments
5
min read
LW
link
[Question]
Is there a fundamental distinction between simulating a mind and simulating *being* a mind? Is this a useful and important distinction?
Thoth Hermes
Apr 8, 2023, 11:44 PM
−17
points
8
comments
2
min read
LW
link
“warning about ai doom” is also “announcing capabilities progress to noobs”
the gears to ascension
Apr 8, 2023, 11:42 PM
23
points
5
comments
3
min read
LW
link
Feature Request: Right Click to Copy LaTeX
DragonGod
Apr 8, 2023, 11:27 PM
18
points
4
comments
1
min read
LW
link
ELCK might require nontrivial scalable alignment progress, and seems tractable enough to try
Alex Lawsen
Apr 8, 2023, 9:49 PM
17
points
0
comments
2
min read
LW
link
GPTs are Predictors, not Imitators
Eliezer Yudkowsky
Apr 8, 2023, 7:59 PM
416
points
100
comments
3
min read
LW
link
3
reviews
4 generations of alignment
qbolec
Apr 8, 2023, 7:59 PM
1
point
0
comments
3
min read
LW
link
The surprising parameter efficiency of vision models
beren
Apr 8, 2023, 7:44 PM
81
points
28
comments
4
min read
LW
link
Random Observation on AI goals
FTPickle
Apr 8, 2023, 7:28 PM
−11
points
2
comments
1
min read
LW
link
Can we evaluate the “tool versus agent” AGI prediction?
Xodarap
Apr 8, 2023, 6:40 PM
16
points
7
comments
LW
link
Relative Abstracted Agency
Audere
Apr 8, 2023, 4:57 PM
14
points
6
comments
5
min read
LW
link
The benevolence of the butcher
dr_s
Apr 8, 2023, 4:29 PM
84
points
33
comments
6
min read
LW
link
1
review
SERI MATS—Summer 2023 Cohort
Aris
,
Ryan Kidd
and
Christian Smith
Apr 8, 2023, 3:32 PM
71
points
25
comments
4
min read
LW
link
AI Proposals at ‘Two Sessions’: AGI as ‘Two Bombs, One Satellite’?
Derek M. Jones
Apr 8, 2023, 11:31 AM
5
points
0
comments
1
min read
LW
link
(www.chinatalk.media)
All images from the WaitButWhy sequence on AI
trevor
Apr 8, 2023, 7:36 AM
73
points
5
comments
2
min read
LW
link
Guidelines for productive discussions
ambigram
Apr 8, 2023, 6:00 AM
38
points
0
comments
5
min read
LW
link
All AGI Safety questions welcome (especially basic ones) [April 2023]
steven0461
Apr 8, 2023, 4:21 AM
57
points
89
comments
2
min read
LW
link
Bringing Agency Into AGI Extinction Is Superfluous
George3d6
Apr 8, 2023, 4:02 AM
28
points
18
comments
5
min read
LW
link
Lagos, Nigeria—ACX Meetups Everywhere 2023
damola
Apr 8, 2023, 3:55 AM
1
point
0
comments
1
min read
LW
link
Upcoming Changes in Large Language Models
Andrew Keenan Richardson
Apr 8, 2023, 3:41 AM
43
points
8
comments
4
min read
LW
link
(mechanisticmind.com)
Consider The Hand Axe
ymeskhout
Apr 8, 2023, 1:31 AM
142
points
16
comments
6
min read
LW
link
AGI as a new data point
Will Rodgers
Apr 8, 2023, 1:01 AM
−1
points
0
comments
1
min read
LW
link
Parametrize Priority Evaluations
SilverFlame
Apr 8, 2023, 12:39 AM
2
points
2
comments
6
min read
LW
link
Pausing AI Developments Isn’t Enough. We Need to Shut it All Down
Eliezer Yudkowsky
Apr 8, 2023, 12:36 AM
271
points
44
comments
12
min read
LW
link
1
review
Humanitarian Phase Transition needed before Technological Singularity
Dr_What
Apr 7, 2023, 11:17 PM
−9
points
5
comments
2
min read
LW
link
[Question]
Thoughts about Hugging Face?
kwiat.dev
Apr 7, 2023, 11:17 PM
7
points
0
comments
1
min read
LW
link
[Question]
Is it correct to frame alignment as “programming a good philosophy of meaning”?
Util
Apr 7, 2023, 11:16 PM
2
points
3
comments
1
min read
LW
link
Select Agent Specifications as Natural Abstractions
lukemarks
Apr 7, 2023, 11:16 PM
19
points
3
comments
5
min read
LW
link
n=3 AI Risk Quick Math and Reasoning
lionhearted (Sebastian Marshall)
Apr 7, 2023, 8:27 PM
6
points
3
comments
4
min read
LW
link
[Question]
What are good alternatives to Predictionbook for personal prediction tracking? Edited: I originally thought it was down but it was just 500 until I though of clearing cookies.
sortega
Apr 7, 2023, 7:18 PM
4
points
4
comments
1
min read
LW
link
Environments for Measuring Deception, Resource Acquisition, and Ethical Violations
Dan H
Apr 7, 2023, 6:40 PM
51
points
2
comments
2
min read
LW
link
(arxiv.org)
Superintelligence Is Not Omniscience
Jeffrey Heninger
7 Apr 2023 16:30 UTC
16
points
21
comments
7
min read
LW
link
(aiimpacts.org)
An ‘AGI Emergency Eject Criteria’ consensus could be really useful.
tcelferact
7 Apr 2023 16:21 UTC
5
points
0
comments
LW
link
Reliability, Security, and AI risk: Notes from infosec textbook chapter 1
Orpheus16
7 Apr 2023 15:47 UTC
34
points
1
comment
4
min read
LW
link
Pre-registering a study
Robert_AIZI
7 Apr 2023 15:46 UTC
10
points
0
comments
6
min read
LW
link
(aizi.substack.com)
Live discussion at Eastercon
Douglas_Reay
7 Apr 2023 15:25 UTC
5
points
0
comments
1
min read
LW
link
[Question]
ChatGTP “Writing ” News Stories for The Guardian?
jmh
7 Apr 2023 12:16 UTC
1
point
4
comments
1
min read
LW
link
Storyteller’s convention, 2223 A.D.
plex
7 Apr 2023 11:54 UTC
8
points
0
comments
2
min read
LW
link
Back to top
Next
N
W
F
A
C
D
E
F
G
H
I
Customize appearance
Current theme:
default
A
C
D
E
F
G
H
I
Less Wrong (text)
Less Wrong (link)
Invert colors
Reset to defaults
OK
Cancel
Hi, I’m Bobby the Basilisk! Click on the minimize button (
) to minimize the theme tweaker window, so that you can see what the page looks like with the current tweaked values. (But remember,
the changes won’t be saved until you click “OK”!
)
Theme tweaker help
Show Bobby the Basilisk
OK
Cancel