Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
Archive
Sequences
About
Search
Log In
All
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
2025
All
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
All
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
Page
2
The Sequences on YouTube
Neil
Jan 7, 2024, 1:44 AM
26
points
9
comments
2
min read
LW
link
AI Risk and the US Presidential Candidates
Zane
Jan 6, 2024, 8:18 PM
41
points
22
comments
6
min read
LW
link
A Challenge to Effective Altruism’s Premises
False Name
Jan 6, 2024, 6:46 PM
−26
points
3
comments
3
min read
LW
link
Lack of Spider-Man is evidence against the simulation hypothesis
RamblinDash
Jan 6, 2024, 6:17 PM
7
points
23
comments
1
min read
LW
link
A Land Tax For Britain
A.H.
Jan 6, 2024, 3:52 PM
6
points
9
comments
4
min read
LW
link
Book review: Trick or treatment (2008)
Fleece Minutia
Jan 6, 2024, 3:40 PM
1
point
0
comments
2
min read
LW
link
Are we inside a black hole?
Jay
Jan 6, 2024, 1:30 PM
2
points
5
comments
1
min read
LW
link
Survey of 2,778 AI authors: six parts in pictures
KatjaGrace
Jan 6, 2024, 4:43 AM
80
points
1
comment
2
min read
LW
link
Project ideas: Epistemics
Lukas Finnveden
Jan 5, 2024, 11:41 PM
43
points
4
comments
LW
link
(www.forethought.org)
Almost everyone I’ve met would be well-served thinking more about what to focus on
Henrik Karlsson
Jan 5, 2024, 9:01 PM
96
points
8
comments
11
min read
LW
link
(www.henrikkarlsson.xyz)
The Next ChatGPT Moment: AI Avatars
kolmplex
and
southpaw
Jan 5, 2024, 8:14 PM
43
points
10
comments
1
min read
LW
link
AI Impacts 2023 Expert Survey on Progress in AI
habryka
Jan 5, 2024, 7:42 PM
28
points
2
comments
7
min read
LW
link
(wiki.aiimpacts.org)
Technology path dependence and evaluating expertise
bhauth
and
Muireall
Jan 5, 2024, 7:21 PM
25
points
2
comments
15
min read
LW
link
The Hippie Rabbit Hole -Nuggets of Gold in Rivers of Bullshit
Jonathan Moregård
Jan 5, 2024, 6:27 PM
39
points
20
comments
8
min read
LW
link
(honestliving.substack.com)
[Question]
What technical topics could help with boundaries/membranes?
Chipmonk
Jan 5, 2024, 6:14 PM
15
points
25
comments
1
min read
LW
link
Catching AIs red-handed
ryan_greenblatt
and
Buck
Jan 5, 2024, 5:43 PM
111
points
27
comments
17
min read
LW
link
AI Impacts Survey: December 2023 Edition
Zvi
Jan 5, 2024, 2:40 PM
34
points
6
comments
10
min read
LW
link
(thezvi.wordpress.com)
Forecast your 2024 with Fatebook
Sage Future
Jan 5, 2024, 2:07 PM
19
points
0
comments
1
min read
LW
link
(fatebook.io)
Predictive model agents are sort of corrigible
Raymond Douglas
Jan 5, 2024, 2:05 PM
35
points
6
comments
3
min read
LW
link
Striking Implications for Learning Theory, Interpretability — and Safety?
RogerDearnaley
Jan 5, 2024, 8:46 AM
37
points
4
comments
2
min read
LW
link
If I ran the zoo
Optimization Process
Jan 5, 2024, 5:14 AM
18
points
1
comment
2
min read
LW
link
Does AI care about reality or just its own perception?
RedFishBlueFish
Jan 5, 2024, 4:05 AM
−6
points
8
comments
1
min read
LW
link
MIRI 2024 Mission and Strategy Update
Malo
Jan 5, 2024, 12:20 AM
223
points
44
comments
8
min read
LW
link
Project ideas: Governance during explosive technological growth
Lukas Finnveden
Jan 4, 2024, 11:51 PM
20
points
0
comments
LW
link
(www.forethought.org)
Hello
S Benfield
Jan 4, 2024, 11:35 PM
6
points
0
comments
2
min read
LW
link
Using Threats to Achieve Socially Optimal Outcomes
StrivingForLegibility
Jan 4, 2024, 11:30 PM
8
points
0
comments
3
min read
LW
link
Best-Responding Is Not Always the Best Response
StrivingForLegibility
Jan 4, 2024, 11:30 PM
10
points
0
comments
3
min read
LW
link
Safety Data Sheets for Optimization Processes
StrivingForLegibility
Jan 4, 2024, 11:30 PM
15
points
1
comment
4
min read
LW
link
The Gears of Argmax
StrivingForLegibility
Jan 4, 2024, 11:30 PM
11
points
0
comments
3
min read
LW
link
Cellular reprogramming, pneumatic launch systems, and terraforming Mars: Some things I learned about at Foresight Vision Weekend
jasoncrawford
Jan 4, 2024, 7:33 PM
28
points
0
comments
8
min read
LW
link
(rootsofprogress.org)
Deep atheism and AI risk
Joe Carlsmith
Jan 4, 2024, 6:58 PM
153
points
22
comments
27
min read
LW
link
Some Vacation Photos
johnswentworth
Jan 4, 2024, 5:15 PM
83
points
0
comments
1
min read
LW
link
AISN #29: Progress on the EU AI Act Plus, the NY Times sues OpenAI for Copyright Infringement, and Congressional Questions about Research Standards in AI Safety
Dan H
and
Corin Katzke
Jan 4, 2024, 4:09 PM
8
points
0
comments
6
min read
LW
link
(newsletter.safe.ai)
EAG Bay Area Satellite event: AI Institution Design Hackathon 2024
beatrice@foresight.org
Jan 4, 2024, 3:02 PM
1
point
0
comments
1
min read
LW
link
AI #45: To Be Determined
Zvi
Jan 4, 2024, 3:00 PM
52
points
4
comments
31
min read
LW
link
(thezvi.wordpress.com)
Screen-supported Portable Monitor
jefftk
Jan 4, 2024, 1:50 PM
16
points
10
comments
1
min read
LW
link
(www.jefftk.com)
[Question]
Which investments for aligned-AI outcomes?
tailcalled
Jan 4, 2024, 1:28 PM
8
points
9
comments
2
min read
LW
link
Non-alignment project ideas for making transformative AI go well
Lukas Finnveden
Jan 4, 2024, 7:23 AM
44
points
1
comment
LW
link
(www.forethought.org)
Fact Checking and Retaliation Against Sources
jefftk
Jan 4, 2024, 12:41 AM
7
points
2
comments
4
min read
LW
link
(www.jefftk.com)
Investigating Alternative Futures: Human and Superintelligence Interaction Scenarios
Hiroshi Yamakawa
Jan 3, 2024, 11:46 PM
1
point
0
comments
17
min read
LW
link
“Attitudes Toward Artificial General Intelligence: Results from American Adults 2021 and 2023”—call for reviewers (Seeds of Science)
rogersbacon
Jan 3, 2024, 8:11 PM
4
points
0
comments
1
min read
LW
link
What’s up with LLMs representing XORs of arbitrary features?
Sam Marks
Jan 3, 2024, 7:44 PM
158
points
64
comments
16
min read
LW
link
Spirit Airlines Merger Play
sapphire
Jan 3, 2024, 7:25 PM
5
points
12
comments
1
min read
LW
link
$300 for the best sci-fi prompt: the results
RomanS
Jan 3, 2024, 7:10 PM
16
points
19
comments
7
min read
LW
link
Agent membranes/boundaries and formalizing “safety”
Chipmonk
Jan 3, 2024, 5:55 PM
26
points
46
comments
3
min read
LW
link
Safety First: safety before full alignment. The deontic sufficiency hypothesis.
Chipmonk
Jan 3, 2024, 5:55 PM
48
points
3
comments
3
min read
LW
link
Practically A Book Review: Appendix to “Nonlinear’s Evidence: Debunking False and Misleading Claims” (ThingOfThings)
tailcalled
3 Jan 2024 17:07 UTC
111
points
25
comments
2
min read
LW
link
(thingofthings.substack.com)
Trivial Mathematics as a Path Forward
ACrackedPot
3 Jan 2024 16:41 UTC
−4
points
2
comments
2
min read
LW
link
Copyright Confrontation #1
Zvi
3 Jan 2024 15:50 UTC
34
points
7
comments
18
min read
LW
link
(thezvi.wordpress.com)
[Question]
Theoretically, could we balance the budget painlessly?
Logan Zoellner
3 Jan 2024 14:46 UTC
4
points
12
comments
1
min read
LW
link
Previous
Back to top
Next
N
W
F
A
C
D
E
F
G
H
I
Customize appearance
Current theme:
default
A
C
D
E
F
G
H
I
Less Wrong (text)
Less Wrong (link)
Invert colors
Reset to defaults
OK
Cancel
Hi, I’m Bobby the Basilisk! Click on the minimize button (
) to minimize the theme tweaker window, so that you can see what the page looks like with the current tweaked values. (But remember,
the changes won’t be saved until you click “OK”!
)
Theme tweaker help
Show Bobby the Basilisk
OK
Cancel